Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
MULTI-RANGE SOLID STATE LIDAR SYSTEM
Document Type and Number:
WIPO Patent Application WO/2020/132417
Kind Code:
A1
Abstract:
A Lidar system includes a first photodetector having a first field of view and a second photodetector having a second field of view. The system includes a first light source aimed at the first field of view and a second light source aimed at the second field of view. The system is designed to distinguish between light reflected by an object in the first field of view and light reflected by an object in the second field of view.

Inventors:
SMITH ELLIOT (US)
Application Number:
PCT/US2019/067796
Publication Date:
June 25, 2020
Filing Date:
December 20, 2019
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
CONTINENTAL AUTOMOTIVE SYSTEMS INC (US)
International Classes:
G01S7/481; G01S17/89
Foreign References:
US20180067195A12018-03-08
US20180106890A12018-04-19
US20180100928A12018-04-12
Attorney, Agent or Firm:
BINKOWSKI, Matthew et al. (US)
Download PDF:
Claims:
CLAIMS

What is claimed is:

1. A system comprising:

a first photodetector having a first field of view;

a second photodetector having a second field of view overlapping the first field of view;

a first light source aimed at the first field of view;

a second light source aimed at the second field of view; and

a first bandpass filter covering the first photodetector and a second bandpass filter covering the second photodetector, the first bandpass filter designed to transmit light in a first bandwidth and the second bandpass filter designed to transmit light in a second bandwidth different than the first bandwidth.

2. The system of claim 1 , wherein the first light source is designed to emit light having a first wavelength in the first bandwidth and the second light source is designed to emit light having a second wavelength in the second bandwidth.

3. The system of claim 2, wherein the difference between the center wavelength of a wavelength curve of the second bandpass filter and the center wavelength of a wavelength curve of the first bandpass filter plus the full-width half-maximum of the wavelength curve of the light emitted from the first light source is greater than the full-width half-maximum of the wavelength curve of the first bandpass filter..

4. The system of claim 1, wherein the first and second light sources are designed to emit pulsed laser light.

5. The system of claim 1, wherein the first field of view is wider than second field of view.

6. The system of claim 5, wherein the first field of view is shorter than second field of view.

7. The system of claim 1, wherein the first receiving unit and the second receiving unit are on a vehicle.

8. The system of claim 1, wherein the first field of view and the second field of view are aimed in substantially the same direction.

9. The system of claim 1, further comprising a controller programmed to emit light simultaneously from the first light source and the second light source.

10. The system of claim 1, further comprising a casing supporting the first and second photodetectors, the first and second light sources, and the first and second bandpass filters.

11. A method comprising:

emitting light from a first light source;

emitting light from a second light source;

filtering to a first bandwidth light that is emitted by the first light source and reflected by a reflecting surface;

filtering to a second bandwidth light that is emitted by the second light source and reflected by a reflecting surface, the second bandwidth being different than the first bandwidth;

with a first photodetector, detecting light in the first bandwidth transmitted by the first bandpass filter; and

with a second photodetector, detecting light in the second bandwidth transmitted by the second bandpass filter.

12. The method as set forth in claim 11, wherein emitting light from the first light source and emitting light from the second light source are substantially simultaneous.

13. The method as set forth in claim 11, wherein emitting light from the first light source includes emitting light at a first wavelength that is within the first bandwidth, and wherein emitting light from the second light source includes emitting light at a second wavelength that is within the second bandwidth.

14. The method of claim 11, wherein the difference between the center wavelength of a wavelength curve of the second bandpass filter and the center wavelength of a wavelength curve of the first bandpass filter plus the full-width half-maximum of the wavelength curve of the light emitted from the first light source is greater than the full-width half-maximum of the wavelength curve of the first bandpass filter.

15. The method of claim 11, wherein the first light source and the second light source emit pulsed laser light.

16. A system comprising:

a first photodetector having a first field of view;

a second photodetector having a second field of view overlapping the first field of view;

a first light source aimed at the first field of view;

a second light source aimed at the second field of view; and

a controller programmed to:

substantially simultaneously emit a pulse of light from the first light source and the second light source;

during a first time period, activate the first photodetector and deactivate the second photodetector; and

during a second time period, activate the second photodetector and deactivate the first photodetector, the second time period initiating after the first time period.

17. The system of claim 16, wherein the first time period initiates simultaneously with emission of light from the first and second light sources.

18. The system of claim 16, wherein the first field of view is wider than second field of view.

19. The system of claim 18, wherein the first field of view is shorter than second field of view.

20. The system of claim 16, wherein the first field of view and the second field of view are aimed in substantially the same direction.

21. The system of claim 16, wherein the first time period and the second time period overlap.

22. A method comprising:

emitting light from a first light source into a field of view of a first photodetector and simultaneously emitting light from a second light source into a field of view of a second photodetector, the field of view of the second photodetector overlapping the field of view of the first photodetector;

during a first time period, activating the first photodetector and deactivating the second photodetector with a first photodetector, detecting light in the first bandwidth transmitted by the first bandpass filter; and

with a second photodetector, detecting light in the second bandwidth transmitted by the second bandpass filter.

23. The method as set forth in claim 22, wherein the first time period initiates simultaneously with emission of light from the first and second light sources.

24. The method as set forth in claim 22, wherein the first time period and the second time period overlap.

Description:
MULTI-RANGE SOLID STATE LIDAR SYSTEM

BACKGROUND

[0001] A solid-state Lidar system includes a photodetector or an array of photodetectors that is essentially fixed in place relative to a carrier, e.g., a vehicle. Light is emitted into the field of view of the photodetector and the photodetector detects light that is reflected by an object in the field of view. For example, a Flash Lidar system emits pulses of light, e.g., laser light, into the field of view. The detection of reflected light is used to generate a 3D environmental map of the surrounding environment. The time of flight of the reflected photon detected by the photodetector is used to determine the distance of the object that reflected the light.

[0002] The solid-state Lidar system may be mounted to a vehicle to detect objects in the environment surrounding the vehicle and to detect distance of those objects for environmental mapping. The output of the solid-state Lidar system may be used, for example, to autonomously or semi -autonomously control operation of the vehicle, e.g., propulsion, braking, steering, etc. Specifically, the system may be a component of or in communication with an advanced driver- assistance system (ADAS) of the vehicle.

[0003] In instances where the vehicle uses both short-range and long-range fields of view to generate the 3D map of the surrounding environment, difficulties may exist in distinguishing long- range reflections and short-range reflections such that they do not influence the distance measurement of the other.

BRIEF DESCRIPTION OF THE DRAWINGS

[0004] Figure l is a perspective view of a vehicle with a Lidar system showing a 3D map of the objects detected by the Lidar system.

[0005] Figure 2 is a block diagram of one example of the Lidar system.

[0006] Figure 3 is a block diagram of another example of the Lidar system.

[0007] Figure 4 is a block diagram of another example of the Lidar system.

[0008] Figure 5 is a block diagram of another example of the Lidar system.

[0009] Figure 6 is a schematic of the operation of the example of Figure 2.

[0010] Figure 7 is a schematic of the operation of the example of Figure 3. [0011] Figure 8 is a graph showing operation of first and second light sources and first and second bandwidth filters.

[0012] Figure 9 is a flow chart of a method of operating the examples of Figures 2 and 3.

[0013] Figure 10 is a schematic of the operation of the example of Figure 4.

[0014] Figure 11 is a schematic of the operation of the example of Figure 5.

[0015] Figure 12 is a method of operating the examples of Figures 4 and 5.

DETAILED DESCRIPTION

[0016] With reference to the Figures, wherein like numerals indicate like parts throughout the several views, a system 10 is generally shown. Specifically, the system 10 is a light detection and ranging (Lidar) system. With reference to Figure 1, the system 10 includes two fields of view FV1, FV2 and emits light into the fields of view FV1, FV2. The system 10 detects the emitted light that is reflected by objects in the fields of view FV1, FV2, e.g., pedestrians, street signs, vehicle 12, etc. As described below, the system 10 separately collects the reflected light from the two fields of view FV1, FV2.

[0017] The system 10 is shown in Figure 1 as being mounted to a vehicle 12. In such an example, the system 10 is operated to detect objects in the environment surrounding the vehicle 12 and to detect distance of those objects for environmental mapping. The output of the system 10 may be used, for example, to autonomously or semi-autonomously control operation of the vehicle 12, e.g., propulsion, braking, steering, etc. Specifically, the system 10 may be a component of or in communication with an advanced driver-assistance system (ADAS) of the vehicle 12. The system 10 may be mounted to the vehicle 12 in any suitable position (as one example, the system 10 is shown on the front of the vehicle 12 and directed forward). The vehicle 12 may have more than one system 10 and/or the vehicle 12 may include other object detection system 10, including other Lidar systems. The vehicle 12 is shown in Figure 1 as including a single system 10 aimed in a forward direction merely as an example. The vehicle 12 shown in the Figures is a passenger automobile. As other examples, the vehicle 12 may be of any suitable manned or un-manned type including a plane, satellite, drone, watercraft, etc.

[0018] The system 10 may be a solid-state Lidar system. In such an example, the system 10 is stationary relative to the vehicle 12. For example, the system 10 may include a casing 14 that is fixed relative to the vehicle 12 and a silicon substrate of the system 10 is fixed to the casing 14. The system 10 may be a staring, non-moving system. As another example, the system 10 may include elements to adjust the aim of the system 10, e.g., the direction of the emitted light may be controlled by, for example, optics, mirrors, etc.

[0019] As a solid-state Lidar system, the system 10 may be a Flash Lidar system. In such an example, the system 10 emits pulses of light into the fields of view FV1, FV2. More specifically, the system 10 may be a 3D Flash Lidar system 10 that generates a 3D environmental map of the surrounding environment, as shown in part in Figure 1. An example of a compilation of the data into a 3D environmental map is shown in the fields of view FV1, FV2 in Figure 1.

[0020] Four examples of the system 10 are shown in Figures 2-5, respectively. Common numerals are used to identify common features among the examples. With reference to Figures 2- 5, the system 10 includes a system controller 16 and two pairs of light sources 18, 22 and photodetectors 20, 24. Specifically, one pair includes a first light source 18 and a first photodetector 20, and the other pair includes a second light source 22 and a second photodetector 24. As described further below, the first light source 18 emits light into the first field of view of the first photodetector 20, i.e., the first field of view FV1, and the second light source 22 emits light into the field of view of the second photodetector 24, i.e., the second field of view FV2. With reference to Figures 3 and 5, the system 10 may also include a third pair having a third light source 26 and a third photodetector 28, in which case the third light source 26 emits light into the field of view of the third photodetector 28, i.e., the third field of view FV3. The system 10 may include two or more pairs of light sources and photodetectors.

[0021] The system 10 may be a unit. In other words, the first light source 18, first photodetector 20, second light source 22, second photodetector 24, and the system controller 16 may be supported by a common substrate that is attached to the vehicle 12, e.g., a casing 14 as schematically shown in Figures 2-5. In the examples shown in Figures 3 and 5, the third light source 26 and the third photodetector 28 are supported by the casing 14. The casing 14 may, for example, enclose the other components of the system 10 and may include mechanical attachment features to attach the casing 14 to the vehicle 12 and electronic connections to connect to and communicate with electronic system 10 of the vehicle 12, e.g., components of the ADAS. The casing 14, for example, may be plastic or metal and may protect the other components of the system 10 from environmental precipitation, dust, etc. In the alternative to the system 10 being a unit, components of the system may be separated and disposed at different locations of the vehicle 12. In such examples, the system 10 may include multiple casings with each casing containing components of the system 10. As one example, one casing may including one or more of the pairs of light sources 18, 22, 26 and photodetectors 20, 24, 28 and another casing may include one or more of the pairs of light sources 18, 22, 26 and photodetectors 20, 24, 28. As another example, in addition or in the alternative, one or more of the pairs of light sources 18, 22, 26 and photodetectors 20, 24, 28 may be split among separate casings. In such examples, the system 10 may include any suitable number of casings.

[0022] The controller 16 may be a microprocessor-based controller or field programmable gate array (FPGA) implemented via circuits, chips, and/or other electronic components. In other words, the controller 16 is a physical, i.e., structural, component of the system 10. For example, the controller 16 may include a processor, memory, etc. The memory of the controller 16 may store instructions executable by the processor, i.e., processor-executable instructions, and/or may store data. The controller 16 may be in communication with a communication network of the vehicle 12 to send and/or receive instructions from the vehicle 12, e.g., components of the ADAS.

[0023] As described further below, the controller 16 communicates with the light sources 18, 22, 26 and the photodetectors 20, 24, 28. Specifically, the controller 16 instructs the first light source 18 to emit light and substantially simultaneously initiates a clock. When the light is reflected, i.e., by an object in the first field of view FV1, the first photodetector 20 detects the reflected light and communicates this detection to the controller 16, which the controller 16 uses to identify object location and distance to the object (based time of flight of the detected photon using the clock initiated at the emission of light from the first light source 18). The controller 16 uses these outputs from the first photodetector 20 to create the environmental map and/or communicates the outputs from the first photodetector 20 to the vehicle 12, e.g., components of the ADAS, to create the environmental map. Specifically, the controller 16 continuously repeats the light emission and detection of reflected light for building and updating the environmental map. While the first light source 18 and first photodetector 20 were used as examples, the controller 16 similarly communicates with second light source 22 and second photodetector 24 and with the third light source 26 and the third photodetector 28.

[0024] The light sources 18, 22, 26 emit light into the fields of view FV1, FV2, FV3, respectively, for detection by the respective photodetector when the light is reflected by an object in the respective field of view FV1, FV2, FV3. The light sources 18, 22, 26 may have similar or identical architecture and/or design. For example, the light sources 18, 22, 26 may include the same type of components arranged in the same manner, in which case the corresponding components of the light sources 18, 22, 26 may be identical or may have varying characteristics (e.g., for emission of different light wavelengths as described below).

[0025] With reference to Figures 2-5, the light sources 18, 22, 26 may each include a light emitter (i.e., a first light emitter 30, a second light emitter 32, a third light emitter 34). For example, the light emitter 30, 32, 34 may be a laser. The light emitter 30, 32, 34 may be, for example, a semiconductor laser. In one example, the light emitter 30, 32, 34 is a vertical-cavity surface-emitting laser (VCSEL). As another example, the light emitter 30, 32, 34 may be a diode-pumped solid-state laser (DPSSL). As another example, the light emitter 30, 32, 34 may be an edge emitting laser diode. The light sources 18, 22, 26 may be designed to emit a pulsed flash of light, e.g., a pulsed laser light. Specifically, the light emitter 30, 32, 34, e.g., the VCSEL, is designed to emit a pulsed laser light. The light emitted by the light emitters 30, 32,

34 may be, for example, infrared light. Alternatively, the light emitted by the light emitters 30, 32, 34 may be of any suitable wavelength, as also described further below.

[0026] As set forth above, the system 10 may be a staring, non-moving system. As another example, the system 10 may include elements to adjust the aim of the system 10. For example, with continued reference to Figures 2-5, any of the light sources 18, 22, 26 may include beam steering device 36 that direct and/or diffuse the light from the light emitter 30, 32, 34 into the respective field of view FV1, FV2, FV3. In the example shown in Figure 1, the light beams are emitted from the system 10 as horizontal lines. In such an example, the beam steering device 36 may emit the light as the horizontal beam and/or may adjust the vertical location of the light beam. While the beam steering devices 36 are shown in Figures 2-5, it should be appreciated that the beam steering devices 36 may be eliminated from Figures 2-5 in an example where the system 10 is a staring, non-moving system. In such an example, the light source 18, 22, 26, e.g., the VCSEL, exposes the entire respective field of view FV1, FV2, FV3 at once, i.e., at the same time.

[0027] In examples including the beam steering device 36, the beam steering device 36 may be a micromirror. For example, the beam steering device 36 may be a micro-electro-mechanical system 10 (MEMS) mirror. As an example, the beam steering device 36 may be a digital micromirror device (DMD) that includes an array of pixel -mirrors that are capable of being tilted to deflect light. As another example, the MEMS mirror may include a mirror on a gimbal that is tilted, e.g., by application of voltage. As another example, the beam steering device 36 may be a liquid-crystal solid-state device. While the first, second, and third beam steering devices are all labeled with reference numeral 36, it should be appreciated that the beam steering devices 36 may of the same type or different types; and in examples in which the beam steering devices 36 are of the same type, the beam steering devices 36 may be identical or may have different characteristics.

[0028] With continued reference to Figures 2-5, the system 10 includes transmission optics through which light exits the system 10. Specifically, the first light source 18 includes first transmission optics 38 through which light emitted by the first light emitter 30 exits the system 10 into the first field of view FV1 and second transmission optics 40 through which light emitted by the second light emitter 32 exits the system 10 into the second field of view FV2. The third light source 26 includes third transmission optics 42 through which light emitted by the third light emitter 34 exits the system 10 into the third field of view FV3. The transmission optics 38, 40, 42 may include any suitable number of lenses. The transmission optics 38, 40, 42 may have similar or identical architecture and/or design. For example, the transmission optics 38, 40, 42 may include the same type of components arranged in the same manner, in which case the corresponding components of the transmission optics 38, 40, 42 may be identical or may have varying characteristics (e.g., for emission of different light wavelengths as described below).

[0029] The first light source 18 is aimed at the first field of view FV1 and the second light source 22 is aimed at the second field of view FV2. Specifically, the system 10 emits light from the first light source 18 into a first field of illumination and emits light from the second light source 22 into a second field of illumination. In the examples shown in Figures 3 and 5, the third light source 26 is aimed at the third field of view FV3. The field of illumination is the area exposed to light emitted from the light sources 18, 22, 26. The first field of illumination may substantially match the first field of view FV1 and the second field of illumination may substantially match the second field of view FV2 (“substantially match” is based on

manufacturing capabilities and tolerances of the light sources 18, 22, 26 and the photodetectors 20, 24, 28), as is the case shown for example in Figure 1. The first field of illumination and the second field of illumination may overlap, as described further below. In the examples shown in Figures 3 and 5, the second field of illumination and the third field of illumination may overlap, as described further below.

[0030] With continued reference to Figures 2-5, the system 10 includes a first receiving unit 44 and a second receiving unit 46. In the examples shown in Figures 3 and 5, the system 10 may include a third receiving unit 48. The first receiving unit 44 includes the first photodetector 20 and may include first receiving optics 50. The second receiving unit 46 includes the second photodetector 24 and may include second receiving optics 52. The third receiving unit 48 includes the third photodetector 28 and may include third receiving optics 54.

[0031] For the purposes of this disclosure, the term“photodetector” includes a single photodetector or an array of photodetectors, e.g., an array of photodiodes. The photodetectors 20, 24, 28 may be, for example, avalanche photodiode detectors. As one example, the photodetectors 20, 24, 28 may be a single-photon avalanche diode (SPAD). As another example, the photodetectors 20, 24, 28 may be a PIN diode. The photodetectors 20, 24, 28 may have similar or identical architecture and/or design. For example, the photodetectors 20, 24, 28 may include the same type of components arranged in the same manner, in which case the

corresponding components of the photodetectors 20, 24, 28 may be identical or may have varying characteristics.

[0032] The first field of view FV1 is the area in which reflected light may be sensed by the first photodetector 20, the second field of view FV2 is the area in which reflected light may be sensed by the second photodetector 24, and the third field of view FV3 is the area in which reflected light may be sensed by the third photodetector 28. The first field of view FV1 and the second field of view FV2 may overlap. In other words, as least part of the first field of view FV1 and at least part of the second field of view FV2 occupy the same space such that an object in the overlap will reflect light toward both photodetectors 20, 24. For example, as shown in Figures 6 and 10, the first field of view FV1 and the second field of view FV2 may be centered on each other, i.e., aimed in substantially the same direction (“substantially the same” is based on manufacturing capabilities and tolerances of the light sources 18, 22, 26 and the photodetectors 20, 24, 28). In the examples shown in Figures 7 and 11, the first field of view FV1 and the second field of view FV2 may be aimed in different directions while overlapping. With continued reference to Figures 7 and 11, the second field of view FV2 and third field of view FV3 may overlap. In Figures 7 and 11, the second field of view FV2 and the third field of view FV3 may be aimed in different directions while overlapping. In the example shown in Figures 7 and 11, the first field of view FV1, the second field of view FV2, and the third field of view FV3 are each aimed in a different direction.

[0033] The fields of view FV1, FV2, FV3 may have different widths and/or lengths. In the examples shown in Figures 6 and 10, the length of the first field of view FV1 is shorter than the length of the second field of view FV2. In other words, the first photodetector 20 has a short range and the second photodetector 24 has a long range. In the examples shown in Figures 6 and 10, the first field of view FV1 is wider than the second field of view FV2.

[0034] Light reflected in the fields of view FV1, FV2, FV3 is reflected to receiving optics 50, 52, 54. The receiving optics 50, 52, 54 may include any suitable number of lenses, filters, etc.

[0035] The system 10 may distinguish between the reflected light that was emitted by the first light source 18 and reflected light that was emitted by the second light source 22 based on differences in wavelength of the light. For example, with reference to Figures 2 and 3, the first light source 18 and the second light source 22 may emit light having different wavelengths lΐ, l2 and the first receiving unit 44 and the second receiving unit 46 may detect light having different wavelengths lΐ, l2. In other words, the first receiving unit 44 may be designed to detect the wavelength lΐ of light that was emitted from the first light source 18 and reflected by a reflecting surface in the first field of view FV1 (and detect little or no light at wavelength l2 emitted from the second light source 22) and the second receiving unit 46 may be designed to detect wavelength l2 of the light that was emitted from the second light source 22 and reflected by a reflecting surface in the second field of view FV2 (and detect little or no light at wavelength lΐ emitted from the first light source 18).

[0036] Similarly, the third receiving unit 48 may be designed to detect light that was emitted from the third light source 26 at a third wavelength l3 and reflected by a reflecting surface in the third field of view FV3 (and detect little or no light emitted at wavelengths lΐ, l2 from the first light source 18 and the second light source 22). As another example, the third receiving unit 48 may be designed to detect light that was emitted from the third light source 26 at the first wavelength lΐ and reflected by a reflecting surface in the third field of view FV3. In such an example, the first receiving unit 44 and the third receiving unit 48 are pointed in different directions such that the first and third fields of view FV1, FV3 do not overlap (see Fig. 7). [0037] Figure 6 is a schematic showing the operation of the example shown in Figure 2 and Figure 7 is a schematic showing the operation of the example, shown in Figure 3. As set forth above, with reference to Figures 6 and 7, the first light source 18 is designed to emit light having a first wavelength lΐ (see Fig. 8) and the second light source 22 is designed to emit light having a second wavelength l2 (see Fig. 8). First wavelength lΐ and second wavelength l2 are different. In addition, the first receiving unit 44 transmits light at the first wavelength lΐ, i.e., light reflected in the first field of view FV1, and filters out light at the second wavelength l2, and the second receiving unit 46 transmits light at the second wavelength l2, i.e., light reflected in the second field of view FV2, and filters out light at the first wavelength lΐ . Using this filtering, the system is able to distinguish between reflections in the first and second fields of view FV1, FV2.

[0038] With reference to Figures 2 and 3, the first light source 18, the second light source 22, the third light source 26 may be designed to emit light at the desired wavelength. As an example, the first light emitter 30, the second light emitter 32, and the third light emitter 34 may be designed to generate and emit light at the desired wavelength. As another example in addition or in the alternative to the design of the light emitters 30, 32, 34, the transmitting optics may include bandpass filters that filter the light emitted from the light emitters 30, 32, 34 to the desired wavelength.

[0039] With reference to Figures 2 and 3, the first receiving optics 50 may include a first bandpass filter 56 and the second receiving optics 52 may include a second bandpass filter 58. With reference to Figure 3, the third receiving optics 54 may include a third bandpass filter 60. The bandpass filters 56, 58, 60 are optical filters, i.e., physical elements that are at least part of the receiving optics 50, 52, 54. In the examples shown in Figures 2 and 3, the first bandpass filter 56 covers the first photodetector 20 and the second bandpass filter 58 covers the second photodetector 24, i.e., reflected light entering the system 10 travels through at least one of the bandpass filters. With reference to Figure 3, the third bandpass filter 60 covers the third photodetector 28. The bandpass filters 56, 58, 60 may be narrow-bandpass filters.

[0040] Figure 8 shows operation of the bandpass filters 56, 58. Figure 8 shows the wavelength curves of the light emitted from the first light source 18 and the second light source 22 shown in solid lines. The dotted lines show the wavelength curve of the first bandpass filter 56 having a first bandwidth BW 1 and the second bandpass filter having a second bandwidth BW2. The first bandpass filter 56 is designed to transmit light in the first bandwidth BW 1 (and attenuate outside the first bandwidth BW1) and the second bandpass filter 58 designed to transmit light in a second bandwidth BW2 (and attenuate outside the second bandwidth BW2). The wavelength range of light emitted from the first light source 18 is in the first bandwidth BW 1 and the wavelength range of light emitted from the second light source 22 is in the second bandwidth BW2. In other words, the first light source 18 is designed to emit light in the first bandwidth BW 1 and the second light source 22 is designed to emit light in the second bandwidth BW2. For example, in Figure 8 both the wavelength curve of light emitted from the first light source 18 and the wavelength curve of the first bandpass filter 56 have a center wavelength at lΐ and the first bandwidth BW1 is larger than the range of wavelengths emitted by the first light source 18. Similarly, both the wavelength curve of light emitted from the second light source 22 and the wavelength curve of the second bandpass filter 58 have a center wavelength at l2 and the first bandwidth BW2 is larger than the range of wavelengths emitted by the second light source 22. In the example shown in Figure 8, the center wavelength of the wavelength curve emitted from the first light source 18 is outside the second bandwidth BW2, and the center wavelength of the wavelength curve emitted from the second light source 22 is outside the first bandwidth BW 1.

[0041] In the example shown in Figure 8, the light sources 18, 22 and bandpass filters 56, 58 are designed such that the difference between the center wavelength CW2 (i.e., at peak transmission) of the wavelength curve second bandwidth BW2 and the center wavelength CW 1 (i.e., at peak transmission) of the second bandwidth BW1 plus the full-width half-maximum of the wavelength curve of light emitted from the first light source 18 is greater than the full-width half-maximum of the curve of the first bandpass filter 56.

[0042] In other words, the light sources 18, 22 and the bandpass filters 56, 58 may be designed according the following relationship:

Dl+ p > V

where

CW 1 = center wavelength of light transmitted by first bandpass filter 56;

CW2 = center wavelength of light transmitted by the second bandpass filter 58;

Dl = CW 1 - CW2; p = FWHM of wavelength curve emitted by first or second light source 18, 22; and V = FWHM of wavelength curve of the bandpass filter 56, 58 covering the first or second photodetector 20, 24.

[0043] This relationship reduces cross-talk, e.g., the first photodetector 20 detecting reflected light generated by the second light source 22 and the second photodetector 24 detecting reflected light generated by the first light source 18. Any remaining“false signals” from cross-talk data points may be removed, for example, by using histogramming.

[0044] While the first and second light sources 18, 22 and bandpass filters 56, 58 are described above, the relationship between the second light source 22 and the third light source 26 may be similar or identical to that described above. As one example, with reference to Figure 7, the first and third light source 22, 26 may be identical and first and third bandpass filters 56, 60 may be identical.

[0045] In such examples shown in Figures 2, 3, 6, and 7, the relationship described above allows for light to be emitted substantially simultaneously from the first light source 18 and the second light source 22 (and the third light source 26 in the example in Figures 3 and 7). In other words, the controller 16 may be programmed to emit light substantially simultaneously from the first light source 18 and the second light source 22 (and the third light source 26 in the example in Figures 3 and 7). In other words, the controller 16 may substantially simultaneously instruct the first light source 18 to emit light and the second light source 22 to emit light and substantially simultaneously initiate a clock for both or each photodetector 20, 24, 28 (and similarly for the third light source 26 in the example in Figures 3 and 7).“Substantially simultaneously” is based on given manufacturing capabilities and tolerances of the light sources 18, 22, 26 and the photodetectors 20, 24, 28.

[0046] Figure 9 shows an example method 900 of operation of the system 10 in Figures 2 and 3. As shown in block 910, the method includes emitting light from the first light source 18 and the second light source 22. Specifically, the method may include substantially

simultaneously emitting light from the first light source 18 and the second light source 22. In the example of Figure 3, the method may also include substantially simultaneously emitting light from the third light source 26. In other words, the controller 16 instructs the first light source 18 and the second light source 22 (and the third light source 26 in Figure 3) to emit light substantially simultaneously.

[0047] Specifically, block 910 may include emitting light from the first light source 18 at the first wavelength lΐ that is within the first bandwidth BW1, i.e., the bandwidth transmitted by the first bandpass filter 56, and may include emitting light from the second light source 22 at the second wavelength l2 that is within the second bandwidth BW2, i.e., the bandwidth transmitted by the second bandpass filter 58. The difference between the second wavelength l2 and the first wavelength lΐ plus the full-width half-maximum of the waveform of the light emitted from the first light source 18 is greater than the full-width half-maximum of the first bandpass filter 56.

[0048] As also described above, block 910 may include emitting light from the first light source 18 and the second light source 22 as pulsed laser light. Similarly, for the example shown in Figure 3, block 910 may include emitting light from the third lights source as pulsed laser light.

[0049] With continued reference to Figure 9, block 920 includes filtering to the first bandwidth BW1, i.e., attenuating light outside the first bandwidth BW1, light that is emitted by the first light source 18 and reflected by a reflecting surface. Specifically, the light emitted by the first light source 18 and reflected at the first receiving unit 44 may be filtered with the first bandpass filter 56, as described above. In block 930, the method includes detecting light that is filtered to the first bandwidth BW 1. Specifically, this filtered light is detected by the first photodetector 20, as described above. Said differently, the first photodetector 20 may detect light in the first bandwidth BW1 transmitted by the first bandpass filter 56.

[0050] With continued reference to Figure 9, block 940 includes filtering to the second bandwidth BW2, i.e., attenuating light outside the second bandwidth BW2, light that is emitted by the second light source 22 and reflected by a reflecting surface. Specifically, the light emitted by the second light source 22 and reflected at the second receiving unit 46 may be filtered with the second bandpass filter 58, as described above. In block 950, the method includes detecting light that is filtered to the second bandwidth BW2. Specifically, this filtered light is detected by the second photodetector 24, as described above. Said differently, the second photodetector 24 may detect light in the second bandwidth BW2 transmitted by the second bandpass filter 58.

[0051] With continued reference to Figure 9, block 960 includes filtering to the third bandwidth light that is emitted by the third light source 26 and reflected by a reflecting surface. Specifically, the light emitted by the third light source 26 and reflected at the third receiving unit 48 may be filtered with the third bandpass filter 60, as described above. In block 970, the method includes detecting light that is filtered to the third bandwidth. Specifically, this filtered light is detected by the third photodetector 28, as described above. Said differently, the third photodetector 28 may detect light in the third bandwidth transmitted by the third bandpass filter 60.

[0052] In block 980, the method includes determining the location and distance of the object that reflected light back to the system 10. Block 980 may include eliminating or reducing“false signals” due to cross-talk, as described above. This may be accomplished, for example, by histogramming. Block 980 may be performed by the controller 16 or by another component of the vehicle 12, e.g., another component of the ADAS.

[0053] Figures 4 and 5 show two examples of the system 10 that distinguishes between the reflected light in the first field of view FV1 in the second field of view FV2. Specifically, the controller 16 controls the timing of emission of light and collection of light to distinguish between reflected light in the first field of view FV1 and in the second field of view FV2, i.e., is temporally based.

[0054] Figure 10 is a schematic showing the operation of the example shown in Figure 4 and Figure 11 is a schematic showing the operation of the example shown in Figure 5. With reference to Figures 4 and 5, the controller 16 is programmed to substantially simultaneously emit a pulse of light from the first light source 18 and the second light source 22. With reference to Figure 5, the controller 16 is programmed to emit a pulse of light from the third light source 26 substantially simultaneously with the first light source 18 and the second light source 22.

[0055] The controller 16 is programmed to, during a first time period, activate the first photodetector 20 and deactivate the second photodetector 24 and, during a second time period, activate the second photodetector 24 and deactivate the first photodetector 20. Specifically, the second time period initiates after the first time period and extends beyond the first time period. With reference to Figure 10, the first field of view FV1 is shorter than the second field of view FV2, as described above, i.e., the first photodetector 20 is short range and the second

photodetector 24 is long range). Accordingly, the time of flight of photons emitted from the second light source 22 to the second field of view FV2 will be greater than the first light source 18 to the first field of view F V 1. Thus, when the first photodetector 20 is activated and the second photodetector 24 is deactivated during the first time period, the first photodetector 20 detects light reflected in the first field of view FV1, including the light emitted by the first light source 18 (which is returning to the system 10 during the first time period). When the first photodetector 20 is deactivated and the second photodetector 24 is activated during the second time period, the second photodetector 24 detects the light reflected in the portion 62 (identified in Figs. 10 and 11) of the second field of view FV2 that extends beyond the first field of view FV1, i.e., the light emitted by the second light source 22. The activation/deactivation of the first and second photodetectors 20, 24 allows the second photodetector 24 to detect the light emitted by the second light source 22 (which is returning to the system 10 during the second time period) and not the first light source 18 (which has already returned to the system 10). Said differently, short-range detection occurs during the first time period and long-range detection occurs during the second time period. For the purposes of this disclosure, an“activated” photodetector detects light and outputs corresponding data and a“deactivated” photodetector does not detect light or output corresponding data, e.g., is unpowered.

[0056] The first time period may initiate simultaneously with emission of light from the first and second light sources 18, 22. The second time period initiates after the first time period and extends beyond the first time period. The first time period and the second time period overlap.

In such an example, the first time period may begin with the simultaneous emission of light from the first and second light source 18, 22, the second time period subsequently begins, the first time period subsequently ends, and the second time period subsequently ends. This timing reduces cross-talk, e.g., the first photodetector 20 detecting reflected light generated by the second light source 22, and the second photodetector 24 detecting reflected light generated by the first light source 18. Any remaining“false signals” from cross-talk data points may be removed by using histogramming. The light emitted from the first and second light sources 18,

22 may be the same or different wavelengths.

[0057] With reference to Figures 5 and 11, the controller 16 may be programmed to, during the first time period, activate the third photodetector 28. In such an example, the first photodetector 20 and the photodetector simultaneously detect reflected light. In the example shown in Figure 11, the first photodetector 20 and the third photodetector 28 may be aimed in different directions such that the first field of view FV1 and the third field of view FV3 do not overlap. The light emitted from the first and third light sources 26 may be the same or different wavelengths.

[0058] Figure 12 shows an example method 1200 of operation the system 10 in Figures 4 and 5. As shown in block 1210, the method includes emitting light from the first light source 18 and the second light source 22. Specifically, the method may include substantially

simultaneously emitting light from the first light source 18 and the second light source 22. In the example of Figure 5, the method may also include substantially simultaneously emitting light from the third light source 26. In other words, the controller 16 instructs the first light source 18 and the second light source 22 (and the third light source 26 in Figure 3) to emit light substantially simultaneously. Specifically, block 1210 may include emitting light from the first light source 18 into first field of view FV1 and simultaneously emitting light from the second light source 22 into the second field of view FV2 that overlaps the first field of view FV1.

[0059] In block 1220, a clock is started. For example, the controller 16 starts the clock and the first and second time periods are based on the clock. The controller 16 may start the clock at the simultaneous emission of light from the first and second light source 18, 22. The clock is used to determine the time of flight of reflected photons detected by the photodetectors 20, 24,

28 to determine distance of the object that reflected the light.

[0060] In block 1230, the method includes, during the first time period, activating the first photodetector 20 and deactivating the second photodetector 24. As described above, during the first time period, the first photodetector 20 is detecting photons reflected in the first field of view FV1 and not photons reflected in the portion 62 of the second field of view FV2 that extends beyond the first field of view FV1 because the reflected photons in the portion 62 of the second field of view FV2 do not return within the first time period. In other words, short-range detection occurs during the first time period.

[0061] Block 1230 may include, during the first time period, activating the third

photodetector 28. As described above, in such an example, the first photodetector 20 may detect photons reflected in the first field of view FV1 simultaneously with the detection of photons reflected in the third field of view FV3 by the third photodetector 28. In such an example, e.g., Figure 11, the first and third field of views FV1, FV3 are aimed in different directions.

[0062] In block 1240, the method includes, during the second time period, deactivating the first photodetector 20 (and deactivating the third photodetector 28 in examples including the third photodetector 28) and activating the second photodetector 24. As described above, during the second time period, the second photodetector 24 is detecting photons reflected in the portion 62 of the second field of view FV2 that extends beyond the first field of view FV1 because the reflected photons from the first field of view FV1 have returned before the second time period and the reflected photons from the portion 62 of the second field of view FV2 that extends beyond the first field of view FV1 return during the second time period. In other words, long- range detection occurs during the second time period.

[0063] In block 1250, the method includes determining the location and distance of the object that reflected light back to the system 10. Block 1250 may include eliminating or reducing“false signals” due to cross-talk, as described above. This may be accomplished, for example, by histogramming. Block 1250 may be performed by the controller 16 or by another component of the vehicle 12, e.g., another component of the ADAS.

[0064] The disclosure has been described in an illustrative manner, and it is to be understood that the terminology which has been used is intended to be in the nature of words of description rather than of limitation. Many modifications and variations of the present disclosure are possible in light of the above teachings, and the disclosure may be practiced otherwise than as specifically described.