Range-Enabled Three-Dimensional Imaging System and Associated Methods
Systems and methods for three-dimensional (3-D) imaging enabled by natural range-dependent processes. Multiple lasers are configured to independently flash illuminate a target object to 3- D image a resultant “scene” onto a focal plane array (FPA). The first laser produces a wavelength non-resonant with an atmospheric absorption line along the illumination path. The second laser produces a wavelength resonant with the atmospheric absorption line, and closely spaced with the non-resonant wavelength. A ratio of the respective intensities recorded at the FPA for the two wavelengths calculates a range to the target object.
The invention described herein may be manufactured, used, and licensed by or for the Government of the United States for all governmental purposes without the payment of any royalty.
BACKGROUNDLaser Detection and Ranging (LADAR), a special category of Light Detection and Ranging (LIDAR or Lidar), is an optical sensing technology that determines distance, velocity, and/or other characteristics of a distant target object by illuminating the target object with pulsed laser light and analyzing the reflected pulses as detected by some number of sensors. Differences in pulsed laser return times and wavelengths may be used to create digital three-dimensional (3-D) representations of the target object.
LADAR technology has terrestrial, airborne, and mobile applications. For example, LADAR technology may be used to create high-resolution maps of the surfaces of distant objects, with applications in geodesy, geomatics, archaeology, geography, geology, geomorphology, seismology, forestry, atmospheric physics, laser guidance, airborne laser swath mapping (ALSM), and laser altimetry. Application of LADAR to enable 3-D imaging for control and navigation of autonomous vehicles is currently a focus of significant research.
As commonly employed in known LADAR solutions, measurement of a signal received by a detector, such as a focal plane array (FPA), may be calculated as follows:
S=ƒ(R)A(R)Q(R)D p
where S denotes a measured signal, ƒ(R) denotes a function of range describing signal reduction due to scattering targets, A(R) denotes atmospheric absorption, Q(R) denotes a laser-target overlap integral, D denotes a detector collection aperture and includes the effect of system transmission and detector sensitivity, and p denotes target reflectivity. By way of definition, absorption is the process by which incident radiant energy is retained by a substance. As referred to herein, that substance is the atmosphere. When the atmosphere absorbs energy, the result is an irreversible transformation of radiation into another form of energy. The FPA is an image sensing device consisting of an array (typically rectangular) of light-sensing pixels at the focal plane of a lens. FPAs are used most commonly for imaging purposes (e.g., taking pictures or video imagery), but may also be used for non-imaging purposes such as spectrometry, lidar, and wave-front sensing.
Range to target is commonly determined by measuring a time between laser emission from a system, and from system measurement of laser light scattered from a target. Time of flight measurements typically require high bandwidth detectors and supporting electronics to precisely measure the time at which a signal was received by the detector. High accuracy measurements also typically require short laser pulses. Unfortunately, such high precision electronics result in high-cost designs to achieve working LADAR solutions.
The accompanying drawings provide visual representations which will be used to more fully describe various representative embodiments and can be used by those skilled in the art to better understand the representative embodiments disclosed and their inherent advantages. The drawings are not necessarily to scale, emphasis instead being placed upon illustrating the principles of the devices, systems, and methods described herein. In these drawings, like reference numerals may identify corresponding elements.
Specific embodiments of the disclosure will now be described in detail with reference to the accompanying figures. While this disclosure is susceptible of being embodied in many different forms, there is shown in the drawings and will herein be described in detail specific embodiments, with the understanding that the present disclosure is to be considered as an example of the principles of the described and not intended to limit the disclosure to the specific embodiments shown and described. In the description below, like reference numerals may be used to describe the same, similar or corresponding parts in the several views of the drawings.
The present disclosure relates generally to systems and associated methods of three-dimensional (3-D) imaging enabled by natural range-dependent processes. In certain embodiments, the present design enables the use of low-cost camera FPA instead of expensive high-speed detector arrays normally employed in conventional LADAR solutions. The technique may be advantageously employed and adapted in various solutions requiring 3-D imaging, such as machine vision for autonomous vehicles and potentially for gaming.
One embodiment of the present disclosure may advantageously exploit differential absorption in the atmosphere.
The atmosphere comprises elements and compounds that may include, inter alia, oxygen (O2), nitrogen (N2), carbon monoxide (CO), carbon dioxide (CO2), water (H2O), and other atmospheric gas or gases. These elements and compounds may provide a laser transmission medium for the lasers described herein. The above-mentioned group of elements and compounds may be selected from for analysis.
Another embodiment of the present disclosure may exploit a differential in measured intensity from a scattering object. Embodiments may leverage the natural range-dependent processes associated with the transmission of a laser beam to a target and the resulting measured signal recorded at a sensor receiver. These embodiments may each comprise simple, low cost 3-D imaging architectures that may be advantageously employed together or separately.
Referring now to
The FPA 135 need not be high bandwidth but instead may be a relatively low-bandwidth detector (e.g., low-bandwidth detector frame rates, such as those of a camera, may suffice). Bandwidth describes the width of a wavelength range with regard to a specific part of the spectrum which transmits incident energy via a filter. Bandwidth may be used to determine the resolution of spectrometers, spectral width of optical communication sources and duration of pulse waveforms. With respect to LADAR detectors, bandwidth typically refers to the frequency content of signals that such a detector is capable of faithfully reproducing. For example, given an optical signal characterized by a sinusoidal variation in time, the detector must have a bandwidth greater than the sine frequency in order to reproduce that sine wave. The higher the bandwidth, the greater the temporal or range resolution a LADAR system may have. In this example, a low bandwidth FPA means that the FPA pixels cannot faithfully reproduce the time variation of the incident optical signal. Instead, such a low bandwidth FPA sums all the incident energy and loses most of the temporal information. The temporal resolution of the FPA is ultimately limited by the frame rate which is typically tens to hundreds of frames per second. In certain embodiments of the present design, the relatively low-bandwidth is typically below 35 megabytes per second.
The lasers 115, 125 may be positioned substantially equidistant from a target object of interest 145 (also referred to herein as a scene) and configured to emit two very closely spaced wavelengths, such that one wavelength is resonant with an atmospheric absorption line defined between the lasers 115, 125 and the scene 145 and such that the other wavelength is just off the absorption line so that this latter wavelength does not suffer absorption. More specifically, the two wavelengths produced by the cooperating lasers 115, 125 may be considered “closely spaced” in that the two wavelengths are far apart enough such that their respective atmospheric transmissions are quite different but also such that their target reflectivities are quite similar.
By way of definition, atmospheric transmission describes how radiation propagates through an atmosphere, such as the earth's atmosphere. Bands, or ranges, of transmission or propagation (atmospheric transmission bands) are used to describe the radiation reflective and absorption properties. The atmospheric transmission band describes the capacity of the atmosphere to transmit electromagnetic energy. In the case of the earth's atmosphere, this atmospheric transmittance is dependent on the air mass penetrated by rays, as well as on the amount of water vapor and dust in the air. A high atmospheric transmission band has a higher rate of transmittance than a low atmospheric transmission band. The term high and low in this instance are relative to one another such that a relatively low atmospheric transmission band is less than a relatively high atmospheric transmission band.
In certain embodiments, the 3-D imaging system 100 of the present design may receive and use a first wavelength Aa that is in a relatively low atmospheric transmission band and a second wavelength Ab that is in a relatively high atmospheric transmission band. In other embodiments, the 3-D imaging system 100 of the present design may receive and use a first wavelength/:.a that is in a relatively high atmospheric transmission band and a second wavelength/:.b that is in a relatively low atmospheric transmission band.
The terms “relatively high . . . ” and “relatively low . . . ” mean that the higher value exceeds the lower value. Because these are relative terms, no specific values impact this disclosure. However, because these quantities are relative to each other, it is an embodiment of the present design that the first wavelength/:.a and the second wavelength/:.b may differ by approximately less than 0.015 micrometer (μm), while a respective reflectivity of the first wavelength/:.a and of the second wavelength/:.b may be approximately similar.
Indeed, the first wavelength/:.a and the second wavelength/:.b are of a relatively laser pulse type. This short laser pulse may be characterized by a pulse width (also referred to as pulse duration) that is less than 100 ms and typically less than 10 ms. In certain embodiments of the present design, for example, and without limitation, relevant laser pulse widths may be less than 10 ms while round trip laser transit times (e.g., from sensor to target and back to sensor) may depend on range to target and approach ˜6.7 ns/m.
For example, and without limitation, the first laser 115 may emit the non-resonant wavelength to flash illuminate the scene, or target, 145 such that the scene, or target, 145 may be imaged onto the FPA 135. The second laser 125 may emit the resonant wavelength to similarly flash illuminate the scene, or target, 145 such that the scene, or target, 145 may be imaged by the FPA 135. For the first laser 115, a first recorded intensity on each pixel of the FPA 135 may be a function of range to target and target reflectivity. For the second laser 125, a second recorded intensity may be a function of range, target reflectivity, and atmospheric absorption (which is also a function of range).
The ratio of the first and second recorded intensities is a function of the atmospheric absorption and, therefore, range. Knowing the atmospheric absorption at the two laser wavelengths, the system 100 may estimate the range to target from the ratio of the two measured intensities.
The ratio of two intensities the first intensity Sa and the second intensity Sb is defined as follows:
where α denotes an atmospheric absorption and scattering attenuation factor for the absorption line, and where R denotes the range.
The above equation is useful to measure, using the focal plane array, a first intensity Sa, of the first backscatter signal and a second intensity Sb of the second backscatter signal and determining, using the focal plane array, a range to the target from a ratio of the first intensity Sa, and the second intensity Sb.
Continuing to refer to
Absorption is the process by which incident radiant energy is retained by a substance. In this case, the substance is the atmosphere. When the atmosphere absorbs energy, the result is an irreversible transformation of radiation into another form of energy.
The ratio of S1i/S2i is given by
By measuring this ratio with a standard FPA 135 and knowing α at each wavelength, the system 100 may estimate, or determine, the R resolution dependence on (e.g., the signal ratio) in the following:
More specifically, the ratio above allows simplification, leaving a simple expression dependent on α and R (i.e., because the wavelengths are so close, the ratio cancels out the reflectivity dependence). The system 100 may determine α (λ1) and a (λ2) based on meteorological conditions, and then calculate the range R.
Referring now to
Operating the present design using an H2O absorption line may lead to complications as absorption may vary with humidity. A humidity measurement device and/or real-time National Oceanic and Atmospheric Administration (NOAA) meteorological humidity data may be employed to remove this variation. Alternatively, operating the present design using a more stable gas absorption line (e.g., oxygen (O2), nitrogen (N2), carbon monoxide (CO), carbon dioxide (CO2)) may alleviate the complication of absorption variability.
For example, and without limitation, the present design may operate along a 950 nanometer (nm) H2O absorption line. Graph 400 at
Graph 420 at
Also for example, and without limitation, consider the present design operating along a 763 nanometer (nm) O2 absorption line. Graph 500 at
Graph 510 at
Graph 520 at
Referring now to
Specular targets typically exhibit specular reflection, also known as regular reflection, which is the mirror-like reflection of waves, such as light, from a surface. The law of reflection states that for each incident ray the angle of incidence equals the angle of reflection, and the incident, normal, and reflected directions are coplanar. In contrast, a non-specular target exhibits non-specular wave reflection of light or sound from rough surfaces, resulting in scattering of wave components. Non-specular reflection is also known as diffuse reflection.
When a non-specular target 610 is illuminated by the light source, such as illuminator or solar illuminator, the incident light is scattered into a hemisphere and the signal has a change rate that changes with range according to the function ƒ(R)=R−2. Measurement of the scattered signal at two or more ranges may be used to estimate range and may also be used to estimate change rate. In the event the laser beam does not fully resolve, or overlap, the object the range dependent signal function will be ƒ(R)=R−4.
As illustrated in
Sj,i=ƒ(R)A(R)Q(R)Dp
where j=1,2 depending on the sensor referenced, and where i is the pixel.
Using a common wavelength in each sensor 615, 625, the ratio of measured intensity for a point in the scene S1,i, S2,i may be determined
The ratio of measured intensities S1,i, S2,i may be used to estimate range for each pixel in the scene.
In an alternative embodiment, the present design may substitute for the two sensors 615, 625 a single sensor that is moving. The single sensor may image a target scene from two positions along its vector (the displacement between these two positions serving as the distance 8 in the formulae above) and then may determine R as described above.
For example, and without limitation, consider the present design operating as shown in
Graph 800 at
Graph 810 at
Graph 820 at
Those skilled in the art will understand that the principles of the present disclosure may be implemented on or in data communication with any type of suitably arranged device or system configured to perform range analysis operations, in any combination. System 900 of the present design may comprise a processor 912 that may be operable to accept and execute computerized instructions, and also a data store 913 that may store data and instructions used by the processor 912. The processor 912 may be in data communication with external computing resources, such as the wide area network (WAN) 930, through a network interface 919. Furthermore, the processor 912 may be configured to direct input received from components of the WAN 930 to the data store 913 for storage. Similarly, the processor 912 may be configured to retrieve data from the data store 913 to be forwarded as output to various components of the WAN 930.
For example, and without limitation, the computerized instructions of the system 900 may be configured to implement absorption analysis subsystem 922 that may be stored in the data store 913 and retrieved by the processor 912 for execution. The absorption analysis subsystem 922 may be operable to determine range to a target object from differential absorption in the atmosphere as described above. Also, for example, and without limitation, the computerized instructions of the system 900 may be configured to implement a scatter analysis subsystem 924 that may be stored in the data store 913 and retrieved by the processor 912 for execution. The scatter analysis subsystem 924 may be operable to determine range to a target object from scattered signals intensities as described above. Also for example, and without limitation, the computerized instructions of the system 900 may be configured to implement Application Programming Interface (API) 926 that may be stored in the data store 913 and retrieved by the processor 912 for execution. The API subsystem 926 may be operable to forward computed range information for use by downstream applications.
Those skilled in the art will appreciate that the present disclosure contemplates the use of computer instructions and/or systems configurations that may perform any or all of the operations involved in 3-D imaging using LADAR. The disclosure of computer instructions that include absorption analysis subsystem 922 instructions, scatter analysis subsystem 924 instructions, and API subsystem 926 instructions is not meant to be limiting in any way. Those skilled in the art will readily appreciate that stored computer instructions and/or systems configurations may be configured in any way while still accomplishing the many goals, features and advantages according to the present disclosure.
It will be appreciated that the systems and methods described above are set forth by way of example and not of limitation. Absent an explicit indication to the contrary, the disclosed steps may be modified, supplemented, omitted, and/or re-ordered without departing from the scope of this disclosure. Numerous variations, additions, omissions, and other modifications will be apparent to one of ordinary skill in the art. In addition, the order or presentation of method steps in the description and drawings above is not intended to require this order of performing the recited steps unless a particular order is expressly required or otherwise clear from the context.
Embodiments disclosed herein may include computer program products comprising computer-executable code or computer-usable code that, when executing on one or more computing devices, performs any and/or all of the steps thereof. The code may be stored in a non-transitory fashion in a computer memory, which may be a memory from which the program executes (e.g., random-access memory associated with a processor), or a storage device such as a disk drive, flash memory or any other optical, electromagnetic, magnetic, infrared or other device or combination of devices. In another implementation, any of the systems and methods described above may be embodied in any suitable transmission or propagation medium carrying computer-executable code and/or any inputs or outputs from same.
Claims
1-20. (canceled)
21. A three-dimensional (3-D) imaging system for determining range to a target and comprising:
- a light source configured to illuminate a non-specular target with a wavelength A characterized by a reflectivity from the non-specular target;
- a first active sensor adapted to receive a first backscatter signal associated with the wavelength A;
- a second active sensor adapted to receive a second backscatter signal associated with the wavelength 1, the first active sensor adapted to measure a first intensity S1i of the first backscatter signal and the second active sensor adapted to measure a second intensity S2i of the second backscatter signal; and
- a focal plane array adapted to determine a range to the target from a ratio of the first intensity S1i and the second intensity S2i.
22. The 3-D imaging system according to claim 21, where the light source comprises one of a laser and a solar illuminator.
23. The 3-D imaging system according to claim 21, where the ratio of the first intensity S1,i and the second intensity S2,i is defined as
- S1,i/S2i=ƒ(R+o)/ƒ(R)=(R+o)2/R2
- where δ denotes a distance between the first active sensor and the second active sensor; and
- where R denotes the range.
24. The 3-D imaging system according to claim 21, where the first backscatter signal and the second backscatter signal are characterized by a change rate of f(R)=R−2.
25. A three-dimensional (3-D) imaging system for determining range to a target and comprising:
- a light source configured to illuminate a non-specular target with a wavelength A characterized by a reflectivity from the non-specular target;
- a movable active sensor adapted to receive a backscatter signal associated with the wavelength A;
- the first active sensor adapted to measure a first intensity S1,i of a first backscatter signal at a first position and to measure a second intensity S2i of a second backscatter signal at a second position; and
- a focal plane array adapted to determine a range to the target from a ratio of the first intensity S1i and the second intensity S2i.
26. The 3-D imaging system according to claim 25, where the light source comprises one of a laser and a solar illuminator and the ratio of the first intensity S1,i and the second intensity S2,i is defined as
- S1,i/S2,i=ƒ(R+o)/ƒ(R)=(R+o)2/R2
- where δ denotes a distance between the first active sensor and the second active sensor; and
- where R denotes the range.
27. The 3-D imaging system according to claim 25, where the first backscatter signal and the second backscatter signal are characterized by a change rate of ƒ(R)=R−2.
28. The 3-D imaging system according to claim 25 adapted to measure a first intensity Sa associated with the first backscatter signal and measure a second intensity Sb associated with the second backscatter signal; and
- determine, independent of reflectivity, a range to the target from a ratio of the first intensity Sa and the second intensity Sb, the ratio of the first intensity Sa and the second intensity Sb is defined as Sα/Sb=e−2R(α(λα)−α(λb)); where α denotes an atmospheric absorption and scattering attenuation factor for the absorption line; and where R denotes the range; where a respective reflectivity of the first wavelength λa and of the second wavelength λb are similar.
29. The 3-D imaging system according to claim 28, where the first wavelength λa is in a relatively low atmospheric transmission band and the second wavelength λb is in a relatively high atmospheric transmission band, whereby the second wavelength transmission band is greater than the first wavelength transmission band.
30. The 3-D imaging system according to claim 28, where the first wavelength λa is in a relatively high atmospheric transmission band and the second wavelength λb is in a relatively low atmospheric transmission band, whereby the first wavelength transmission band is greater than the second wavelength transmission band.
31. The 3-D imaging system according to claim 29, where the first wavelength λa and the second wavelength λb differ by less than 0.015 micrometer (μm).
32. The 3-D imaging system according to claim 30, where the first wavelength λa and the second wavelength λb differ by less than 0.015 micrometer (μm).
33. The 3-D imaging system according to claim 28, where the first wavelength λa and the second wavelength λb have a pulse duration less than 10 ms.
34. A three-dimensional (3-D) imaging system for determining range to a specular target and comprising:
- a light source configured to illuminate a specular target with a wavelength A characterized by a reflectivity from the specular target;
- a first active sensor adapted to receive a first regular reflection associated with the wavelength A;
- a second active sensor adapted to receive a second regular reflection associated with the wavelength 1, the first active sensor adapted to measure a first intensity S1i of the first regular reflection and the second active sensor adapted to measure a second intensity S2i of the second regular reflection; and
- a focal plane array adapted to determine a range to the target from a ratio of the first intensity S1i and the second intensity S2i.
35. The 3-D imaging system according to claim 34 where the ratio of the first intensity S1,i and the second intensity S2,i is defined as
- S1,i/S2,i=ƒ(R+o)/ƒ(R)=(R+o)2/R2
- 1 where δ denotes a distance between the first active sensor and the second active sensor; and
- where R denotes the range.
36. The 3-D imaging system according to claim 35, where the first regular reflection and the second regular reflection are characterized by a change rate of ƒ(R)=R−2.
37. The 3-D imaging system according to claim 35 where the ratio of the first intensity Sa and the second intensity Sb is defined as Sa Sb = e - 2 R ( α ( λ a ) - α ( λ b ) );
- where α denotes an atmospheric absorption and scattering attenuation factor for the absorption line; and
- where R denotes the range;
- where a respective reflectivity of the first wavelength λa and of the second wavelength λb are similar.
38. The 3-D imaging system according to claim 37, where the first wavelength λa and the second wavelength λb differ by less than 0.015 micrometer (μm).
39. The 3-D imaging system according to claim 38, where the first wavelength λa and the second wavelength λb have a pulse duration less than 10 ms.
40. The 3-D imaging system according to claim 25, where the first regular reflection and the second regular reflection are characterized by a change rate of ƒ(R)=R−2.
Type: Application
Filed: Oct 3, 2022
Publication Date: Feb 16, 2023
Inventor: Christian Keyser (Shalimar, FL)
Application Number: 17/937,647