Abstract:
A LIDAR system which reduces or suppress the frequency shift induced by the movement of objects in a scene relative to the LIDAR, and which comprises a light source, an input aperture (101), a splitter (2) configured to split a reflected light into a reference channel (4) and a first imaging channel (3), a first imaging optical IQ receiver (5) configured to obtain a first interference signal, a reference optical IQ receiver (6) configured to obtain a reference interference signal, an imaging oscillator (111), configured to be temporarily coherent with the reflected light, at least a mixer (12), connected to the first imaging optical IQ (5) and to the reference optical IQ (6) and configured to obtain a first intermodulation product with a higher frequency and an intermodulation product of interest with its Doppler Shift scaled.
Abstract:
Verfahren zur Dynamikerweiterung von Roh-Phasenbilder einer Lichtlaufzeitkamera bzw. Lichtlaufzeitkamerasystem, bei dem wenigstens zwei Tiefenbilder mit unterschiedlichen Belichtungszeiten aufgenommen werden.
Abstract:
An optical sensing system (100) for performing distance measurements. The optical sensing system comprises an emitter (110) configured to emit electromagnetic radiation modulated at a known frequency. The optical sensing system further comprises a detector (130) configured to sample incident electromagnetic radiation at the known frequency, convert the sampled electromagnetic radiation into charge carriers, and collect the charge carriers in a storage component (IG1,IG2;IG10,IG20,IG30,IG40) to produce an electronic signal. The optical sensing system further comprises a processor (150) configured to determine a correction by applying a non-linear polynomial function to the electronic signal.
Abstract:
Die Erfindung betrifft ein Verfahren zum Ermitteln einer eine Umgebung eines LiDAR-Sensors repräsentierenden Punktwolke bereitgestellt, umfassend die folgenden Schritte: Durchführen einer Laufzeitmessung mittels des LiDAR-Sensors, Ermitteln einer Punktwolke, welche eine Umgebung des LiDAR-Sensors repräsentiert, basierend auf der Laufzeitmessung, wobei die Punktwolke eine höhere Winkelauflösung für Objekte aufweist, welche sich mindestens in einer ersten Distanz zum LiDAR-Sensor befinden, als für Objekte, welche sich in einer Distanz zum LiDAR-Sensor befinden, welche kleiner oder kleiner-gleich als die erste Distanz ist. Die Erfindung betrifft weiter einen LiDAR-Sensor sowie ein Computerprogramm.
Abstract:
A method of operating an optical device, the method comprising tuning a first emission wavelength of a first output radiation of a laser device continuously within a first wavelength spectrum by modulating a drive current thereof with a first drive current modulation having a frequency of at least 100 kHz. The first wavelength spectrum comprises a first spectral feature associated with at least part of a gas absorption spectrum of at least one gas. The method comprises the steps of modulating the first output radiation of the laser device with a first output modulation, the first output modulation comprising a first plurality of binary pulses, scanning the first wavelength spectrum at a rate of at least 1 µm per second, projecting the first output radiation towards a first target area, receiving scattered radiation from the first target area, and processing the scattered radiation.
Abstract:
Techniques for de-aliasing depth ambiguities included within infrared phase depth images are described herein. An illuminator emits reference light towards a target object. Some of this light is reflected back and detected. A phase image is generated based on phase differences between the reference light and the reflected light. The phase differences represent changes in depth within overlapping sinusoidal periods of the reference and reflected light. The phase image also includes ambiguities because multiple different depths within the phase image share the same phase difference value, even though these depths actually correspond to different real-world depths. The phase image is fed as input to a machine learning (ML) component, which is configured to de-alias the ambiguities by determining, for each pixel in the phase image, a corresponding de-aliasing interval. A depth map is generated based on the phase image and any de-aliasing intervals generated by the ML component.
Abstract:
A time-of-flight (ToF) system (100) disclosed herein provides a method of separating a direct component of light (112) collected by a ToF detector from a global component of light (112a, 112b) collected by the ToF detector (120), the method comprising acquiring three or more images represented by three or more matrices in response to illuminating a target (150) with a light source using a first spatial pattern (132a, 132b, 132c) at three or more different modulation frequencies, acquiring an additional image represented by an additional matrix in response to illuminating the target with the light source using a second spatial pattern (134), the second spatial pattern being different than the first spatial pattern, and determining one or more parameters of the direct component of light (112) and the global component of light (112a, 112b) based on analysis of the three or more matrices and the additional matrix. The direct component is due to the direct illumination of a point in the scene by the source of light and the global component is due to the indirect illumination of the scene point. A ToF module (102) includes a computing module (104) that includes a processor (106) and a spatial pattern data store (108). The spatial patterns (132a, 132b, 132c and 134) are two dimensional patterns that map to the two-dimensional grid of light source elements (110a, 110b). Light source (110) may include a large number of light source pixels. The scattering of each of the direct light component (112) and the global light component (112a) at the point (154) results in a total light signal (114) that is captured by a camera (120) of the ToF module (102). The combination of imaging lens (116), light sampling array (124), and sampler (124) receives and converts the total light signal (114) into a sampled signal (126) that is communicated to a multipath mitigation module (128). The multipath mitigation module 128 includes an image analyzer (130) that analyzes images captured by the camera (120) to determine various parameters of a direct component and a global component of the total light signal (114. 44). A denoising module (140) generates a denoising coefficient ξ(m,n) to increase the signal to noise ratio (SNR) where m and n denote a size of a kernel matrix applied to each pixel location.
Abstract:
The present disclosure pertains to a device which has a circuitry that obtains image data of a scene being representative of a time of flight measurement of light reflected from the scene, wherein the image data is based on a pattern of light being illuminated on the scene, wherein the pattern of light includes high intensity light areas and low intensity light areas; obtains, based on the image data, first image data being representative of the high intensity light areas; obtains, based on the image data, second image data being representative of the low intensity light areas; estimates direct component image data based on the first image data and the second image data; and generates a depth map of the scene based on the direct component image data and the second image data.
Abstract:
A method for facilitating removal of specular reflection noise from light data can include illuminating, using an illumination unit, a target with a light source. The illumination unit is configured to project light with a spatial light pattern onto the target. The method can also include acquiring, with a sensor unit, light data that is reflected from the target. The light data may comprise a directly reflected spatial light pattern and a specular reflected spatial light pattern. The directly reflected spatial light pattern and the specular reflected spatial light pattern comprise at least one spatial distinction that distinguishes the directly reflected spatial light pattern from the specular reflected spatial light pattern. The method can further comprise processing the light data to distinguish the directly reflected spatial light pattern from the specular reflected spatial light pattern based upon the at least one spatial distinction.