Why Use SiPM Sensors for Automotive LiDAR Applications?
                    
                    
                    Information on our surrounding environment  can be obtained by sensing technologies to benefit our everyday lives:  either to improve task efficiency, our safety, or purely for  entertainment purposes. LiDAR  is one such sensing technique that uses laser light to measure the  distance to objects and can be used to create a 3D model of the  surrounding environment. Each pixel in the image captured by the LiDAR system  will have a depth associated with it. This allows for better object  identification and removes any ambiguity that might be present in a 2D  image obtained by an image sensor alone.
 
How does LiDAR build a 3D point cloud? LiDAR typically  uses the direct time-of-flight (dToF) technique to measure the distance  to an object. A short laser pulse is sent out and some of that light is  reflected back by objects in the scene and detected by a sensor, such  as the ArrayRDM-0112A20-QFN,  to accurately record the time it took for the round trip of the laser  pulse (see Fig. 1). Using the known speed of light, the distance can be  calculated from this dToF measurement. This gives a single distance  measurement within the field of view.
 
In order to build up a complete picture of  the surroundings, this point measurement needs to be repeated at many  different locations across the scene. This can be achieved by having  fixed sensors and lasers that rotate and scan throughout the scene, or  by using beam steering technology such as MEMS (micro-electro-mechanical  systems) mirrors.
 
Figure 1. Illustration of dToF technique.
 
LiDAR systems,  in general, rely upon the following key components: the illumination  source, the sensor, optics, beam steering, signal processing and power  management (Fig.2). For performance, the most critical elements are the  illumination source and the sensor. The illumination is typically  limited by eye safety considerations so that the greatest impact on  system performance is often down to the sensor.
 
Figure 2. Anatomy of a dToF LiDAR System including the sensor element.
 
In many scenarios, the system will need to  operate with the limited signal return, from distant or low-reflectivity  objects where the signal may consist of only a small number of photons.  Therefore, the sensor should be as sensitive as possible. The  sensitivity of a LiDAR sensor  is a combination of different factors. Firstly, the detection  efficiency, which is the probability that an incident photon will  produce a signal, is of prime importance. Then there is the sensitivity  to low incidence flux or minimum detectable signal. Some sensors, such  as PIN diodes have no internal gain and so a single detected photon will  not register above the inherent sensor noise. An avalanche photodiode  (APD) has some internal gain (~100x) but still, an incident signal  composed of a small number of photons does not register above the noise,  which requires it to integrate the returned signal for a certain  duration of time. Sensors that operate in the Geiger mode, such as SiPMs (silicon photomultipliers)  and SPADs (single-photon avalanche diodes), have internal gains of the  order of a million (1,000,000x), and so even a single photon generates a  signal that can be reliably detected above the internal sensor noise.  This allows one to set a low threshold to detect the faintest of return  signals.
 
While SiPMs and SPADs overcome many noise issues due to their high gain, in practical LiDAR  applications there is another source of noise that needs to be  considered – the ambient solar background or simply put sunlight. We are  often seeking to detect very faint LiDAR return  signals while being bombarded by unwanted light from the sun. So the  problem becomes one of maximizing the signal (the returned laser light)  while ignoring or minimizing the noise (sunlight). One way to do this is  to take advantage of the single-photon sensitivity of the sensor and  look for photons that are correlated in time.
 
This method of multi-shot dToF measurement  is achieved by repeating the procedure a number of times (multiple laser  pulses resulting in a dToF measurement for each). Instead of  calculating a distance for each measurement, each ToF value is added to a  histogram or distribution plot. The result is a plot that looks like  that shown in Fig. 3. The background counts are uncorrelated in  time – that is, they arrive randomly in time relative to the time the  pulse fired. These counts can be ignored since they are the noise due to  sunlight. The peak represents counts that are correlated in  time – a significant number of counts that all arrived with the same  time value, indicating a signal from a target. This peak value can be  translated to a distance for a particular frame and the process can  start again. Even with several dozens of laser cycles per pixel per  frame, frame rates of 30 fps can be achieved.
 
Figure 3. Example LiDAR ToF histogram
 
While a SiPM or  SPAD sensor can use its single-photon sensitivity in conjunction with  time-correlation techniques to see faint return signals, a PIN diode or  APD sensor would simply miss these counts due to being lost in the solar  background. Hence, these other sensor types simply cannot range as far  or as efficiently.
 
How is depth information being employed in the real-world and how can LiDAR help?  Consumer mobile applications have to date enabled many features via  image sensor technology alone, for example using structured light.  Time-of-flight (ToF) technology has been incorporated to some extent in  mobile phones for a few years now to add depth sensing and enable  photography features such as fast autofocus and “bokeh” portrait  effects. Most recently, dToF imaging LiDAR sensors have been incorporated in the latest consumer mobile devices,  which provide better depth information over previous techniques and  will no doubt greatly increase the number of mobile applications that  make use of this data. The 3D information can be used to enable 3D  mapping applications and improved augmented reality and virtual reality  (AR/VR) experiences.
 
In automotive  and industrial applications, where safety is key, the limitations of  image sensors alone for object identification and hence autonomous  decision-making and navigation, highlight the need for additional  information via a fusion of different sensing modalities. LiDAR can be used in conjunction with other sensing technologies, such as cameras,  ultrasonic and radar, to provide added redundancy, increasing the  confidence level of the decision-making algorithms responsible for  navigating or interacting with the environment. Each of these techniques  has unique characteristics providing varying levels of information,  with advantages and disadvantages in different situations.
 
Figure 4. Comparison of different sensor technologies
 
To enable high-performance LiDAR systems for automotive, a highly sensitive sensor, such as a SiPM, is the most efficient receiver. The SiPMs from ON Semiconductor  offer an unparalleled combination of performance and operational  parameters: high photon detection efficiency and low noise and dark  count rates combined with low operational voltage, temperature  sensitivity and process uniformity.
 
The ArrayRDM-0112A20-QFN, a 12-pixel linear array of SiPMs, addresses the market need for LiDAR.  It features industry-leading 18% photon detection efficiency at 905 nm,  the typical wavelength for a cost-effective, broad market LiDAR system. In addition, it is the first linear array of SiPMs commercially available and the very first automotive-qualified SiPM in the market. Learn more about the ArrayRDM-0112A20-QFN, or check out our Design Resources below!