2008.02411v3
2008.02411v3
Accurate 3D imaging is essential for machines to map and interact with the physical world[1, 2].
While numerous 3D imaging technologies exist, each addressing niche applications with varying
degrees of success, none have achieved the breadth of applicability and impact that digital image
sensors have achieved in the 2D imaging world[3–10]. A large-scale two-dimensional array of coherent
detector pixels operating as a light detection and ranging (LiDAR) system could serve as a universal
3D imaging platform. Such a system would offer high depth accuracy and immunity to interference
from sunlight, as well as the ability to directly measure the velocity of moving objects[11]. However,
due to difficulties in providing electrical and photonic connections to every pixel, previous systems
have been restricted to fewer than 20 pixels[12–15]. Here, we demonstrate the first large-scale
coherent detector array consisting of 512 (32 × 16) pixels, and its operation in a 3D imaging system.
Leveraging recent advances in the monolithic integration of photonic and electronic circuits, a dense
array of optical heterodyne detectors is combined with an integrated electronic readout architecture,
enabling straightforward scaling to arbitrarily large arrays. Meanwhile, two-axis solid-state beam
steering eliminates any tradeoff between field of view and range. Operating at the quantum noise
limit[16, 17], our system achieves an accuracy of 3.1 mm at a distance of 75 metres using only
4 mW of light, an order of magnitude more accurate than existing solid-state systems at such
ranges. Future reductions of pixel size using state-of-the-art components could yield resolutions
in excess of 20 megapixels for arrays the size of a consumer camera sensor. This result paves the
way for the development and proliferation of low cost, compact, and high performance 3D imaging
cameras, enabling new applications from robotics and autonomous navigation to augmented reality
and healthcare.
The digital complementary metal–oxide– mance, such as in mobile devices and augmented reality
semiconductor (CMOS) image sensor revolutionized systems. Optical phased arrays are a promising solid-
2D imaging, borrowing technology from silicon micro- state approach, but developing long-range 2D-scanning
electronics to produce a flexible and scalable camera systems has proven challenging, with current demonstra-
sensor[18]. As a focal plane array (FPA), the digital tions limited to less than 20 metres[21–23]. As such, no
image sensor uses a lens to focus light and form an current technology addresses the needs of these diverse
image on the detector. A key advantage of this scheme use cases.
is that the field of view and light collection efficiency are
not set by the image sensor, but by the choice of lens. Here, we demonstrate a fully solid-state, integrated
Furthermore, the CMOS image sensor can be optimized photonic LiDAR based on the same FPA concept as the
for high performance or low-cost. The flexibility of this CMOS image sensor. By efficient use of light, our sys-
arrangement made the digital CMOS sensor the sensor tem achieves the range, depth accuracy and field of view
of choice for most 2D imaging. needed by demanding applications such as self driving
vehicles[1] and drone- or land-based 3D mapping[24, 25].
In contrast, 3D imaging is characterized by an assort- The centerpiece of our system is the coherent receiver
ment of competing technologies, each addressing a nar- array, a focal plane array of compact optical heterodyne
row niche. Long range and high precision applications detectors operating at the quantum noise limit[16, 17].
such as autonomous vehicles and construction site map- To eliminate tradeoffs between field of view and range,
ping are dominated by expensive and fragile mechani- the receiver is paired with solid-state beam steering that
cally steered LiDAR systems[3, 4]. Solid-state solutions sequentially illuminates the scene in small patches. The
such as structured light[5] and time-of-flight arrays[6– coherent receiver array allows our architecture to operate
10, 19, 20] are used when affordability, compactness, and using a frequency-modulated continuous-wave (FMCW)
reliability must be achieved at the expense of perfor- coherent LiDAR scheme[26, 27]. In contrast to time-of-
2
flight LiDARs that use pulses of light, FMCW LiDAR This switching approach to beam steering is robust and
uses a linearly chirped laser. Scattered light received can be scaled up to arbitrarily large arrays, with optical
from the target is mixed with local oscillator light in a losses limited only by waveguide scattering[32] and the
heterodyne receiver, producing a beat frequency propor- extinction ratio of the switching trees. Meanwhile, the
tional to round-trip travel time and hence distance to the receiver consists of a dense FPA of miniaturized hetero-
target. dyne receivers. All receiver pixels that correspond to the
The FMCW scheme confers several advantages relative illuminated area are simultaneously read out in parallel.
to time-of-flight schemes. First, heterodyne detection is Since the angular resolution is defined by the point spread
immune to interference from sunlight and other nearby function of the lens, which drops off very quickly, there
LiDAR systems since it selectively detects light close in is negligible crosstalk between different receiver pixels.
frequency to the local oscillator light[11]. Second, co- To avoid wasting local oscillator light, a second switch-
herent LiDAR directly measures target velocity through ing tree provides only the activated subset of the receiver
Doppler shifts of the received light[26, 27]. Third, high FPA with local oscillator light.
depth accuracy, depending upon only chirp bandwidth Parallel readout in the receiver is fundamental to this
and signal-to-noise ratio[28], is achieved using relatively architecture’s scalability. First, resolution is defined by
low frequency receiver electronics. In contrast, depth ac- the number of pixels in the receiver FPA, rather than
curacy is limited by receiver bandwidth for time-of-flight the number of steering positions. This increases the
schemes. Finally, the FMCW system emits photons con- maximum resolution for a given chip size since hetero-
tinuously and thus is well suited for photonic integration, dyne receiver pixels are ∼ 10× smaller than thermo-optic
where non-linear effects constrain peak power[29, 30]. switches. Second, fast thermo-optic switching is unneces-
Conversely, time-of-flight schemes emit photons in short sary because the pixel rate is decoupled from the switch-
high-power bursts. ing rate. Finally, parallel readout proportionally reduces
Previous 3D imaging systems based on coherent re- the receiver signal frequencies for an FMCW scheme by
ceiver arrays have been limited to fewer than 20 pix- allowing longer ramp times, simplifying the readout elec-
els due to their reliance on direct electrical connections tronics.
to each pixel[12–15]. To address scalability, we imple-
mented our LiDAR system on a silicon photonics pro-
cess with monolithically integrated radio-frequency (RF) DESIGN OF HYBRID CMOS-PHOTONICS CHIP
CMOS electronics[31]. A highly multiplexed electronic
readout architecture is integrated into the receiver array, An optical micrograph of our demonstrator chip is
minimizing external electrical connections while main- shown in Fig. 1(c). The transmitter consists of a 1 × 16
taining signal integrity. Our 512-pixel prototype array thermo-optic switch tree with 16 grating couplers in
can thus be scaled to arbitrarily large numbers of pix- the transmit FPA. Meanwhile, the receiver consists of
els by increasing the size of the array. Furthermore, the a 32 × 16 (512) pixel array of heterodyne receivers, with
use of a standard commercial foundry process facilitates local oscillator light provided by a 1 × 8 switch tree. A
immediate mass production at minimal cost. subsection of the receiver is schematically illustrated in
Fig. 2(a). Each pixel collects scattered light from the
scene using a grating coupler. Local oscillator light is
SCALABLE 3D IMAGING ARCHITECTURE provided to each pixel via a network of silicon waveg-
uides. Scattered light and local oscillator (LO) are mixed
As shown in Fig. 1(a), our architecture is based on two on a balanced detector consisting of a 50-50 directional
FPAs. The first acts as a transmitter, and the second as coupler and germanium PIN photodiodes, producing a
a receiver. Chirped laser light for the FMCW scheme heterodyne tone corresponding to the target’s distance.
is generated externally by modulating a fixed-frequency The signal is then amplified by a transimpedance ampli-
1550 nm laser with a silicon-photonic IQ Mach-Zehnder fier (TIA) integrated within the pixel. A buffer amplifier
modulator (MZM), which is in turn driven by an arbi- at the end of each row transmits the signal to the edge
trary waveform generator. This approach ensures chirp of the chip, maintaining wide bandwidths while driving
linearity and enables the use of a simple, low-noise laser. the large parasitic capacitances of the wiring and mul-
Long-range performance is achieved by sequentially il- tiplexed circuitry. Finally, a set of 8 output amplifiers
luminating and reading out the scene in small patches. transmits the signal off-chip, enabling parallel readout.
By illuminating only the portion of the scene correspond- As shown in Fig. 2(b), the individual pixels are turned
ing to the pixels being read out, no illumination light on and off using a power switch built into each TIA.
is wasted. As illustrated in Fig. 1(b), the transmit- In general, minimizing the input-referred noise of the
ter consists of a switching tree terminated by a FPA of electronic signal chain improves the pixel’s sensitivity
grating couplers. Light is directed to one transmit grat- and detection probability. Furthermore, higher receiver
ing at a time, illuminating a small subset of the scene. bandwidths are desirable since this increases the max-
3
a IQ modulator
Power splitter Laser
MZM
Receiver chip
Digital
signal
processing
Beamsplitter Lens
Heterodyne
detector pixel
Transmit Local
light oscillator
1×M switch Focal plane array 1×N switch Focal plane array
Receiver block
c
FIG. 1. Solid-state 3D imaging architecture. (a) Our architecture consists of two focal plane arrays (FPAs): a transmitter
FPA that sequentially illuminates patches of the scene, and a receiver FPA that detects scattered light from the scene. The
frequency-modulated continuous-wave scheme is used for ranging. An optional microlens array can be used to shape the
illumination pattern to more closely match the receiver array, thereby improving system efficiency. (b) On-chip steering of
light is provided by thermo-optic switching trees on both the transmitter and receiver chips. (c) Optical micrograph of our
demonstrator chip, showing the switching trees and focal plane arrays for both the transmit and receive functionality.
imum range for a given integration time. The TIA gain and bandwidths above 280 MHz, as shown in Fig.
feedback resistance determines gain, bandwidth, and 3(a)-(b). As seen in Fig. 3(d), our integrated TIA has
noise, with bandwidth and noise decreasing with larger a similar gain-bandwidth product and a 2 − 3× lower
resistance[33]. Compact waveguide-coupled photodiodes noise floor than conventional systems with photodiodes
and tight integration between the photodiodes and TIAs and amplifiers on separate chips.
yields a small parasitic capacitance of 1.5 fF. We thus The low noise on-chip amplifier chain allows the re-
achieve low noise in the electrical signal chain for 20 kΩ ceiver FPA to operate at the quantum limit for sensitiv-
4
Power, Transimpedance
ground amplifier Power + ground
vias RF output wires
via
LO light from
switching tree
50/50
coupler
Photodiode
Grating
coupler
Receiver pixel
Si (undoped)
Si (n-doped)
Si (p-doped) Electronic amplifiers
Row RF
Ge Metal wires outputs
Row amplifier
Power bus
Block and
Grating column enable
coupler
Pixel signal
Directional output bus
couplers
Ground bus
LO bus
Ge Transimpedance Inter-
Optical
Balanced amplifier stage
Detector filter Electrical 50 μm
FIG. 2. Receiver focal plane array design. (a) Schematic of a receiver block in our receiver focal plane array. Within the
receiver block, local oscillator (LO) light is distributed to a dense array of heterodyne detector pixels via a network of silicon
waveguides. Meanwhile, each pixel collects scattered light from the scene using a grating coupler, which is combined with LO
light on a balanced detector to produce a detectable photocurrent. The photocurrent is amplified in two stages: first by a
transimpedance amplifier within the pixel, and again by an amplifier at the end of each row. For clarity, we have omitted
control wires from the diagram. (b) Electrical schematic of the heterodyne detector pixel. (c) Optical micrograph of a small
subset of the receiver focal plane array.
ity, which is reached when local oscillator shot noise dom- the size of the LO distribution tree since many receiver
inates all other noise sources[16, 17]. As shown in Fig. pixels can simultaneously share LO power.
3(e), shot noise reaches parity with amplifier noise with Monolithic integration of electronics into the receiver
only 5 µW of LO power for a typical receiver pixel. In FPA facilitates the use of an actively multiplexed read-
contrast, coherent receivers for telecommunications typ- out architecture, allowing the receiver to be scaled to
ically require ∼ 10 − 100× more LO power. Combined arbitrarily large numbers of pixels. In our demonstrator
with the excellent 30 − 40 dB common-mode rejection chip, multiple levels of multiplexing and amplification are
ratio of the balanced heterodyne detectors, as shown in used to map 512 pixels to 8 outputs while maintaining
Fig. 3(c), this makes the receiver array significantly less signal integrity. As illustrated in Extended Data Fig.
susceptible to LO noise sources such as laser relative in- 2(a-c), the pixels are read out in blocks of 8 at a time.
tensity noise, optical amplifier noise, and chirp generator The output analog signals are fed into a bank of off-chip
noise. Furthermore, the low LO power per pixel reduces analog-to-digital converters for digitization, followed by
5
a b 120 c
f-3dB = 280 MHz 120
Occurrences
Occurrences
80 80
−5
60 60
−10 40 40
20 20
−15 0 0
1 2 3
10 10 10 0.8 1.0 1.2 1.4 20 30 40 50
Frequency (MHz) Noise floor (pA/√Hz) CMRR (dB)
d 2 e 5
10
4
Noise floor (pA/√Hz)
2
1 Previous work
10
1 Amplifier
ise
noise
no
ot
0
This work
Sh
10
0.5
12 13 14 15 −2 −1 0 1
10 10 10 10 10 10 10 10
Gain-bandwidth product (Ω-Hz) LO power (μW)
FIG. 3. Receiver electro-optic performance. (a) Measured frequency response of the receiver readout chain for an optical signal
supplied to a single pixel, showing a cutoff frequency of 280 MHz. (b, c) Histograms of input-referred amplifier noise and
common-mode rejection ratio (CMRR) respectively throughout the full array, showing tight distributions for both parameters.
(d) Largely due to tight integration between our photodiodes and transimpedance amplifiers, we have achieved a high gain-
bandwidth product with significantly improved noise performance compared to previous designs. (e) Input-referred noise as a
function of optical local oscillator (LO) power for a single pixel, demonstrating shot-noise limited detection using < 10 µW of
LO power.
a Frequency (MHz) b c
0.40 0.45 0.50 0.55 Rotating basketball at 17 m (1 rpm)
3000 200
3 Static object 18 m target (85% reflectance) Pixel at left edge of ball
10
Occurrences
Occurrences
150
Up-chirp 2000
Photons
2
10
1 Down-chirp 100
10 1000
0 50
10
0 0
3 Moving object 75 m target (30% reflectance) Pixel at right edge of ball
10 400
Occurrences
Occurrences
150
Photons
2
10 300
1 100
10 200
0
100 50
10
0 0
15 16 17 18 19 −10 −5 0 5 10 −15 −10 −5 0 5 10 15
Distance (m) Depth error (mm) Velocity (mm/s)
d e Rotating basketball at 17 m
10
Velocity (mm/s)
0.2
0
0.1
-10
y (m)
0.0
−0.1
1 rpm
−0.2
17.8
f 10
17.7
Velocity (mm/s)
17.6 5
17.5 0
)
(m
17.4
z
−5
−0.2 17.3
−0.1
0.0 17.2 −10
x (m) 0.1 −15 −10 −5 0 5 10 15
0.2
x (cm)
g h
Boxes at 54 m Exterior wall at 75 m
0.4
0.6
0.2 0.3
y (m)
y (m)
0.0 0.0
−0.2 −0.3
76.0
−0.6 −0.4 54.8 −1.0
−0.2 0.0 −0.5 75.5
54.6 0.0
)
(m
x (m)
(m
0.6 1.0
z
FIG. 4. 3D imaging system characterization. (a) Representative signals from a receiver pixel, showing Doppler splitting between
the up- and down-chirps for the moving target. (b) Depth noise for targets at 18 m and 75 m, with standard deviations of
1.8 mm and 3.1 mm respectively. (c) Velocity histograms for a basketball rotating at 1 rpm, exhibiting a standard deviation of
1.0 mm/s. (d) Velocity annotated point cloud of a basketball at 17 m rotating about its vertical axis at 1 rpm. (e) Photograph
of the basketball setup. (f) Horizontal linecut of velocity across the middle of the basketball. (g, h) Point clouds of (g) stacked
cardboard boxes at 54 m, and (h) an exterior wall at 75 m. Distance to the target is indicated by colour in (g) and (h). The
missing band of points in the middle of the point clouds is due to a narrow gap in the receiver array for electrical and optical
routing.
7
C. Geyer, M. Gittleman, S. Harbaugh, M. Hebert, T. M. Lasers and Electro-Optics (Optical Society of America,
Howard, S. Kolski, A. Kelly, M. Likhachev, M. Mc- 2018) p. JTh5C.2.
Naughton, N. Miller, K. Peterson, B. Pilnick, R. Ra- [23] C. V. Poulton, M. J. Byrd, P. Russo, E. Timurdogan,
jkumar, P. Rybski, B. Salesky, Y.-W. Seo, S. Singh, M. Khandaker, D. Vermeulen, and M. R. Watts, IEEE
J. Snider, A. Stentz, W. R. Whittaker, Z. Wolkowicki, Journal of Selected Topics in Quantum Electronics 25, 1
J. Ziglar, H. Bae, T. Brown, D. Demitrish, B. Litk- (2019).
ouhi, J. Nickolaou, V. Sadekar, W. Zhang, J. Struble, [24] J. Wang, W. Sun, W. Shou, X. Wang, C. Wu, H.-Y.
M. Taylor, M. Darms, and D. Ferguson, Journal of Field Chong, Y. Liu, and C. Sun, Journal of Intelligent &
Robotics 25, 425 (2008). Robotic Systems 79, 417 (2015).
[2] Q. Wang and M.-K. Kim, Advanced Engineering Infor- [25] A. Kasturi, V. Milanovic, B. H. Atwood, and J. Yang,
matics 39, 306 (2019). in Laser Radar Technology and Applications XXI , Vol.
[3] D. D. Lichti, ISPRS Journal of Photogrammetry and Re- 9832, edited by M. D. Turner and G. W. Kamerman,
mote Sensing 61, 307 (2007). International Society for Optics and Photonics (SPIE,
[4] J. Kidd, Performance Evaluation of the Velodyne VLP- 2016) pp. 206 – 215.
16 System for Surface Feature Surveying, Master’s thesis, [26] H. D. Griffiths, Electronics Communication Engineering
University of New Hampshire (2017). Journal 2, 185 (1990).
[5] J. Salvi, J. Pagès, and J. Batlle, Pattern Recognition 37, [27] J. Riemensberger, A. Lukashchuk, M. Karpov, W. Weng,
827 (2004), agent Based Computer Vision. E. Lucas, J. Liu, and T. J. Kippenberg, Nature 581, 164
[6] A. Corti, S. Giancola, G. Mainetti, and R. Sala, Robotics (2020).
and Autonomous Systems 75, 584 (2016). [28] K. Thurn, R. Ebelt, and M. Vossiek, in 2013 IEEE MTT-
[7] P. McManamon, Optical Engineering 51, 1 (2012). S International Microwave Symposium Digest (MTT)
[8] P. F. McManamon, P. S. Banks, J. D. Beck, D. G. Fried, (2013) pp. 1–3.
A. S. Huntington, and E. A. Watson, Optical Engineer- [29] H. K. Tsang, C. S. Wong, T. K. Liang, I. E. Day, S. W.
ing 56, 1 (2017). Roberts, A. Harpin, J. Drake, and M. Asghari, Applied
[9] S. W. Hutchings, N. Johnston, I. Gyongy, T. Al Abbas, Physics Letters 80, 416 (2002).
N. A. W. Dutton, M. Tyler, S. Chan, J. Leach, and [30] H. Rong, A. Liu, R. Jones, O. Cohen, D. Hak, R. Nico-
R. K. Henderson, IEEE Journal of Solid-State Circuits laescu, A. Fang, and M. Paniccia, Nature 433, 292
54, 2947 (2019). (2005).
[10] A. Ronchini Ximenes, P. Padmanabhan, M. Lee, Y. Ya- [31] K. Giewont, K. Nummy, F. A. Anderson, J. Ayala,
mashita, D. Yaung, and E. Charbon, IEEE Journal of T. Barwicz, Y. Bian, K. K. Dezfulian, D. M. Gill,
Solid-State Circuits 54, 3203 (2019). T. Houghton, S. Hu, B. Peng, M. Rakowski, S. Rauch,
[11] B. Behroozpour, P. A. M. Sandborn, M. C. Wu, and J. C. Rosenberg, A. Sahin, I. Stobert, and A. Stricker,
B. E. Boser, IEEE Communications Magazine 55, 135 IEEE Journal of Selected Topics in Quantum Electronics
(2017). 25, 1 (2019).
[12] F. Aflatouni, B. Abiri, A. Rekhi, and A. Hajimiri, Opt. [32] Y. Shen, N. C. Harris, S. Skirlo, M. Prabhu, T. Baehr-
Express 23, 5117 (2015). Jones, M. Hochberg, X. Sun, S. Zhao, H. Larochelle,
[13] A. Martin, D. Dodane, L. Leviandier, D. Dolfi, D. Englund, and M. Soljačić, Nature Photonics 11, 441
A. Naughton, P. O’Brien, T. Spuessens, R. Baets, G. Lep- (2017).
age, P. Verheyen, P. De Heyn, P. Absil, P. Feneyrou, and [33] B. Razavi, Solid-State Circuits Magazine 11, 10 (2019).
J. Bourderionnet, Journal of Lightwave Technology 36, [34] J. Y. Wang, Appl. Opt. 21, 464 (1982).
4640 (2018). [35] D. Rebolj, Z. Pučko, N. Čuš. Babič, M. Bizjak, and
[14] D. Inoue, T. Ichikawa, A. Kawasaki, and T. Yamashita, D. Mongus, Automation in Construction 84, 323 (2017).
Opt. Express 27, 2499 (2019). [36] A. Samani, E. El-Fiky, M. Morsy-Osman, R. Li, D. Patel,
[15] C. Li, X. Cao, K. Wu, X. Li, and J. Chen, Opt. Express T. Hoang, M. Jacques, M. Chagnon, N. Abadı́a, and
27, 32970 (2019). D. V. Plant, Journal of Lightwave Technology 37, 2989
[16] M. Collett, R. Loudon, and C. Gardiner, Journal of Mod- (2019).
ern Optics 34, 881 (1987). [37] A. Stove, IEE Proceedings F (Radar and Signal Process-
[17] M. A. Rubin and S. Kaushik, Opt. Lett. 32, 1369 (2007). ing) 139, 343 (1992).
[18] A. El Gamal and H. Eltoukhy, IEEE Circuits and Devices [38] V. Winkler, in 2007 European Radar Conference (2007)
Magazine 21, 6 (2005). pp. 166–169.
[19] B. L. Stann, K. Aliberti, D. Carothers, J. Dammann, [39] H.-S. Chen and C. R. N. Rao, Journal of Physics D: Ap-
G. Dang, M. M. Giza, W. B. Lawler, B. C. Redman, plied Physics 1, 1191 (1968).
and D. R. Simon, in Laser Radar Technology and Appli- [40] Z. Sheng, Z. Wang, C. Qiu, L. Li, A. Pang, A. Wu,
cations IX , Vol. 5412, edited by G. W. Kamerman and X. Wang, S. Zou, and F. Gan, IEEE Photonics Jour-
G. W. Kamerman, International Society for Optics and nal 4, 2272 (2012).
Photonics (SPIE, 2004) pp. 264 – 272. [41] N. C. Harris, Y. Ma, J. Mower, T. Baehr-Jones, D. En-
[20] K. Hu, Y. Zhao, M. Ye, J. Gao, G. Zhao, and G. Zhou, glund, M. Hochberg, and C. Galland, Opt. Express 22,
IEEE Sensors Journal 17, 5547 (2017). 10487 (2014).
[21] C. V. Poulton, A. Yaacobi, D. B. Cole, M. J. Byrd, [42] M. Mendez-Astudillo, M. Okamoto, Y. Ito, and T. Kita,
M. Raval, D. Vermeulen, and M. R. Watts, Opt. Lett. Opt. Express 27, 899 (2019).
42, 4091 (2017). [43] M. G. Ahmed, T. N. Huynh, C. Williams, Y. Wang, P. K.
[22] S. A. Miller, C. T. Phare, Y.-C. Chang, X. Ji, O. A. J. Hanumolu, and A. Rylyakov, IEEE Journal of Solid-
Gordillo, A. Mohanty, S. P. Roberts, M. C. Shin, State Circuits 54, 834 (2019).
B. Stern, M. Zadka, and M. Lipson, in Conference on
9
Methods
Design and fabrication. velocity measurements were performed using a series of
fifty 17 µs up-chirps, followed by fifty 17 µs down-chirps,
The demonstration chips used as transmitter and as shown in Extended Data Figure 5(d). The multiple
receiver FPAs were fabricated using GlobalFoundries’ chirps were coherently combined using a two-dimensional
CMS90WG 300 mm silicon photonics process[31], which FFT to extract the beat frequencies, maintaining the
monolithically integrates photonic devices with 90 nm same signal-to-noise ratio as the single-chirp case [37, 38].
silicon-on-insulator (SOI) RF CMOS electronics. All
photonic devices used in the design, with the exception of
the directional couplers, were provided in the foundry’s
Optical setup.
standard process development kit (PDK). By doing so
the photonic architecture had correct-by-construction de-
vice placement and connectivity, verifiable using Mentor A narrow-linewidth (< 100 Hz) fiber laser (NKT Ad-
Graphics’ Calibre Design Rule Checker. The integrated justik) operating at 1550 nm was used as the seed laser for
electronics followed a standard design flow using Cadence the FMCW ranging system. A linear chirp was applied
Virtuoso and Spectre for circuit design and layout, and to the laser light using a silicon photonic IQ modula-
Mentor Graphics’ Calibre for verification of design rules, tor fabricated at the University of Southampton, which
comparing layout-versus-schematic, and extracting par- was driven by a microwave chirp produced by an arbi-
asitics. The two domains are merged into a single hi- trary waveform generator (Tektronix AWG70002A). The
erarchy enabling connectivity verification at the receiver chirped laser light was amplified by erbium-doped fiber
photodiodes along with design rule verification of closely amplifiers (EDFAs) in two stages (a Keopsys CEFA-C-
intertwined photonics and electronics across the chip. HG-PM followed by an NKT Boostik). The amplified
Due to the limited number of dies available from the light was then coupled on-chip via single-mode optical
multi-project wafer shuttle, we were able to fully test the fiber V-grooves into two identical demonstration chips,
functionality of 5 dies. We did not observe any defects used as a transmitter and receiver respectively. The light
such as dead pixels or inoperative thermo-optic switches emitted by the transmitter FPA was structured using a
across these dies. 32x16 microlens array (PowerPhotonic), which was pre-
cisely matched to the receiver array’s pixel pattern. This
created a structured illumination pattern and minimized
Optical chirp scheme. any waste of light due to transmit optical power being
incident on the gaps between pixels.
A linearly chirped optical field E(t) has the form To precisely match the fields of view of the transmitter
and receiver FPAs, we took advantage of the fact that
E(t) = exp i2πf0 t + iπrt2
each receiver grating coupler emits a small amount of LO
= cos πrt2 + i sin πrt2 exp(i2πf0 t),
(1) light due to backreflections from the balanced detectors.
where f0 is the carrier frequency, and r is the chirp ramp An infrared camera was then used to align the patterns
rate. Thus, by coherently modulating fixed-frequency of spots produced by the receiver and transmitter FPAs.
light with a microwave chirp of the form cos πrt2 + The physical aperture of the output lens is 25 mm in
i sin πrt2 , we produce a linear chirp in the optical do- diameter. The size of the mode corresponding to a single
main. grating coupler on the receiver focal plane array at this
In our demonstrator system, we use a series of lin- same lens (which can be thought of as the entrance pupil
ear up-chirps immediately followed by a series of down- of our system) is slightly elliptical and is 11 mm along the
chirps. The mean and difference of the up-chirp and first axis and 16 mm along the second axis. For the long
down-chirp beat frequencies allow separate measurement range 75 m measurements, the output lens was adjusted
of range and velocity of a target respectively[11, 26, 27]. such that the transmit and receive beams were essentially
For low-velocity measurements, which includes all the collimated to infinity. For shorter range measurements,
measurements presented here except those in Extended the lens position was adjusted to simulate the effects of
Data Figure 5(f), we use a single up-chirp and single a smaller aperture.
down-chirp, each with a length of 850 µs. This chirp Due to the use of coherent detection, the receiver de-
scheme is illustrated in Extended Data Figure 5(c). The tects a single polarization of scattered light. In coher-
beat frequencies for each chirp were then extracted us- ent LiDAR systems, the receiver polarization is typically
ing fast-Fourier transforms (FFTs). For fast moving ob- chosen to be the same as the transmitter polarization[11,
jects with sufficiently large Doppler shifts, however, the 26, 27] as is done here, since most materials preferentially
beat frequency wraps around zero, resulting in ambigu- scatter light into the same polarization as the illuminat-
ous measurements. To compensate for this effect, high ing light[39].
11
thermo-optic phase shifters, and can be resolved by using differences in on-chip optical waveguide lengths, in ad-
existing designs for compact thermo-optic shifters [42]. dition to subtly differing paths taken through the free
Correcting for the 4× mismatch between the number of space optics by light from different pixels. Since these
transmitter and receiver positions, our equivalent point path length differences are static, they can be eliminated
rate is 2 × 104 per second. Combined with a 4 mW emit- using straightforward calibration measurements.
ter power, this yields an optical efficiency of 0.2 µJ/point. Thus, the key parameter for our system is depth noise,
The combination of high optical efficiency and long range the variation in depth measurements due to stochastic
necessary for autonomous navigation applications[1] is noise in our system. Depth noise was measured by ac-
typically met using mechanically steered LiDARs, such quiring 40 sequential frames of a static test target for the
as the commonly used Velodyne VLP-16. This 100 m histograms shown in Fig. 4(b), and 20 sequential frames
class mechanical LiDAR uses the same 0.2 µJ of light for the depth precision measurements in Extended Data
per point as our system[4], and has a much poorer depth Fig. 5(b). The mean distance value for each pixel was
accuracy of 3 cm. taken to be the true distance, and depth error was de-
fined as the deviation from the true distance for each
pixel. Finally, we defined measurement precision as the
Electro-optic characterization. standard deviation of the depth error.
Measurement error in our 3D imaging system can be Fig. 3(d) plots input-referred noise current density
divided into two categories: systemic errors due to non- against the transimpedance gain-bandwidth product for
idealities in our system, and random fluctuations in the several state-of-the-art CMOS and BiCMOS optical re-
measured beat frequencies due to shot noise, laser rela- ceiver publications[43–47]. A custom design must simul-
tive intensity noise, laser frequency fluctuations, and elec- taneously meet requirements for gain, noise, and band-
tronic noise sources. Systemic errors in our system are width. Generally, the gain-bandwidth product will be
very tightly controlled. Since the frequency chirps in our constant for a target technology and power consumption.
system are generated using direct digital synthesis in an In a resistive shunt-feedback configuration, the input-
AWG, distance accuracy is fundamentally derived from referred current noise in,rms is typically dominated by
the speed of light, a fixed physical constant, and the tim- the feedback resistance RF :
ing accuracy of the clocks in the AWG and ADCs, which r
4kT
are controlled to within a few parts per million. The in,rms = · BW−3dB (4)
RF
only remaining source of systemic error comes from op-
tical path length differences between pixels, which mani- where k is the Boltzmann constant, T is the tempera-
fest as static offsets in measured depth. These are due to ture, and BW−3dB is the 3-dB bandwidth. The gain is
13
approximately equal to the feedback resistance, and the Pixel Size Reduction.
bandwidth is determined by the pole at the input, where
CT is the total capacitance at the TIA input, and A0 is When considering the maximum resolution sensor that
the open-loop gain of the TIA. can fit on a chip with a given area, the pixel size is the
dominant factor. The switching trees are negligible in
size compared to the receiver FPA if existing compact
1 + A0 designs for thermo-optic phase shifters are used. For ex-
BW−3dB = (5) ample, efficient thermo-optic phase shifters as short as
2π · RF · CT
35 µm in length have been demonstrated [42].
The current pixel size is limited by the use of foundry
PDK devices, which were not designed to minimize foot-
The negative feedback acts to reduce the input print. In terms of individual photonic components, effi-
impedance looking into the TIA. Due to our low band- cient grating couplers[48] with a footprint of 3 × 3 µm2 ,
width requirement (< 1 GHz), and small diode and para- and 2 × 2 couplers[49] as small as 3 × 1 µm2 have been
sitic capacitance at the TIA input, we can use a large re- demonstrated. Meanwhile, photodiodes with a footprint
sistor to get high TIA gain resulting in a reasonable gain- of 3 × 1 µm2 are feasible due to the short absorption
bandwidth product while allowing a low input-referred length of germanium[50]. Lastly, using advanced deep
noise density. Having low-noise electronics improves the sub-micron CMOS technology nodes, the TIA could be
system’s detection probability, providing longer range for reduced to 3 × 3 µm2 in size. This should allow the com-
a given optical power. ponents for a single pixel to fit in an 8 × 5 µm2 footprint.
14
a b c
Input
Heater 1
Output 1
Output 2
Heater 2
Heater 1
Heater 2
1 2
Outputs
Extended Data Figure 1. Thermo-optic switching tree demonstration. (a) The thermo-optic switches consist of a Mach-Zehnder
interferometer with an electrical heater on each arm. (b) Tuning curve for a single thermo-optic switch, showing optical power
in the two outputs as a function of applied heater power. The use of two heaters allows the average electrical power consumption
per switch to be halved. (c) Output power distribution of the 1×16 transmitter switch tree for all switch settings, demonstrating
clean switching. Output power was monitored using a set of monitor photodiodes at the output of the switch tree. (d) On
and off transients for a representative thermo-optic switch, demonstrating 90% − 10% switching times of 9.1 µs and 12.1 µs
respectively. Due to minor thermal crosstalk between switches, the switching transients are not perfect decaying exponentials.
15
a b
Output driver amplifiers
TX Gratings
Transmitter Ch0 Ch1 Ch2 Ch3 Ch4 Receiver
(+ – )(+ – ) (+ – )(+ – ) (+ – )
...
8 Output Drivers
0 8
RX Block 0 RX Block 4
1 9 3 4
column 0 1 2 5 6 7 row
Row Buffers
1×16 2 10 0 RX Block 1 RX Block 5
TX Switch 1
3 11
2 RX Block 2 RX Block 6
4 12 3
RX Block 0
5 13 4 RX Block 3 RX Block 7
5
6 14 6
1×8
7 15 7 LO Switch
c d
Pixel Array Buffers Drivers
RX Block 0, Row 0 Channel 0 TX Optical Steering TX Grating 0 TX Grating 1 TX Grating 2
RX Column Select 0 1 7 0 1 7 0
RX Block 0, Row 7 Channel 7
Acquire Pixel Column
RX Block 7, Row 7 Time
Readout 8 pixels
Column 0 1 7 Chip
Optical
Boundary
Active circuits during readout example Electrical
Extended Data Figure 2. Transmitter (TX) and receiver (RX) synchronization and readout architecture. (a) The TX steers
light through a 4 level tree of 1 × 2 switches to feed the focal plane array of 16 output grating couplers. Each leaf contains a
fractional tap and monitor photodiode enabling electronic calibration of the tree. (b) The RX array is divided into 8 blocks of 64
pixels. Imaging an 8-pixel column requires both steering the local oscillator (LO) light to the block and enabling the associated
electronics (pixel column and row buffer amplifiers). Signals from the active pixel column are driven by 8 output amplifiers
for parallel readout. (c) Several levels of multiplexing are used to map 512 pixels down to 8 output channels. An active RX
block has one active pixel per row, with the other disabled pixels within the row presenting high output impedance (no drive
strength). The row buffers are similarly passively multiplexed between the blocks. The 8 drivers are always activated during
readout. (d) Timing diagram showing synchronization between the optical switching trees (TX and RX) and the electrical
readout circuitry.
16
Receiver
chip
Objective
lens
Lens
Polarizing
beamsplitter
Lens
To target
Extended Data Figure 3. Free-space optics schematic of the demonstration system. Much of the complexity in the optical
system is to match the receiver and transmitter focal plane arrays, which can be corrected in the future by adjusting the chip
layouts. For inexpensive consumer versions of the system, the Faraday rotator and polarizing beamsplitter could be replaced
by a 50-50 beamsplitter, at the cost of a 4× reduction in signal strength. Although it is possible to implement this experiment
using a single chip for both transmit and receive functions, we have used two identical chips acting as the transmitter and
receiver respectively to simplify the experimental setup.
17
a b
Extended Data Figure 4. Far-field infrared camera images of transmitter steering. (a) Images of several representative steering
positions. The receiver fields of view corresponding to the 16 steering positions are indicated by the dashed lines, with the
currently active block indicated by a solid outline. The light from the active transmit grating is first slightly defocused to
completely illuminate the active block, and then structured by the microlens array. Due to this defocusing and the soft edges
of the beam, a small fraction of the transmitted light falls outside of the active block. (b) A zoomed-in image showing the
structured illumination pattern produced by the microlens array. The locations of the bright spots coincide with the receiver
pixel grating couplers.
18
a b
Vertical edge Horizontal edge
Retroreflector
Retroreflector
Retroreflector Air
Vertical edge
Horizontal edge
Detection threshold
4
Frequency (GHz)
Up-chirp Down-chirp
850 μs 850 μs
Time
d f
Multi-chirp FMCW
4
Frequency (GHz)
Up-chirps Down-chirps
50 × 17 μs 50 × 17 μs
Time
Extended Data Figure 5. Additional characterization of system performance. (a) Imaging contrast measured using retroreflec-
tive sheeting. Our system achieves > 25 dB contrast for a 1 pixel displacement, > 50 dB contrast for a 4 pixel displacement,
and reaches the system noise floor thereafter, illustrating the excellent pixel-to-pixel isolation in our system. Here, the error
bars represent the standard error. (b) Depth precision and detection probability as a function of distance for a 44% reflectance
target. The error bars on the depth precision represent the 95% confidence intervals. (c, d) Single-chirp (c) and multi-chirp
(d) frequency-modulated continuous-wave waveforms used for measuring slow and fast objects respectively. (e, f) Measured
velocity as a function of rotation rate for a 30 cm diameter styrofoam cylinder at a distance of 17 m using (e) single-chirp and
(f) multi-chirp waveforms, with error bars indicating the standard deviation.