Sensors 23 04417 v3
Sensors 23 04417 v3
Communication
On-Chip Compressive Sensing with a Single-Photon Avalanche
Diode Array
Chenxi Qiu † , Peng Wang † , Xiangshun Kong, Feng Yan, Cheng Mao *, Tao Yue and Xuemei Hu *
School of Electrical Science and Engineering, Nanjing University, Nanjing 210023, China;
[email protected] (C.Q.); [email protected] (P.W.); [email protected] (T.Y.)
* Correspondence: [email protected] (C.M.); [email protected] (X.H.)
† These authors contributed equally to this work.
Abstract: Single-photon avalanche diodes (SPADs) are novel image sensors that record photons at
extremely high sensitivity. To reduce both the required sensor area for readout circuits and the data
throughput for SPAD array, in this paper, we propose a snapshot compressive sensing single-photon
avalanche diode (CS-SPAD) sensor which can realize on-chip snapshot-type spatial compressive
imaging in a compact form. Taking advantage of the digital counting nature of SPAD sensing,
we propose to design the circuit connection between the sensing unit and the readout electronics
for compressive sensing. To process the compressively sensed data, we propose a convolution
neural-network-based algorithm dubbed CSSPAD-Net which could realize both high-fidelity scene
reconstruction and classification. To demonstrate our method, we design and fabricate a CS-SPAD
sensor chip, build a prototype imaging system, and demonstrate the proposed on-chip snapshot
compressive sensing method on the MINIST dataset and real handwritten digital images, with both
qualitative and quantitative results.
1. Introduction
Due to its digitized sensing nature of light and high sensitivity, single-photon avalanche
Citation: Qiu, C.; Wang, P.; Kong, X.; diode (SPAD) imagers have been applied in a variety of applications, such as low light
Yan, F.; Mao, C.; Yue, T.; Hu, X. imaging, high dynamic range imaging, etc. [1–3]. Realizing a two-dimensional SPAD sensor
On-Chip Compressive Sensing with a with a large pixel number is a long-pursued goal, which could be readily applied for com-
Single-Photon Avalanche Diode puter vision tasks [4–8]. However, due to the working principle of SPAD, the requirement
Array. Sensors 2023, 23, 4417. of the self-contained controlling circuit, in-pixel or periphery data acquisition, or storage
https://doi.org/10.3390/s23094417 memory become the bottleneck for fabricating the SPAD arrays with a large pixel num-
Academic Editor: Lei Huang ber [9]. Furthermore, the growing requirement of bandwidth with the increase of the pixel
number poses another challenge. To realize format SPAD array, either column-wise shared
Received: 2 April 2023 periphery circuits [10] or 3D stacking technology [11] is used, facing the trade-off between
Revised: 21 April 2023
the readout speed and high cost, with non-reduced data bandwidth. Therefore, how to
Accepted: 28 April 2023
efficiently realize a large-format SPAD array is still an unresolved problem. Since 2006, the
Published: 30 April 2023
theory of compressed sensing (CS) [12–14] has been proposed for efficient data sampling
and high-fidelity sparse recovery. Implementing the idea of CS on the CMOS (comple-
mentary metal–oxide–semiconductor) sensor chip has been recently proposed [15–21] and
Copyright: © 2023 by the authors.
shows promising data throughput reduction on conventional sensing techniques.
Licensee MDPI, Basel, Switzerland. In this paper, to realize efficient SPAD array sensing, we introduce compressed imaging
This article is an open access article technology to the design of the SPAD array. Our CS-SPAD array chip could implement
distributed under the terms and compressive sensing in a snapshot way and reduce the total bandwidth of imaging data
conditions of the Creative Commons to 25%. Besides, the required sensor area for readout electronics with memory is reduced
Attribution (CC BY) license (https:// by 25%.
creativecommons.org/licenses/by/ The entire pipeline of the proposed scheme is shown in Figure 1. Specifically, we
4.0/). propose to realize compressive sensing on a chip with a 32 × 32 SPAD array. Through
introducing different electronic connections between the readout electronics and the sens-
ing pixels, photons collected by different combinations of sensor pixels can be captured
in a summation way. The connections are set according to the compressive sensing ma-
trix and the measurement of each readout electronics corresponds to a CS measurement.
Furthermore, to retrieve information from the compressively sensed data of the CS-SPAD
imaging sensor, we propose a CSSPAD-Net which could realize both image recovery and
image classification based on the captured data of the CS-SPAD array. To demonstrate the
proposed method, we design and fabricate the CS-SPAD sensor array, build a prototype
imaging system based on the chip, and realize both image recovery and classification based
on the output of the sensor, with high fidelity. In all, with the proposed CS-SPAD sensor,
we could realize SPAD array sensing efficiently, i.e., with 25% data throughput and 25%
periphery electronics required for two-dimensional sensing. The performance of imaging
and classification of the CS-SPAD imaging system is demonstrated on the MNIST [22]
dataset and real handwritten digital images, both quantitatively and qualitatively.
Reconstructed
result
..
..
..
..
..
Classification
Target ‘2’ result
CS-SPAD Compressive
scene measurements Multi-task network
2. Related Work
Large-format SPAD imaging arrays have been a long-pursued goal [9–11,23–25], with
wide application in low-light imaging [1,26], high-dynamic-range imaging [2,3], 3D imag-
ing [1,27], etc. To realize the photon-resolving working principle of SPAD, the peripheral
circuit, including the quenching circuit, readout electronics, etc., for generating the photon
counting results are required and occupy a large area of the SPAD sensor, preventing the
implementation of a large-format SPAD imager. Commonly, to realize large-format SPAD
arrays, the peripheral circuits or counting electronics with memory are commonly shared
among columns or rows of pixels [10], facing a trade-off between the readout speed and
required sensor area. To realize large-format SPAD imaging without sharing the readout
electronics, 3D stacking technology is introduced and large-format SPAD imaging can
be realized, at the expense of high cost [11,28]. Furthermore, the requirement of data
bandwidth are not reduced, becoming a potential bottleneck for large-format SPAD arrays.
Thus, how to realize a SPAD imager with an efficient sensor area is still an open problem.
Compressive sensing [12], firstly proposed in 2006, provides an elegant solution to
reduce the required data transmission bandwidth through sampling the scene information
with a sub-Nyquist sampling rate and compressive reconstruction, taking full advantage
of nature image redundancy. Mathematically, the scene redundancy is modeled with
sparsity in the image transform domain, e.g., Fourier or wavelet transform domain [29],
learned transform domain [30,31], etc. Through restricting the sparseness of the signal
in the transform domain, the original signal can be recovered with high fidelity from the
compressive measurement. Based on the compressive sensing theory, efficient imaging
technology based on an optical system is proposed, such as single-pixel imaging [32], ghost
imaging [33], low-light imaging [34] and mid-infrared imaging [35], which could largely
reduce the required data transmission bandwidth.
Sun et al. [36] proposed to build a discrete micromirror device modulation-based
compressive SPAD imaging system to realize a high-resolution SPAD imager. However, the
requirement of a spatial light modulator leads to a large increase of the imaging complexity.
Besides, multiple measurements are required for different compressed codes, which largely
restricts the speed of compression acquisition and prevents real-time imaging.
Sensors 2023, 23, 4417 3 of 13
In order to avoid complex optical systems, some on-chip compressed sensing schemes
have been proposed in conventional imaging sensors, which could realize efficient data
reading and reduced data bandwidth. Conventional CMOS image sensor converts light
intensity into electrical signals for each pixel individually, while CS CMOS image sensors
only sample a small set of random pixel summations [15–21], which can reduce the size
of output data, analog to digital conversion (ADC) operations and the sensor power
consumption.
In this paper, we propose to implement compressive sensing on the SPAD sensor
array, which could realize compressive sensing on the chip and help to reduce the data
throughput and required sensor area for data reading and memory. Specifically, we design
the electrical circuits between the sensing unit and the readout electronics to realize the
compressive sensing process. Each set of readout electronics is designed to count the pixels
in a local unit and integrate the process of data compression into the chip through the local
coupling of pixels on the chip, which reduces the sensor area required for data storage and
data throughput to 25%.
To demonstrate our methods, we propose CSSPAD-Net to process the captured data
by the CS SPAD chip, fabricate and tape out the CS-SPAD image sensor, and build a
prototype imaging system. The effectiveness and efficiency of the proposed CS-SPAD
sensor are demonstrated, both quantitatively and qualitatively, on the MNIST dataset [22]
or real handwritten digital images. We will introduce the details of our method in the
methods section.
3. Methods
In order to realize efficient SPAD array sensing, we design a novel compressive sensing
SPAD array which can directly record the compressively sensed data. In the decoding
process, we propose a neural network designed to directly process the compressively
sensed data, which can reconstruct the scene and realize classification. This section will
introduce our methods, including the proposed compressive imaging chip design and
information processing network architecture for compressed data.
.
.
Circuits 1 Circuits 2 Circuits 3 Circuits 4
.
.
.
(c) (d)
1
4
2
3
Figure 2. (a) Classic sensor array (b), basic compressed sensing imaging unit of CS-SPAD, (c–f) four
different CS connection settings.
Sensors 2023, 23, 4417 4 of 13
3.1.2. CS-SPAD
Figure 2b shows the overall layout of the CS-SPAD chip we designed. The entire
pixel array of CS-SPAD consists of the basic compressed coding unit described above. For
each pixel in the basic compressed coding unit, we use a single-photon avalanche diode
as a solid-state photodetector to record the number of photons arriving at each pixel. The
readout is the sum of the photon counts of all its connected pixels. For each 4 × 4 SPAD
local block, four readout circuits are required to capture the compressively sampled data,
i.e., the use of readout circuits is also reduced to 25%.
layers to eliminate the block artifacts caused by the block processing of the CS-SPAD. For
further utilization of the statistical prior of nature images, we utilize a similar structure
through a global–local residual learning way. Global residual learning [37] could enforce
the overall goal of the network to learn the residual details of the initially reconstructed
image and largely improve the learning difficulty compared to directly learning the image
itself [38], and we introduce the local residual learning in the residual dense block (RDB) to
further helps the fusion of deep and shallow features in the network.
Reconstructed
On-chip compressive
result
sampling
RDB
RDB
RDB
conv
conv
conv
conv
FC
CS-SPAD c +
...
... Reconstruction branch
...
...
...
...
Conv+ Conv+ Conv+ Conv+
Bridge operation BN+ BN+ BN+ BN+
AvgPool AvgPool AvgPool AvgPool
ResBlock
ResBlock
ResBlock
ResBlock
conv
‘2’
FC
c + + +
Classification
Target Classification branch results
scene
(a) Overview
Conv
Conv
Conv
Conv
Conv
Conv
Conv
Relu
Relu
Relu
Relu
Relu
Relu
BN
BN
+ c c c c
Figure 4. (a) Overview of CSSPAD-Net, (b) the structure of residual block (Resblock), (c) the structure
of residual dense block (RDB).
1 rec
LMSE = k x − x k2F . (2)
N
where || F (.)|| is the Frobenius norm. In the classification branch, the loss function is
the cross entropy (CE) loss and the learning rate is set to 0.1. Given predicted vector
xpred = [ x1 , x2 , . . . , xC ] with C classes and the ground truth class j, the CE loss can be
calculated as:
pred
xj
CE e
L = −log pred
. (3)
∑iC=1 e xi
We train our CSSPAD-Net on Nvidia GeForce RTX 2080 for 200 epochs and the learning
rate of both is reduced by half every 50 epochs. The batch size is set to 128. The Adam
optimizer [40] is adopted with β 1 = 0.9, β 2 = 0.999 and e = 1 × 10−8 .
Sensors 2023, 23, 4417 6 of 13
4. Experiments
We verify the CS-SPAD sensor in real scenes and use CSSPAD-Net to complete scene
reconstruction and perceptual classification. Unlike commercial cameras that integrate
focusing and control systems, the CS-SPAD we proposed is just a computational sensor
with a photoelectric conversion function, which requires an additional focusing lens and
control system to cooperate with the CS-SPAD sensor. In this section, we first introduce
our CS-SPAD sensor chip and the optical system, after which we will demonstrate the
effectiveness of the CS-SPAD sensor and the proposed CSSPAD-Net with experimental
results on MNIST data [22].
(a)
(b) (c)
Figure 5. (a) CS SPAD chip overview, (b) the diagram of CS-SPAD sensor chip design, (c) single
pixel layout.
The chip system architecture is shown in Figure 6, consisting of three main parts: a
32 × 32 SPAD detector array, 256-row readout circuits, and an address decoding circuit.
The 32 × 32 SPAD detector circuit includes the SPAD detector, the corresponding gating
quenching circuit, and the logic circuits for multiple “wired AND” operations required by
compressed sensing. The row readout circuit includes a shaping circuit, a transmission
circuit for integrating time, a 12-bit counter circuit, and a 12 bit-buffer circuit. The address
decoding circuit adopts two identical 8 × 16 two-stage decoding modules, which work
synchronously and transmit data through two 12-bit IO data ports.
Sensors 2023, 23, 4417 7 of 13
Ctrans
256×1 12-bit Counter
BL Plus Address
256×1 12-bit Latch
Selector
2×12-bit IO
The chip’s internal structure and the pixel layout are shown in Figure 5b,c, respectively.
We adopt the 0.18 µm 1P6M CMOS technology, and the SPAD pixel size is 15 µm. The array
size is 32 × 32 and the bit depth of the counter is 12 bit. The calibrated dark count of the
proposed CS-SPAD is 200 cps at room temperature, i.e., 300 K. The dead time of the sensor
is about 20 ns. Our CS-SPAD works in avalanche mode with single-photon sensitivity and
the quantum efficiency is about 15%. The performance summary of the CS-SPAD sensor
chip is shown in Table 1.
Parameter Value
Technology 0.18 µm 1P6M CMOS
Chip size 2.9 mm × 3.3 mm
Array size 32 × 32
Pixel size 15 µm
Counter width 12 bit
Dark counts rate 200 cps
Dead time 20 ns
1.8 V (Digital)/3.3 V (Analog)
Power supply
/12 V (SPAD cathode voltage)
Power consumption 10 mW@12 V SPAD cathode voltage
As shown in Figure 7, we build a prototype imaging system based on the proposed CS-
SPAD sensor. From left to right are the target scene, the lens used for focusing, the CS-SPAD
sensor, and the host computer for controlling the automatic execution of the system. The
working pipeline of the CS-SPAD imaging is as follows: first, the host computer controls
the display to display a pattern to be sampled on the monitor, then the CSSPAD is exposed,
and finally the compressed data sampled by the CS-SPAD sensor are read out and the
reconstruction and classification results can be implemented with the trained network on
the host computer.
Camera
Screen lens
CS SPAD
Chip
Camera lens (b)
Host CS SPAD
computer Chip
Peripheral
circuit
(a) (c)
Figure 7. (a) Overview of the prototype imaging system, (b) the detail of camera lens and CS SPAD
chip, (c) the detail of peripheral circuits.
Sensors 2023, 23, 4417 9 of 13
CS-SPAD:
Simulation:
Ground Truth:
CS-SPAD:
Simulation:
Ground Truth:
We also analyzed the reconstruction quality and classification accuracy of different cat-
egories of digital numbers, as shown in Figure 9 Although the PSNR of different categories
of the reconstructed digital numbers fluctuates a little, it has little effect on the classification
results of the classification branch.
40
36.88
99.9 99.2 99.6 99.1 99.7 99.1 99.3
100 98.7 98.8 33.03
97.7
31.33 31.11 31.13 31.29 30.73 30.58 31.03
30 29.42
95
Accuracy/%
PSNR/dB
90
20
85
80
10
75
70 0
0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9
Digits Digits
(a) (b)
Figure 9. Reconstruction and classification results for different digital number categories by CSSPAD-
Net. (a) Reconstruction PSNR with different digital number categories and (b) classification accuracy
with different digital number categories.
CS-SPAD:
Real
handwritten
Digits:
CS-SPAD:
Real
handwritten
Digits:
5. Conclusions
In this paper, we propose CS-SPAD to realize on-chip spatial compressive imaging,
which could reduce the required sensor area for readout electronics and the data throughput
to 25%. CSSPAD-Net is further proposed to recover images from compressively sampled
data and implement high-accuracy perceptual classification. We taped out the CS-SPAD
array and built a prototype imaging system to demonstrate the effectiveness of the sensor.
Quantitative and qualitative experiments on MNIST dataset [22] and real hand-written
digits were conducted to demonstrate the effectiveness and efficiency of our proposed
CS-SPAD imaging sensor. For future work, we plan to further improve the performance of
the CS-SPAD imaging chip and extend the on-chip CS idea to 3D imaging. Specifically, since
existing research on compressed sensing [52–56] have shown that the joint optimization of
the reconstruction network and the compressive sensing matrix can greatly improve the
imaging efficiency, we plan to introduce the end-to-end optimization of the sensing matrix
and reconstruction algorithm to further improve the CS imaging efficiency of our CS-SPAD
imager. Beyond that, since a large-format SPAD array with a time-to-digital converter
(TDC) module is in high demand for 3D imaging, which encounters more severe challenges
of high data bandwidth and large sensor area for peripheral circuits of TDC, we plan to
further develop an on-chip CS-SPAD with a TDC module, based on the proposed CS-SPAD
method, to realize efficient 3D CS detection.
Author Contributions: Conceptualization, X.H. and T.Y.; methodology, C.Q.; sensor design and
fabrication, X.K. and C.M.; software, C.Q. and P.W.; validation, C.Q. and P.W.; formal analysis, C.Q.
and P.W.; investigation, C.Q. and P.W.; resources, C.Q. and P.W.; data curation, C.Q., X.H. and P.W.;
writing—original draft preparation, C.Q. and P.W.; writing—review and editing, X.H.; visualization,
C.Q.; supervision, X.H.; project administration, X.H., T.Y. and F.Y.; funding acquisition, X.H. and T.Y.
All authors have read and agreed to the published version of the manuscript.
Funding: This work was supported by NSFC Projects 61971465, National Key Research and Develop-
ment Program of China (2022YFA1207200), and Fundamental Research Funds 244 for the Central
Universities, China (Grant No. 0210-14380184).
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: The Mnist [22] data presented in this study are openly available and
can be founed at http://yann.lecun.com/exdb/mnist/, accessed on 27 April 2023.
Conflicts of Interest: The authors declare no conflict of interest.
References
1. Shin, D.; Xu, F.; Venkatraman, D.; Lussana, R.; Villa, F.; Zappa, F.; Goyal, V.K.; Wong, F.N.; Shapiro, J.H. Photon-efficient imaging
with a single-photon camera. Nat. Commun. 2016, 7, 12046. [CrossRef] [PubMed]
2. Ingle, A.; Velten, A.; Gupta, M. High flux passive imaging with single-photon sensors. In Proceedings of the IEEE/CVF
Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6760–6769.
3. Liu, Y.; Gutierrez-Barragan, F.; Ingle, A.; Gupta, M.; Velten, A. Single-photon camera guided extreme dynamic range imaging. In
Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 4–8 January 2022;
pp. 1575–1585.
4. Campos, C.; Elvira, R.; Rodríguez, J.J.G.; Montiel, J.M.; Tardós, J.D. Orb-slam3: An accurate open-source library for visual,
visual–inertial, and multimap slam. IEEE Trans. Robot. 2021, 37, 1874–1890. [CrossRef]
5. Zhang, K.; Zhang, Z.; Li, Z.; Qiao, Y. Joint face detection and alignment using multitask cascaded convolutional networks. IEEE
Signal Process. Lett. 2016, 23, 1499–1503. [CrossRef]
6. Zhu, X.; Wang, Y.; Dai, J.; Yuan, L.; Wei, Y. Flow-guided feature aggregation for video object detection. In Proceedings of the IEEE
International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 408–417.
7. Fang, H.S.; Xie, S.; Tai, Y.W.; Lu, C. Rmpe: Regional multi-person pose estimation. In Proceedings of the IEEE International
Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2334–2343.
8. Carreira, J.; Zisserman, A. Quo vadis, action recognition? A new model and the kinetics dataset. In Proceedings of the IEEE
Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 June 2017; pp. 6299–6308.
Sensors 2023, 23, 4417 12 of 13
9. Bruschini, C.; Homulle, H.; Antolovic, I.M.; Burri, S.; Charbon, E. Single-photon avalanche diode imagers in biophotonics:
Review and outlook. Light. Sci. Appl. 2019, 8, 87. [CrossRef]
10. Morimoto, K.; Ardelean, A.; Wu, M.L.; Ulku, A.C.; Antolovic, I.M.; Bruschini, C.; Charbon, E. Megapixel time-gated SPAD image
sensor for 2D and 3D imaging applications. Optica 2020, 7, 346–354. [CrossRef]
11. Morimoto, K.; Iwata, J.; Shinohara, M.; Sekine, H.; Abdelghafar, A.; Tsuchiya, H.; Kuroda, Y.; Tojima, K.; Endo, W.; Maehashi, Y.;
et al. 3.2 megapixel 3D-stacked charge focusing SPAD for low-light imaging and depth sensing. In Proceedings of the 2021 IEEE
International Electron Devices Meeting (IEDM), San Francisco, CA, USA, 11–16 December 2021; pp. 20–22.
12. Donoho, D.L. Compressed sensing. IEEE Trans. Inf. Theory 2006, 52, 1289–1306. [CrossRef]
13. Brunton, S.L.; Proctor, J.L.; Tu, J.H.; Kutz, J.N. Compressed sensing and dynamic mode decomposition. J. Comput. Dyn. 2015,
2, 165. [CrossRef]
14. McMackin, L.; Herman, M.A.; Chatterjee, B.; Weldon, M. A high-resolution SWIR camera via compressed sensing. In Proceedings
of the Infrared Technology and Applications XXXVIII, Baltimore, MD, USA, 23–27 April 2012; Volume 8353, p. 835303.
15. Dadkhah, M.; Deen, M.J.; Shirani, S. Block-based CS in a CMOS image sensor. IEEE Sens. J. 2012, 14, 2897–2909. [CrossRef]
16. Katic, N.; Kamal, M.H.; Kilic, M.; Schmid, A.; Vandergheynst, P.; Leblebici, Y. Power-efficient CMOS image acquisition system
based on compressive sampling. In Proceedings of the 2013 IEEE 56th International Midwest Symposium on Circuits and Systems
(MWSCAS), Columbus, OH, USA, 4–7 August 2013; pp. 1367–1370.
17. Majidzadeh, V.; Jacques, L.; Schmid, A.; Vandergheynst, P.; Leblebici, Y. A (256 × 256) pixel 76.7 mW CMOS imager/compressor
based on real-time in-pixel compressive sensing. In Proceedings of the 2010 IEEE International Symposium on Circuits and
Systems, Paris, France, 30 May–2 June 2010; pp. 2956–2959.
18. Oike, Y.; El Gamal, A. CMOS image sensor with per-column Σ∆ ADC and programmable compressed sensing. IEEE J. Solid-State
Circuits 2012, 48, 318–328. [CrossRef]
19. Leitner, S.; Wang, H.; Tragoudas, S. Design of scalable hardware-efficient compressive sensing image sensors. IEEE Sens. J. 2017,
18, 641–651. [CrossRef]
20. Lee, H.; Kim, W.T.; Kim, J.; Chu, M.; Lee, B.G. A compressive sensing CMOS image sensor with partition sampling technique.
IEEE Trans. Ind. Electron. 2020, 68, 8874–8884. [CrossRef]
21. Park, C.; Zhao, W.; Park, I.; Sun, N.; Chae, Y. A 51-pJ/pixel 33.7-dB PSNR 4× compressive CMOS image sensor with column-
parallel single-shot compressive sensing. IEEE J. Solid-State Circuits 2021, 56, 2503–2515. [CrossRef]
22. Deng, L. The mnist database of handwritten digit images for machine learning research [best of the web]. IEEE Signal Process.
Mag. 2012, 29, 141–142. [CrossRef]
23. Wayne, M.; Ulku, A.; Ardelean, A.; Mos, P.; Bruschini, C.; Charbon, E. A 500 × 500 dual-gate SPAD imager with 100%
temporal aperture and 1 ns minimum gate length for FLIM and phasor imaging applications. IEEE Trans. Electron Devices 2022,
69, 2865–2872. [CrossRef]
24. Bruschini, C.; Burri, S.; Bernasconi, E.; Milanese, T.; Ulku, A.C.; Homulle, H.; Charbon, E. LinoSPAD2: A 512x1 linear SPAD
camera with system-level 135-ps SPTR and a reconfigurable computational engine for time-resolved single-photon imaging. In
Proceedings of the Quantum Sensing and Nano Electronics and Photonics XIX, SPIE, San Francisco, CA, USA, 29 January–2
February 2023; Volume 12430, pp. 126–135.
25. Ma, J.; Zhang, D.; Elgendy, O.A.; Masoodian, S. A 0.19 e-rms read noise 16.7 Mpixel stacked quanta image sensor with 1.1
µm-pitch backside illuminated pixels. IEEE Electron Device Lett. 2021, 42, 891–894. [CrossRef]
26. Sundar, V.; Ma, S.; Sankaranarayanan, A.C.; Gupta, M. Single-Photon Structured Light. In Proceedings of the IEEE/CVF
Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2022; pp. 17865–17875.
27. Gutierrez-Barragan, F.; Ingle, A.; Seets, T.; Gupta, M.; Velten, A. Compressive single-photon 3D cameras. In Proceedings of the
IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2022; pp. 17854–17864.
28. Hutchings, S.W.; Johnston, N.; Gyongy, I.; Al Abbas, T.; Dutton, N.A.; Tyler, M.; Chan, S.; Leach, J.; Henderson, R.K. A
reconfigurable 3-D-stacked SPAD imager with in-pixel histogramming for flash LIDAR or high-speed time-of-flight imaging.
IEEE J. -Solid-State Circuits 2019, 54, 2947–2956. [CrossRef]
29. Candès, E.J.; Wakin, M.B. An introduction to compressive sampling. IEEE Signal Process. Mag. 2008, 25, 21–30. [CrossRef]
30. Hu, X.; Suo, J.; Yue, T.; Bian, L.; Dai, Q. Patch-primitive driven compressive ghost imaging. Opt. Express 2015, 23, 11092–11104.
[CrossRef] [PubMed]
31. Kulkarni, K.; Lohit, S.; Turaga, P.; Kerviche, R.; Ashok, A. Reconnet: Non-iterative reconstruction of images from compressively
sensed measurements. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA,
27–30 June 2016; pp. 449–458.
32. Katkovnik, V.; Astola, J. Compressive sensing computational ghost imaging. JOSA A 2012, 29, 1556–1567. [CrossRef]
33. Duarte, M.F.; Davenport, M.A.; Takhar, D.; Laska, J.N.; Sun, T.; Kelly, K.F.; Baraniuk, R.G. Single-pixel imaging via compressive
sampling. IEEE Signal Process. Mag. 2008, 25, 83–91. [CrossRef]
34. Morris, P.A.; Aspden, R.S.; Bell, J.E.; Boyd, R.W.; Padgett, M.J. Imaging with a small number of photons. Nat. Commun. 2015,
6, 5913. [CrossRef] [PubMed]
35. Wang, Y.; Huang, K.; Fang, J.; Yan, M.; Wu, E.; Zeng, H. Mid-infrared single-pixel imaging at the single-photon level. Nat.
Commun. 2023, 14, 1073. [CrossRef]
Sensors 2023, 23, 4417 13 of 13
36. Sun, Q.; Dun, X.; Peng, Y.; Heidrich, W. Depth and transient imaging with compressive spad array cameras. In Proceedings of
the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 273–282.
37. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual dense network for image super-resolution. In Proceedings of the IEEE
Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2472–2481.
38. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on
Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778.
39. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings
of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125.
40. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980.
41. Parmesan, L.; Dutton, N.; Calder, N.J.; Krstajic, N.; Holmes, A.J.; Grant, L.A.; Henderson, R.K. A 256 × 256 SPAD array with
in-pixel time to amplitude conversion for fluorescence lifetime imaging microscopy. Memory 2015, 900, M5.
42. Perenzoni, M.; Massari, N.; Perenzoni, D.; Gasparini, L.; Stoppa, D. A 160 × 120 Pixel Analog-Counting Single-Photon Imager
With Time-Gating and Self-Referenced Column-Parallel A/D Conversion for Fluorescence Lifetime Imaging. IEEE J. Solid-State
Circuits 2015, 51, 155–167.
43. Lee, C.; Johnson, B.; Jung, T.; Molnar, A. A 72 × 60 angle-sensitive SPAD imaging array for lens-less FLIM. Sensors 2016, 16, 1422.
[CrossRef]
44. Gyongy, I.; Calder, N.; Davies, A.; Dutton, N.A.; Duncan, R.R.; Rickman, C.; Dalgarno, P.; Henderson, R.K. A 256 × 256, 100-kfps,
61% Fill-Factor SPAD Image Sensor for Time-Resolved Microscopy Applications. IEEE Trans. Electron Devices 2017, 65, 547–554.
[CrossRef]
45. Burri, S.; Bruschini, C.; Charbon, E. LinoSPAD: A compact linear SPAD camera system with 64 FPGA-based TDC modules for
versatile 50 ps resolution time-resolved imaging. Instruments 2017, 1, 6. [CrossRef]
46. Zhang, C.; Lindner, S.; Antolovic, I.M.; Wolf, M.; Charbon, E. A CMOS SPAD imager with collision detection and 128 dynamically
reallocating TDCs for single-photon counting and 3D time-of-flight imaging. Sensors 2018, 18, 4016. [CrossRef] [PubMed]
47. Ulku, A.C.; Bruschini, C.; Antolović, I.M.; Kuo, Y.; Ankri, R.; Weiss, S.; Michalet, X.; Charbon, E. A 512 × 512 SPAD image sensor
with integrated gating for widefield FLIM. IEEE J. Sel. Top. Quantum Electron. 2018, 25, 6801212. [CrossRef]
48. Henderson, R.K.; Johnston, N.; Hutchings, S.W.; Gyongy, I.; Al Abbas, T.; Dutton, N.; Tyler, M.; Chan, S.; Leach, J. 5.7 A 256 × 256
40 nm/90 nm CMOS 3D-stacked 120 dB dynamic-range reconfigurable time-resolved SPAD imager. In Proceedings of the 2019
IEEE International Solid-State Circuits Conference (ISSCC), San Francisco, CA, USA, 17–21 February 2019; pp. 106–108.
49. Okino, T.; Yamada, S.; Sakata, Y.; Kasuga, S.; Takemoto, M.; Nose, Y.; Koshida, H.; Tamaru, M.; Sugiura, Y.; Saito, S.; et al. 5.2 A
1200 × 900 6 µm 450 fps Geiger-mode vertical avalanche photodiodes CMOS image sensor for a 250 m time-of-flight ranging
system using direct-indirect-mixed frame synthesis with configurable-depth-resolution down to 10 cm. In Proceedings of the
2020 IEEE International Solid-State Circuits Conference (ISSCC), San Francisco, CA, USA, 16–20 February 2020; pp. 96–98.
50. Kumagai, O.; Ohmachi, J.; Matsumura, M.; Yagi, S.; Tayu, K.; Amagawa, K.; Matsukawa, T.; Ozawa, O.; Hirono, D.; Shinozuka,
Y.; et al. 7.3 A 189 × 600 back-illuminated stacked SPAD direct time-of-flight depth sensor for automotive LiDAR systems. In
Proceedings of the 2021 IEEE International Solid-State Circuits Conference (ISSCC), San Francisco, CA, USA, 13–22 February
2021; Volume 64, pp. 110–112.
51. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE
Trans. Image Process. 2004, 13, 600–612. [CrossRef] [PubMed]
52. Shi, W.; Jiang, F.; Zhang, S.; Zhao, D. Deep networks for compressed image sensing. In Proceedings of the 2017 IEEE International
Conference on Multimedia and Expo (ICME), Hong Kong, China, 10–14 July 2017; pp. 877–882.
53. Chen, Z.; Guo, W.; Feng, Y.; Li, Y.; Zhao, C.; Ren, Y.; Shao, L. Deep-learned regularization and proximal operator for image
compressive sensing. IEEE Trans. Image Process. 2021, 30, 7112–7126. [CrossRef]
54. Cui, W.; Liu, S.; Jiang, F.; Zhao, D. Image Compressed Sensing Using Non-local Neural Network. IEEE Trans. Multimed. 2021, 25,
816–8303. [CrossRef]
55. Fan, Z.E.; Lian, F.; Quan, J.N. Global Sensing and Measurements Reuse for Image Compressed Sensing. In Proceedings of the
IEEE Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 8954–8963.
56. Mou, C.; Zhang, J. TransCL: Transformer makes strong and flexible compressive learning. IEEE Trans. Pattern Anal. Mach. Intell.
2022, 45, 5236–5251. [CrossRef] [PubMed]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.