Technical Seminar Report
Technical Seminar Report
PRAMOD H N (1SI21EC117)
CERTIFICATE
PO1 PO2 PO3 PO4 PO5 PO6 PO7 PO8 PO9 PO10 PO11 PSO1 PSO2
CO-1 3 2 2
CO-2 3 2 3 2 1 1 1 2
CO-3 3 3 3 3 1 2 2 2 2
CO-4 3 2 2 2 2 1 2 2 2 2
Average 3 2 3 3 2 2 1 2 2 2 2
The primary objective of this work is to provide a comprehensive overview of how three-
dimensional (3D) stacking techniques—such as pixel-level and circuit-level stacking—have
revolutionized image sensor performance. Key technologies discussed include through-
silicon vias (TSVs), hybrid bonding, and wafer-to-wafer stacking, as well as the integra-
tion of advanced analog and digital circuits. By examining recent research trends and
practical implementations, this study demonstrates how stacking enhances sensor capa-
bilities, optimizes form factors, and overcomes traditional trade-offs between resolution,
sensitivity, and speed.
The implementation aspect focuses on the design challenges and opportunities presented
by stacked architectures, including thermal management, interconnect strategies, and sig-
nal integrity. Practical case studies are highlighted to showcase how stacking enables
innovations such as high dynamic range imaging, low-light performance, and real-time
image processing. This study ultimately emphasizes the transformative impact of stacked
device technologies on the future of image sensors, positioning them as critical enablers
for intelligent imaging systems in an increasingly connected world.
i
Contents
Abstract i
1 Introduction 1
1.1 Literature Survey . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.2 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
2 Analysis Of Methodology 5
2.1 Front-Illuminated and Back-Illuminated Structures . . . . . . . . . . . . . 5
2.2 Stacked Device Architecture . . . . . . . . . . . . . . . . . . . . . . . . . . 5
2.3 Advanced Sensor Architectures . . . . . . . . . . . . . . . . . . . . . . . . 6
2.4 Movie Recording With More Than Megapixels . . . . . . . . . . . . . . . 6
2.4.1 Double-Column ADC Architecture . . . . . . . . . . . . . . . . . . 7
2.4.2 Three-Layer Stacked CMOS Image Sensor With DRAM . . . . . . 8
2.4.3 Chip-on-Wafer (CoW) Technology for Large Optical Format . . . . 9
2.5 Pixel-Parallel Architectures . . . . . . . . . . . . . . . . . . . . . . . . . . 10
2.5.1 Cu–Cu Connections in Stacked Image Sensors . . . . . . . . . . . . 10
2.5.2 Stacked Pixel Circuit Extensions . . . . . . . . . . . . . . . . . . . 11
2.5.3 Pixel-Parallel ADC . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
2.5.4 Pixel-Parallel Photon Counter . . . . . . . . . . . . . . . . . . . . . 11
2.6 Extension of Sensing Capabilities . . . . . . . . . . . . . . . . . . . . . . . 12
2.7 Spatial Depth . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
2.7.1 Temporal Contrast Sensing . . . . . . . . . . . . . . . . . . . . . . 13
2.7.2 Invisible-Light Imaging . . . . . . . . . . . . . . . . . . . . . . . . . 14
2.8 Intelligent Vision Sensor . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
2.9 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
3 Analysis Of Results 24
3.1 Sensor Performance Evaluation . . . . . . . . . . . . . . . . . . . . . . . . 24
3.2 Application Demonstrations . . . . . . . . . . . . . . . . . . . . . . . . . . 25
3.3 Comparison With Previous Designs . . . . . . . . . . . . . . . . . . . . . . 25
3.4 Trend Analysis and Scalability . . . . . . . . . . . . . . . . . . . . . . . . . 25
3.5 Energy Efficiency and Practical Relevance . . . . . . . . . . . . . . . . . . 26
4 Conclusion 27
Bibliography 27
Chapter 1
Introduction
Image sensors play a crucial role in modern digital imaging applications, ranging from con-
sumer electronics to industrial and automotive sectors. Since the development of charge-
coupled devices (CCDs) in 1969, solid-state image sensors have undergone significant
advancements, leading to their widespread adoption in digital cameras, video recording
devices, and machine vision systems. Over the past two decades, CMOS (Complemen-
tary Metal-Oxide-Semiconductor) image sensors have replaced CCDs as the dominant
technology due to their lower power consumption, faster readout speeds, and improved
integration capabilities.
The increasing demand for high-performance imaging in smartphones, security cameras,
autonomous vehicles, and industrial automation has driven continuous innovation in
CMOS image sensor technology. A major breakthrough in this field was the introduction
of back-illuminated (BI) image sensors, which addressed the limitations of conventional
front-illuminated (FI) structures by improving light sensitivity and optimizing pixel ar-
rangement. The adoption of BI structures has enabled further advancements, particularly
in the development of stacked CMOS image sensors, which enhance imaging performance
by integrating logic circuits and signal processing elements directly onto separate layers.
Stacked image sensor architectures leverage wafer-bonding and advanced wafer-thinning
techniques to achieve compact, high-performance designs. This approach allows for the in-
tegration of highly parallel column-parallel analog-to-digital converters (ADCs), enabling
high-resolution and high-frame-rate imaging. Furthermore, pixel-pitch Cu–Cu connec-
tions facilitate pixel-parallel digitization, further improving sensor efficiency. The evolu-
tion of stacked CMOS sensors has also led to the integration of edge computing function-
alities, allowing real-time image processing and AI-based decision-making directly on the
sensor chip.
This report provides a comprehensive review of recent advancements in stacked CMOS
image sensors, focusing on their architectural evolution, improved sensing capabilities,
and the integration of AI-driven edge computing. The subsequent sections explore vari-
1
Advancements in Image Sensor Architectures Using Stacked Device Technologies 2024-25
ous sensor configurations, advanced pixel circuit implementations, extended sensing func-
tionalities such as spatial depth and contrast sensing, and emerging AI-integrated vision
sensors. Through these advancements, stacked CMOS image sensors continue to redefine
the future of digital imaging, offering superior performance, enhanced functionality, and
broader application potential.
1.2 Summary
This literature survey explores the advancements, design principles, and applications of
CMOS image sensors (CIS), highlighting their advantages over CCD sensors, such as
low power consumption, high-speed operation, and system-on-chip (SoC) integration. It
5
Advancements in Image Sensor Architectures Using Stacked Device Technologies 2024-25
and high-frame-rate imaging through column-parallel ADCs.
Figure 2.2: Implementations of stacked CMOS image sensors: (a) Connections by TSVs
between photodiodes and logic circuits, (b) First stacked CMOS image sensor, (c) Double-
column ADC architecture [1]
Figure 2.3: Implementations of stacked CMOS image sensors: (a) TSV connections be-
tween photodiodes and logic circuits, (b) first stacked CMOS image sensor, (c) double-
column ADC architecture [1]
The sensor chip was fabricated using a 90-nm custom process for photodiodes and NMOS
logic, while the logic chip used a 65-nm standard CMOS process. As the ADCs were
independent of the sensor chip, they could be highly integrated. Redundant parallel
ADCs also improved noise characteristics by averaging multiple analog-to-digital (AD)
conversions, as illustrated in Fig.2.3 .
Increasing pixel count and parallel ADCs results in a higher data output. In 2017, a
three-layer stacked CMOS image sensor was developed to record slow-motion videos at
960 fps, as depicted in Fig. 2.4. TSVs connected the layers, and data from parallel ADCs
were buffered in the second-layer DRAM. This sensor enabled full-HD resolution at 960
fps by buffering digital data through a 102-Gbit/s bus.
When capturing a movie at 30 fps, the readout speed changes to 960 fps upon detecting
rapid motion or user triggers. Up to 63 full-HD frames can be stored in DRAM and
outputted during subsequent movie capture.
Figure 2.4: Three-layer stacked CMOS image sensor with DRAM for high-frame-rate
video capture [1]
Stacked CMOS image sensors are generally fabricated using wafer-on-wafer (WoW) bond-
ing. However, WoW is suboptimal for large optical formats, as sensor and logic chip sizes
must match. The alternative CoW bonding process, illustrated in Fig. 2.5, optimizes
area efficiency by allowing logic chip sizes to differ from the sensor.
A CoW-bonded stacked CMOS image sensor was introduced in 2016 to achieve a global-
shutter image sensor for broadcast cameras with a super-35-mm optical format. Two
diced 65-nm logic chips with parallel ADCs and microbumps were stacked on a large
custom sensor chip. Microbumps with a 40 µm pitch created around 38,000 connections.
This sensor enabled 480 fps imaging with 8M pixels.
Fig. 2.7 presents performance trends in large optical-format image sensors, where a full-35-
mm format sensor with 50M pixels reached 250 fps by 2021. The WoW process achieves
high performance by increasing parallel ADCs and SRAM buffers, while CoW balances
cost and performance. Recent advancements include a 3.6-inch, 127M-pixel image sen-
sor with four stacked logic chips, aiming for higher productivity through improved chip
placement throughput.
Stacked CMOS image sensor technology has revolutionized high-speed imaging by en-
abling higher frame rates, reduced noise, and better integration of processing units.
Figure 2.5: Area efficiencies of WoW and CoW bonding processes for a large optical-
format image sensor [1]
Double-column ADC architectures, three-layer stacking with DRAM, and CoW bond-
ing have significantly contributed to advancements in digital imaging, facilitating super
slow-motion recording and large optical-format imaging. Future developments focus on
improving chip stacking efficiency and further enhancing image sensor performance.
The connections between the sensor and the logic layers have been changed from TSVs to
hybrid-bonded Cu–Cu connections, as shown in Fig. 2.8(a). In the TSV configuration, the
signal lines are delivered to the logic layer at the periphery of the pixel array. In contrast,
Cu–Cu connections can be integrated directly under the pixels, allowing an increase in
the number of linkages. The recent trends regarding Cu–Cu connection pitches are shown
in Fig.2.8(b).
Figure 2.6: Stacked CMOS image sensor using CoW bonding process for large optical
format [1]
2021. The BI SPAD pixel array is stacked on a bottom chip with readout circuitry via
pixel-parallel Cu–Cu connections.
Figure 2.8: (a) Cu–Cu hybrid bonding structure in stacked image sensors. (b) Trends in
Cu–Cu connection pitches [1]
sensitivity. The 10 µm-pitch SPAD pixel, featuring a 7 µm-thick silicon layer, achieves a
PDE above 31.4% at 850 nm and 14.2% at 940 nm, with a dark count rate (DCR) below
2 cps/µm2 at 60◦ C.
A 189×600 SPAD direct time-of-flight (ToF) sensor for automotive LiDAR applications
was reported in 2021. The sensor utilizes a BI-stacked SPAD array, where pixel front-end
circuits are implemented on the bottom chip, achieving a distance accuracy of 30 cm over
a range of up to 200 m under 117k lux sunlight conditions.
Figure 2.9: Pixel circuit configurations for single and dual conversion gains [1]
Figure 2.10: Pixel circuit configurations of voltage-domain global shutter stacked via
pixel-parallel Cu–Cu connections [1]
low-light scene capture and decoding temporally modulated structured light patterns for
depth estimation.
As shown in Figure 2.20, the trend of reducing pixel pitch in EVSs has led to sub-5 µm
pixels, enabling megapixel resolutions and practical deployments in real-time machine
vision applications.
Figure 2.24 illustrates the InGaAs image sensor fabrication process, where a 5-µm pixel
photodiode array (PDA) is connected to a readout integrated circuit (ROIC) using Cu–
Cu bonding. This approach replaces traditional flip-chip bumps, enabling fine-pitch pixel
scaling. As a result, InGaAs sensors achieve high-definition SWIR imaging with applica-
tions in inspection and security monitoring, even under foggy conditions.
Figure 2.12: Chip implementation of pixel-parallel ADC. (a) Chip micrograph. (b)
Rolling-shutter image. (c) Global-shutter image [1]
computations was reported in 2021. This sensor features an integrated solution that
performs complete image capture and transfers data to a CNN inference processor, as
shown in Figure 2.25. The system operates at 120 frames per second (fps) and includes
on-chip CNN processing using a 4.97 TOPS/W DSP. The processing block consists of an
ISP for CNN input preprocessing, a DSP subsystem optimized for CNN processing, and
an 8MB L2 SRAM for storing CNN weights and runtime memory.
Fig.2.24 presents examples of CNN inference results using MobileNet v1, demonstrating
similar performance to TensorFlow. The intelligent vision sensor can run a complete CNN
inference process on the sensor itself, outputting both raw image data and CNN inference
results in the same frame via the MIPI interface. Additionally, the sensor supports the
output of CNN inference results solely through the SPI interface, enabling a small camera
system with reduced power consumption and lower cost.
The embedded CNN inference processor allows users to program their AI models into
memory and reprogram them as required by the specific conditions of deployment. For
example, when installed at a facility entrance, it can count visitors; when placed on a store
shelf, it can detect stock shortages; and when mounted on a ceiling, it can perform heat
mapping of store visitors. Intelligent vision sensors are anticipated to provide low-cost
edge AI solutions with flexible AI models for various applications.
Figure 2.13: Photon-counting image sensor. (a) Chip configuration. (b) Simplified pixel
circuit [1]
2.9 Summary
The proposed methodology leverages stacked device technologies to enhance vision sen-
sor capabilities in depth sensing, temporal contrast detection, and invisible-light imaging.
The BI-stacked SPAD sensor improves near-infrared (NIR) spectral response and depth
accuracy for automotive LiDAR. Event-based vision sensors (EVS) enhance temporal con-
trast detection using asynchronous delta modulation, enabling high-speed machine vision
applications. Additionally, stacked device integration with InGaAs and Ge-on-Si photode-
tectors enables high-resolution short-wavelength infrared (SWIR) imaging for industrial
and security applications. Intelligent vision sensors with embedded AI processing provide
efficient edge AI solutions for applications like visitor counting, stock monitoring, and
heat mapping, reducing dependency on cloud computing.
Figure 2.14: Timing diagram for photon-counting with subframe extrapolation [1]
Figure 2.15: Measured results of photon-counting. (a) Dynamic range and SNR. (b)
Captured HDR image. (c) Captured image with motion-artifact suppression [1]
Figure 2.16: SPAD device structures. (a) FI SPAD. (b) BI-stacked SPAD [1]
Figure 2.17: BI-stacked SPAD with direct ToF depth sensor [1]
Figure 2.18: Pixel block diagram of the event-based vision sensor [1]
Figure 2.19: BI-stacked EVS and examples of its applications: (a) Chip micrographs; (b)
Application examples demonstrating low-light sensitivity and 3D structured light decod-
ing. [1]
Figure 2.20: Trend of pixel pitch in EVSs, demonstrating the evolution toward smaller
pixels enabled by stacked device technologies [1]
Figure 2.21: Fabrication process of InGaAs image sensors using Cu–Cu bonding [1]
Figure 2.22: Trend in contact pitch of flip-chip bump and Cu–Cu bonding as well as
application examples of InGaAs sensors [1]
Figure 2.25: Evolution and future prospects of stacked CMOS image sensors [1]
• Resolution: 1280×720
• Temporal Resolution: 1 µs
These results validate the sensor’s ability to perform under real-time constraints while
maintaining low power consumption and high event throughput.
24
Advancements in Image Sensor Architectures Using Stacked Device Technologies 2024-25
27
Bibliography
[1] Y. Oike, “Evolution of Image Sensor Architectures With Stacked Device Technolo-
gies,” in IEEE Transactions on Electron Devices, June 2022 vol. 69, no. 6, pp. 2757-
2765.
[2] B. Pain, “Fabrication and initial results for a back-illuminated monolithic APS in
a mixed SOI/bulk CMOS technology,” in Proc. IEEE Workshop CCD Adv. Image
Sensors, Jun. 2005, pp. 102–104.
[5] H. Rhodes et al., “The mass production of BSI CMOS image sensors,” in Proc. Int.
Image Sensor Workshop, Jun. 2009, pp. 27–32.
[9] Y. Nitta et al., “High-speed digital double sampling with analog CDS on column
parallel ADC architecture for low-noise active pixel sensor,” in IEEE Int. Solid-State
Circuits Conf. (ISSCC) Dig. Tech. Papers, Feb. 2006, pp. 500–501.
28
Advancements in Image Sensor Architectures Using Stacked Device Technologies 2024-25
[10] S. Yoshihara et al., “A 1/1.8-inch 6.4 MPixel 60 frames/s CMOS image sensor with
seamless mode change,” in IEEE Int. Solid-State Circuits Conf. (ISSCC) Dig. Tech.
Papers, Feb. 2006, pp. 492–493.
[11] A. Suzuki et al., “A 1/1.7-inch 20 Mpixel back-illuminated stacked CMOS image sen-
sor for new imaging applications,” in IEEE Int. Solid-State Circuits Conf. (ISSCC)
Dig. Tech. Papers, Feb. 2015, pp. 1–3.
[12] T. Haruta et al., “A 1/2.3-inch 20 Mpixel 3-layer stacked CMOS image sensor with
DRAM,” in IEEE Int. Solid-State Circuits Conf. (ISSCC) Dig. Tech. Papers, Feb.
2017, pp. 76–77.
[13] Y. Oike et al., “An 8.3 M-pixel 480 fps global-shutter CMOS image sensor with gain-
adaptive column ADCs and 2-on-1 stacked device structure,” in Proc. IEEE Symp.
VLSI Circuits (VLSI-Circuits), Jun. 2016, pp. 222–223.
[15] Image Courtesy: Evolution of Image Sensor Architectures With Stacked Device Tech-
nologies, IEEE Transactions on Electron Devices, June 2022, vol. 69, no. 6.