Jaechoi 1
Jaechoi 1
by
Jaehyuk Choi
Doctroral Committee:
My tenure as a graduate student at the University of Michigan has been a great time
in my life. I experienced many things and learned a lot, and importantly I had an exciting
First I would like to thanks my advisor Professor Euisik Yoon. He has supported my
study in the graduate school and always provided a motivation and trust. I learned
I also would like to thank my dissertation committee Professor Ken Wise, David D.
Wentzloff, and James W. Cutler. I greatly appreciate their feedback and encouragement.
I would like to thank current image group members Seok-Jun Park and Jihyun Cho.
We spent considerable time together especially during tapeout. They helped me a lot of
things in both design and testing. I appreciate their contribution to this work. I would also
like to thank previous image group members, Kwang-Hyun Lee, Sang-Wook Han and
Seong-Jin Kim. Due to their advice and instruction, my research about the multi-
I give thanks to current and past lab members Xia Lou, Khaled AlAshmouny, Fan Wu,
Sung-Yun Park, Kyunghwan Na, Yu-Chih Chen, Patrick Ingram, Roberto Miguez, Mike
McCormick, Maesoon Im, Il-Joo Cho, Hyung-Kew Lee, Jaehoon Chung, Young-Ji Kim,
Hwa Sung Shin, Dr. Songjun Lee, Dr. Bo Yang and Dr. Ararta Nakajima. I had
ii
Finally, I want to give my thanks to my mother. Thank you for constant
iii
TABLE OF CONTENTS
Acknowlegements.............................................................................................................. ii
Abstract………………………………………………………………………………xiii
iv
2.4.2 4-T Pixel with pinned photodiode .................................................................. 29
2.4.3 Shared-pixel architecture................................................................................ 30
v
3.4.2 Region-of-Interest (ROI) readout ................................................................... 89
3.4.3 Variable-resolution image Readout ................................................................ 91
3.6 Summary…………………………………………………………………………..95
4.2 Pixel and column circuits for reconfigurable modes ............................................. 104
vi
5.4.2 Power figure of merit (FOM) comparison with previous low-power imagers
……………………………………………………………………...140
vii
LIST OF FIGURES
viii
Figure 2.16 Shared pixel circuit. ............................................................................... 31
Figure 3.4 Pixel circuit with load capacitance of in-pixel source follower .............. 47
Figure 3.5 Correlated double sampling circuit with programmable gain amplification
........................................................................................................................... 48
Figure 3.10 Digital signal readout circuit and the operation .................................... 54
Figure 3.12 Overall power consumption in the signal chain according to the spatial
Figure 3.13 Analog power consumption in the signal chain according to the spatial
Figure 3.14 Digital power consumption in the signal chain according to the spatial
ix
Figure 3.15 Digital power consumption according to the power supply voltage (320
× 240) ................................................................................................................ 60
Figure 3.16 Overall power consumption according to the pixel pitch (a) 320 × 240
Figure 3.19 Charge injection and clock feedthrough in the reset transistor ............. 66
Figure 3.21 Reset level boosting with capacitive coupling (1) ................................. 69
Figure 3.22 Reset level boosting with capacitive coupling (2) ................................. 69
Figure 3.27 In-pixel nMOS common source amplifier with p-n photodiode ........... 76
Figure 3.28 In-pixel pMOS common source amplifier with pinned photodiode ...... 77
Figure 3.33 Lateral overflow scheme (a) Pixel circuit (b) Operation procedure ...... 83
x
Figure 3.36 In-pixel comparator for the frame difference ........................................ 88
Figure 3.39 Pixel and column architecture of the multi-resolution image sensor .... 93
Figure 3.40 Multi-resolution readout: (a) Block diagram (b) Timing diagram ........ 94
Figure 4.2 Operation and block diagram of the adaptive image sensor with
Figure 4.4 Pixel architecture and equivalent circuits in each mode ....................... 106
Figure 4.5 Timing diagram of the pixel control in each mode ............................... 107
Figure 4.8 Column-parallel ADC for the reconfigurable modes ............................ 109
Figure 4.9 Timing diagram of column-parallel ADC (a) Single ended input (normal,
Figure 4.11 Comparison with conventional latch in single-slope ADC ................. 112
Figure 4.12 Conventional digital image signal readout circuit and operation ........ 112
Figure 4.14 Digital image signal readout circuit with in-equality readout ............. 114
xi
Figure 4.16 Dual exposure scheme: pixel output response and conventional
Figure 5.6 Equivalent pixel circuit for motion sensing and low-power imaging ... 131
Figure 6.2 Illustration of imaging Cubic (iCube) sensor platform. ........................ 148
xii
LIST OF TABLES
xiii
ABSTRACT
In this thesis, an energy-efficient CMOS image sensor has been studied to adaptively
imaging systems.
In wireless sensor nodes, image sensors typically have four main challenges to
address: low power operation from limited energy sources such as batteries or energy
harvesting units, high dynamic range imaging to cover a wide range of illumination, high
spatial resolution for high-quality imaging, and high temporal resolution for video
streaming output. However, it is difficult to optimize all of these as there are trade-offs
among these four parameters and additionally the environmental conditions continuously
One of the solutions to this trade-off is implementing an image sensor with situation-
aware adaptability. In this work, three on-chip adaptive functions have been investigated
imaging for high-dynamic range, and (3) object-adaptive imaging for temporal-spatial
bandwidth suppression.
For proofs of concepts of this adaptive imaging, we designed two prototype chips.
Both prototype sensors have been fabricated and fully characterized. In the first chip, we
xiv
Most of the time, the sensor operates in a monitoring mode at an extremely low power
(less than 1.36 μW/frame) with the power supplied from energy harvesting units. It
reconfigurability for sensitivity and dynamic range. The sensor can provide a high
sensitivity of 23.9 V/lx·s at low illumination, but it can be reconfigured to generate a high
suppress the redundancies in spatial and temporal bandwidths in image signals. The
sensor will wake up when triggered by motion and extracts features from the captured
image for the detection of objects-of-interest. Full image capture operation is performed
only when the objects-of-interest are found, resulting in a significant reduction of total
power consumption at the sensor node. Object detection has been performed using the
classification algorithm and has achieved a 94.5 % success rate for human recognition
changing in the focal plane. Using the three adaptation schemes, the implemented image
sensors provide high dynamic range images (> 99 dB) at low power consumption (< 1.36
μW) while maintaining low bandwidth (< 3.5 %) for image signal transmission.
xv
CHAPTER 1 INTRODUCTION AND BACKGROUND
according to their applications and technologies. In the 1st generation, the emergence of
digital imaging systems enabled the development of a variety of devices such as digital
still cameras, camcorders, webcams and so on. Industry also benefitted from the digital
imaging solutions, for example, broadcasting systems, machine visions, and inspection
systems. In the 1st generation of imaging systems, Charge Coupled Devices (CCD) had
been widely used since developed in the 1970s. In the 2nd generation, imaging systems
were integrated with handheld devices such as cell phones, laptops, and tablet PCs during
the 1990s. This evolution was mainly enabled by the development of a low-cost CMOS
from the 1st generation imaging systems which were mainly used for consumer
electronics and industry are that the 2nd generation imaging systems are operated by
battery and are linked with the networks. As such the image sensor consumes less power,
has a small form factor, and has a low cost for the application of portable devices. Due to
these demands, CMOS image sensors replaced CCDs in most imaging products [1.1].
Currently, imaging systems are expected to be distributed in a wide area due to the
sensors also integrate image sensors for applications in endoscopy, microscopy and
retinal implants. This 3rd generation of imaging systems are distributed, fully networked,
1
and operated with extremely low-power consumption from energy harvesting in order to
monitor wide or unreachable areas. The networked distributed imaging systems have
3. Statistical convergence: we can get the information based on the gross statistics of
the environment
Generation 1 2 3
Appliance, Distributed
Handheld devices
industry network
Applications Digital still camera, Cell phone,
Surveillance networks,
Micro vehicles,
camcorder, laptop,
Wireless biomedical
Industrial automations tablet
imaging
The main applications of distributed image sensor network are military surveillance,
healthcare / patient monitoring. The other main applications of 3rd generation imaging
systems are biomedical sensors. One such commercialized sensor is the capsule
endoscope [1.3] which integrates an image sensor, a controller, and a wireless transmitter.
These image sensors can also be integrated with a wireless microscope [1.4] or with a
2
retinal stimulator [1.5, 1.6], and so on. The main features of each generation are
sites serially by applying clock signals. Incident photons are converted to charge by a
transferred horizontally through horizontal CCD using the clock signal. In order to
enhance the charge transfer efficiency, a high voltage clock signal is used. In the output
stage, transferred charge is converted to voltage through the amplifier. The basic structure
CCDs have been used for high-end imaging systems due to low noise and high
quantum efficiency. However, it consumes large power due to a charge transfer operation
with high voltage. Power consumption is critical factor for mobile devices or any other
3
power-limited applications which require extremely low power consumption. Scalability
is another important issue because it is difficult to scale the pixel array larger due to
limited speed of charge transfer. Moreover, a CCD image sensor is difficult to integrate
CMOS image sensors have not been widely used although it has longer history than a
CCD image sensor. Since MOS image sensors have been developed in 1960’s, low SNR
of a CMOS image sensor made it behind from the market, and main focus of research in
In 1980’s, a MOS image sensor was resurfaced due to the development of CMOS
fabrication technology. CMOS image sensor technology was rapidly developed due to the
An imaging system with CMOS image sensor can be implemented with a single chip,
can be fabricated using standard CMOS technology with low cost. By integrating ADCs,
periphery circuits, and image signal processors into one chip, the size of imaging system
can be reduced. On the other hand, fabrication of CCD is specialized and optimized for
effective photo-carrier generation and charge transfer. Imaging systems using CCD
image processors. Accordingly, the cost for imaging systems with CCD imager is higher
CMOS image sensors can be categorized into two types: a passive pixel sensor or an
active pixel sensor. A passive pixel sensor (PPS) consists of a photodiode and a selection
4
transistor as shown in figure 1.2 (a). The photocurrent which is generated from incident
light is converted to voltage in the column-parallel charge amplifier [1.8]. A PPS has a
high fill factor which is defined as the ratio of photodiode area over pixel area. In general,
high fill factor provides higher sensitivity in a given form factor because a photodiode
can collect more electrons. However, a PPS has low SNR for several reasons. First, small
photo-current is difficult to read out due to large capacitance of column line. Second,
considerable leakage current through the selection transistor corrupts the signal.
Due to the low SNR of a PPS, most CMOS imagers have an active pixel sensor (APS)
architecture. As shown in figure 1.2 (b), an active pixel has an in-pixel amplifier in order
to drive the column line, which has large parasitic capacitance. Fill factor is decreased
compared with passive pixel because of the area from in-pixel circuits. However, the
isolation from the column line through an in-pixel buffer provides higher SNR compared
CMOS image sensors output the voltage using an in-pixel amplifier, whereas the
charge is transferred serially in CCD image sensors. Therefore, readout with high frame
rate and random access is possible. Moreover, photo-detectors such as photo-gates and
pinned photodiodes provide enhanced sensitivity and better noise immunity compared
with conventional p-n junction photodiodes. The architecture and the operation
5
Figure 1.2 Passive pixel and active pixel sensor architecture.
have four main challenges in image sensors as shown in figure 1.3. The first is power
consumption because imaging systems use limited energy sources such as a battery or
energy harvesting from the environment. State-of-art image sensors have large power
The power consumption of CMOS image sensors is difficult to be scaled down because
the voltage scaling directly suppresses the SNR which brings loss of image quality. The
second challenge is wide dynamic range. In outdoor applications, the illumination level
varies from extremely dark to extremely bright. In biomedical sensors, illumination level
is typically low, especially when monitoring inside the body or monitoring cells.
Accordingly, the dynamic range and the sensitivity of image sensors should be high in
6
Power Consumption Illumination Range
(Energy harvesting, battery or
(Saturated or dark image)
wireless power delivery)
schemes do not have enough dynamic range and sensitivity. The third challenge is large
bandwidth (spatial) for large pixel arrays. The image signal has 8-bit (gray-scale) or > 10-
bit (color) bit-depth. The number of pixels is typically larger than 10,000 (100×100
pixels). Therefore we need to transmit 80,000-bit for the transmission of one image frame,
such as JPEG and DPCM can reduce the bandwidth. However, the image compression
entails the loss of information and requires power-consuming processing. The last
challenge is high temporal resolution (or frame rate). Conventional image sensors have
fixed frame rate from 15 to 60 frames per second (fps). Some high-speed sensors have
high frame rate (> 1,000 fps) in order to suppress the motion-blur from moving objects
[1.9-1.11]. As mentioned, transmitting 80k-bit with 15 fps requires huge bandwidth and
power consumption. However, it is rare that the event we want to monitor occurs 15
times in a second. Therefore, imaging with constant frame rate is redundant. If sensors
can monitor specific event and provide event-driven image transmission, temporal
trade-off relationships each other. For example, the reduction of power consumption
entails the reduction of dynamic range and bandwidth. Bandwidth reduction also entails
the reduction of dynamic range or the increase of power consumption when we use
that consider these four parameters: power consumption, dynamic range & sensitivity,
spatial bandwidth and temporal bandwidth. However, this optimization point varies
applicable to the 3rd generation imaging system including distributed image sensor
networks and biomedical sensors, i.e., implementing an image sensor with low-power
consumption, wide dynamic range, and low spatial-temporal bandwidth. Considering four
main parameters (power consumption, dynamic range & sensitivity, spatial bandwidth,
and temporal bandwidth), an efficient way to implement an image sensor is providing the
adaptability, because these four parameters always have trade-off. Figure 1.4 shows the
suppress the redundancy of data and reduce the power consumption by adapting to the
environment. In the adaptive imaging, the first procedure starts from motion sensing to
reduce temporal bandwidth. The sensor is in sleep mode, while only the integrated
motion sensor is turned on, and monitors with extremely low-power consumption. The
sensor is woken up and is triggered to operation mode from the movement of the object.
8
transmission with fixed frame rate. Therefore, we can save both the power consumption
sensor monitors the object in the focal plane. As mentioned, not all of the objects in the
scene are of interests. For example, a swaying tree in the background is not interesting.
The sensor detects target objects in the 2nd procedure, i.e., it classifies whether the
object in the focal plane is of our interest or not. In the 2nd procedure, instead of
generating an image signal which has a large amount of data and redundancy, we
generate a low-bandwidth feature for the object detection. The generated feature can be
processed for the object detection in the host or in the local node. The processor classifies
the object and feeds back the request signal to the sensor. If a target object is detected,
imaging operation is initiated. This object-of-interest imaging for the object adaptation
9
After thhe object deetection, thee sensor staarts the imaaging operaation with loow power
c
consumptio n. The mainn goal of thhe imaging is
i to providde a quality image signal in wide
r
ranges of iilluminationn with limitted energy source. Foor low pow
wer consum
mption, the
p
proposed seensor has ann energy soource adaptaation using adaptive power supplyy voltage.
F wide dynamic
For d rannge, the seensor also hhas an illuumination aadaptation which
w has
a
adaptive dynnamic rangee and sensittivity for varrying illumiination condditions.
f
follows:
1. An adaptive C
CMOS imagge sensor w
with reconfigurable m
modes of ooperations:
adapptation
10
2. A CMOS image sensor with embedded feature extraction for motion-triggered
chapter 3, the power consumption of CMOS image sensor is estimated first. From this
estimation, the challenge and strategy of low-power CMOS imager will be discussed.
Chapter 3 also illustrates previous works in three categories that are essential in the
wireless image sensing systems: low-power CMOS image sensors, wide dynamic range
& high-sensitivity image sensors, and image sensors with integrated image processing for
adaptive CMOS image sensor with reconfigurable modes of operations will be explained.
11
Chapter 1 References
12
C
CHAPTER
R 2 INTROD
DUCTION TO CMOS
S IMAGE SENSORS
S
f
final outputt from imagge sensors iss the digitall image signnal that quaantizes the nnumber of
i
incident phootons. An im
mage signal chain is shoown in figuure 2.1. Inciddent photonns generate
e
electrons inn photo-deteectors. The photo-geneerated electrrons are sw
wept by elecctric field,
w
which is ffrom reversse-biased voltage appllied to phooto-detectorrs. In this way, the
p
photocurren
nt (iph) flow
ws in photoo-detectors. The photoocurrent is integrated in photo-
d
detectors annd is converrted into thee voltage VPD. This vooltage is readd out thouggh in-pixel
a
amplifier cirrcuits. Finallly, output voltage
v VOUUT is convertted to the digital signal..
13
Figure 2.2 shows one example of the CMOS image sensor architectures. CMOS
image sensors include pixel array, column-parallel correlated double sampling (CDS)
Row scanner
10-b image
The operation is as follows: (1) incident photons are converted into electrons, and
electrons are accumulated in the photodiode in each pixel during the integration time
(TINT, also called exposure time) (2) in-pixel circuits convert integrated electrons into
voltage. (3) The row scanner scans and selects one row out of pixel arrays. One row of
pixels selected by the row scanner is read out through the column line. (4) CDS circuits
cancel the fixed pattern noise that is inherent in pixel arrays. CDS circuits also amplify
the signal and suppress the input-referred noise. (5) Noise-cancelled signals from CDS
circuits are converted into the digital signals through ADCs. Converted digital signals are
temporarily stored in latches before the readout. (6) Column scanners scan latches
14
column by column, read out digital signals. (7) The procedure repeats for entire rows, and
2.2 Photogeneration
A photon is the carrier of electromagnetic radiation of all wavelengths. It is massless
and is traveling at the speed of light. The energy of photon can be expressed as Eq. 2.1.
hc
E = hν = (2.1)
λ
Silicon is the most widely used material for photo-detectors as well as for
contemporary VLSI circuits. With incident photons, electron-hole pairs (e-h pairs) are
generated. This process is called photogeneration. Separated e-h pairs can be combined
and this process is called recombination. The photogeneration and recombination process
is shown in figure 2.3. Silicon is in the equilibrium state without any incident photons. In
the equilibrium state, the number of electrons in the conduction band (n0) is equal to the
number of holes in the valence band (p0). Electron-hole (e-h) pairs are continuously
generated from the thermal energy. However, generation rate (G) is equal to
When there are incident photons, the equilibrium is broken. When the photon energy
(E=hν) is higher than the bandgap energy (Eg), photons collide with electrons in the
lattice, and electrons absorb the energy of photons to become free electrons. Therefore,
n = n + ∆n, p = p + ∆p (2.2)
15
,where n0 is electron concentrations in the conduction band (@ equilibrium) and p0: hole
concentrations in the valence band (@ equilibrium). Generated e-h pairs are diffused and
recombined within minority carrier lifetime (τ). Recombination rate (R) can be expressed
as Eq. 2.3.
∆n
R= (2.3)
τ
Incident photons into silicon surface travel in silicon. After traveling some distance, they
are finally absorbed. Figure 2.4 shows the photonflux. The photon flux F can be
dF(x)
= −αF(x) → F(x) = F e (2.4)
dx
16
,where F0 is the photon-flux of incident photons (F(0)), α is the absorption coefficient
[cm-1] that is material and wavelength dependent. The inverse of absorption coefficient
α-1 is the penetration depth at which the photon-flux is reduced to e-1 (=0.37). Note that
photons are absorbed after traveling the distance of α-1 (penetration depth). The
penetration depth is dependent on both the material and the wavelength. As shown in
figure 2.4, photons that have the wavelength of blue travel short distance, whereas
photons that have the wavelength of red travel long distance because the penetration
Red
The generation rate of e-h pairs at the distance of x is proportional to the photon flux
d(F − F(x))
G(x) = = αF(x) = αF e (2.5)
dx
17
2 Photodeetectors
2.3
Since phooto-generateed electronss are lost bby the recoombination,, we have to collect
e
electrons inn order to m
measure thee number oof photons. In order too prevent thhe loss of
e
electrons, w
we can just m
move electroons by electtric field andd collect in the collectioon sites as
s
shown in figgure 2.5. Inn order to reaach a detecttable signal level, electtrons are acccumulated
i the capaccitor during the integrattion time. Thhe device thhat providess the electricc field and
in
Figu
ure 2.5 Charrge collection.
2
2.3.1 P-N photodiode
p
A p-n juunction photodiode is the photodeetector thatt is availablle from thee standard
C
CMOS proccess. Figure 2.6 shows tthe p-n juncction photoddiode.
Figu
ure 2.6 P-N
N photodiod
de.
18
The p-n junction is typically formed with n+/pwell or with nwell/p-sub. Photodiodes
operate in the reverse biased condition. In the reverse biased condition, incident photons
p - type n - type
generation
generation
generation
recombination recombination
E-field
Reverse bias
photocurrent
potential
Diffusion
length
Depletion
region
x
LP W LN
d1 d2
Photocurrent
depletion regions. Note that photo-generated electrons in neutral regions are mostly
recombined. In neutral regions, only a small portion of electrons that are generated within
the diffusion length (Lp, Ln) from depletion regions diffuse into depletion regions and
contributes to the photocurrent (Jph(n), Jph(p), diffusion currents). Electrons from depletion
19
Figure 2.8 shows the components of photocurrent. The total photocurrent density can
J =J ( ) +J +J ( )
(2.6)
≈J = −q G(x)dx = qF e −e
photodiodes should be optimized in order to have large depletion region. This can be
achieved by (1) lowering the doping concentration of the bulk (2) shallow junction and (3)
Dark current
Without any illumination, there is a leakage current in the photodiode. This is called
“dark current”. It is typically process dependent variable [2.1]. The dark current cause
shot noise and fixed pattern noise, therefore, it decrease the SNR. The mechanism of the
dark current can be categorized by two: one is the reverse-bias leakage current and the
other is the surface generation current [2.2, 2.3]. Figure 2.9 shows the dark current
20
mechanism. The reverse bias leakage current mostly consists of two components: (1)
generation current from the depletion current (thermal generation), (2) minority carrier
diffusion current in neutral regions. The source of surface generation current is two: (1)
the surface of n and (2) the interface with STI. Carriers are generated in the near vicinity
of a surface via the interaction with interfacial taps due to the discontinuity of the lattice
n
Generation current: J = q W (2.7)
τ
D n
Diffusion current: J =q (2.8)
τ N
qG
Surface generation current: J = (2.9)
2
τ ), W is the depletion width, GS is the surface generation rate (GS = sgni), sg is the
surface generation velocity. The intrinsic concentration ni is shown in Eq. 2.11 and the
/
T
n = N N e /
=A e / (2.11)
300
2ϵ(N + N )(V − V)
W= (2.12)
qN N
21
,where NC, NV are carrier densities, Eg is the bandgap energy, NA, ND are doping
According to Eq. 2.7 ~ 2.10, we can notice the factors affecting the dark current: (1)
the generation current and ni2 term in the diffusion current. (2) Doping concentration: if
doping concentration increases (accordingly more defects), dark current increases due to
the decrease of τG. (3) Reverse bias voltage: more reverse-biased voltage increases dark
The dark current in p-n photodiodes decreases the SNR. This dark current is mostly
from the surface generation, which is generated from the interface defects in the surface
and in the STI region. Moreover, electrons that are generated near the surface can be
easily trapped in the surface states and recombined, which do not contribute to the
By process variations, the surface can be isolated from the interface. Therefore, dark
current can be suppressed. The photodiode which has isolated surface is called “pinned
22
C
Charge inteegration an
nd detection
n
The phooto-current (I
( ph) is integgrated in thee junction capacitance
c of p-n juncctions, and
l
lowers the potential
p off cathode (inn electron-ppotential, electrons are accumulateed and the
p
potential inncreases) ass shown in figure 2.10. Note that the darkk current (IId) is also
i
integrated. Operation procedurees are as follows: ((1) Before the integgration of
p
photocurren
nt, the cathoode of phootodiode is reset (VR) (2) photocuurrent dischharges the
j
junction caapacitance ((CD) duringg the integrration time (TINT). Noote that thee parasitic
c
capacitance (CP) also contributes
c to the sensse node volttage VPD. (33) After TINNT, VPD is
r
read out using in-pixel buffers. (4) Reset againn, next fram
me starts.
F well caapacity
Full
e
electrons arre full of thee capacitancce, electronss are spilledd over. The definition
d o full well
of
c
capacity is tthe maximuum capacityy of charge sstorage of pphotodiodes. A full welll capacity
223
1
N = C (V)dV [electrons] (2.13)
q
C (V) ≈ C = A ( )( )
for p-n junction
/
≈A for p-n+ junction
( )
,where CPD is the photodiode capacitance, Cj is the junction capacitance, VRESET is the
reset voltage and VSWING is maximum voltage swing of the photodiode. Assuming CPD is
C V
N = [electrons]
q
(2.14)
Q =C V [C]
Conversion gain
The conversion gain is defined as the amount of voltage variation from one electron.
q
A =
Capacitance of detection node
(2.15)
q
= [μV/electron]
C
Higher conversion gain means that small amount of photons can induce high voltage
swing. Therefore, higher conversion gain increases the sensitivity, which will be
explained in the next subsection. In p-n photodiodes, the capacitance of detection node is
CPD. It means that higher the conversion gain, the full well capacity is lowered. If we can
separate the detection node from photodiodes, we can achieve both high full well
capacitance and high conversion gain. This will be revisited in the section 2.3.2.
24
Sensitivity
The sensitivity is defined as the ratio of output signal [V] to an incident illumination
level [lx] in a second. The unit of sensitivity is [V/lx·s]. Note that high conversion gain
increase the sensitivity, however, high sensitivity does not guarantee that the pixel has
high conversion gain. The sensitivity can be increased from process optimization,
Fill factor
Pixels consist of photodiodes and in-pixel readout circuits (typically source follower).
The circuit area is shielded by metal layers in order to prevent from the photogeneration.
Only the area of photodiode is exposed. Fill factor is defined as the portion of the pixel
area that contributes to photo-sensitivity, i.e., the ratio of photodiode area to the pixel
A
FF = [%] (2.16)
A
Larger fill factor increases the sensitivity because photodiodes collect more photons.
One way to increase fill factor is shared pixel architecture. In shared pixel architecture,
neighboring pixels share one readout circuit, and improves fill factor. Another way to
achieve high fill factor is the backside illumination (BSI) technology. Using the BSI
technology, photodiodes are exposed to the backside and the metal lines are located in the
front side.
junction, shallow p+ region is formed. This additional p+ implant isolates the n-region
25
from the surface and also from the STI region in order to suppress the dark current.
Moreover, it offers better blue response compared with p-n photodiodes since the surface
is isolated [2.2]. Note that n-region is fully depleted in order to collect more photons.
In this structure, the profile has a potential well due the p+-n-p structure. Therefore,
photogenerated electrons can be collected in the potential well. The minimum potential of
the well is called “pinning potential” which is generated from the profile of p+-n-p
structure. Another advantage from pinned photodiodes is that we can completely transfer
the collected electrons into the sense node without residual charges. Since the pinning
potential is higher than the potential of sense node (which is reset to the power supply
voltage), electrons in the well are completely transferred to the sense node with the aid of
has two types: p-n photodiodes and pinned photodiodes. In-pixel circuit drives the
column line that has large parasitic capacitance. A source follower is typically used as an
26
in-pixel readout circuit. According to the number of transistors in pixel, the pixel is
shown in the figure) is connected with column line (SIG) and is shared by all rows. The
The operation procedure is described with electron potential diagrams in figure 2.13.
The operation procedure is as follows: (1) Reset: the PD node is reset to VDD – VTHR due
to the VT drop in the reset transistor. In the reset operation, kTC reset noise (VkTC1) that is
thermal noise generated in the R-C circuit is added. (2) Integration: VPD is decreased by
integration of photocurrents. ‘Q’ is the integrated charge. (3) Signal readout: after
integration time (TINT), VPD is read out through source followers. (4) Reset: reset PD
node again for the next frame. (5) Reset level readout: this phase is for the correlated
27
Figure 2.13 Operation of 3-T pixel.
The FPN is mainly from the VT variation (in reset transistors and in source follower
transistors) which is different from each pixel. By subtracting the reset level from the
signal level, VTHR term is removed. In this phase, another kTC noise is added (VkTC2). (6)
Double sampling: VTHR term is removed. However, VkTC is not cancelled from the double
sampling because two kTC noise are not correlated each other, i.e., these two kTC noise
are added in different timing. Since the kTC noise power can be expressed as kT/C, 3-T
Moreover, as explained in the previous section, full well capacity and conversion gain
are in trade-off relationships, i.e., larger well capacity induces lower conversion gain.
Therefore, it is desirable to separate the detection node from photodiodes. The 4-T pixel
28
separates the detection node by using an additional transistor between the photodiode and
detection node which is called floating diffusion ‘FD’ is separated from photodiodes. The
operation procedure is shown in figure 2.15. Assuming that FD node is reset before the
integration, the operation is as follows: (1) Integration (2) Reset (FD): before the signal
readout, reset FD node. The kTC noise is added (VkTC1) (3) Reset level readout: read out
the reset level for noise cancelling (4) Charge transfer: charges in photodiodes are
transferred into the floating diffusion (TX is on). As explained before, pinning potential
is higher than FD node, charges are fully transferred. (5) Signal readout: note that the
signal has same kTC noise as in the reset phase (VkTC1) since charges are dumped into the
floating diffusion which already has kTC noise VkTC1. (6) Correlated double sampling:
the CDS operation cancels both the FPN and the kTC noise.
29
Figure 2.15 Operation of 4-T pixel.
Since kTC noise is cancelled in 4-T pixel, a 4-T pixel architecture offers better SNR
compared with a 3-T pixel. Moreover, conversion gain of 4-T pixel (AC(4T) = q/CFD) is
higher than 3-T pixel (AC(3T) = q/CPD). Note that the capacitance of FD (CFD) consists of
the gate capacitance of one transistor and the junction capacitance of two transistors. The
capacitance of FD is small (typically < 5 fF) for the high conversion gain.
of the pixel area and decreases the sensitivity in a given pixel size. In order to increase
the fill factor, neighboring pixels can share one readout circuit. This is called shared-pixel
architecture [2.13-2.16]. Figure 2.16 shows 4-shared pixel circuit which has one readout
circuit shared by four vertically neighboring pixels. In the shared pixel architecture, two
30
or four neighboring pixels share one floating diffusion (FD) node, one reset transistor and
In addition to small pixel size, we can easily perform the pixel merging in the shared
pixel architecture. The pixel merging is the process to add signals from multiple pixels
and provides better sensitivity and SNR (especially at low illumination) while reducing
the spatial resolution. In the shared pixel architecture, integrated charges from
Contemporary CMOS imager typically uses two or four shared architecture. If more
pixels are shared, we can achieve smaller pixel size with high fill factor. However, the
31
capacitance of FD node induces decreased conversion gain. Moreover, the leakage
current from the other photodiodes in a shared group can corrupt the signal especially at
high illumination. Asymmetry of the layout is another issue in the shared pixel
architecture because only one readout path exists in multiple pixels. Due to the
fixed pattern noise (VFPN) and the kTC noise (VkTC). The fixed pattern noise is mainly
from the variation of threshold voltage in in-pixel readout circuits. The kTC noise is the
thermal noise in the RC circuit that is generated from the reset transistor and the sense
node capacitance. In order to cancel the noise (VN = VFPN + VkTC), CDS circuits sample
reset voltage (VRST + VN) and signal voltage (VSIG + VN) successively, and subtract each
other. One simple CDS circuit is shown in figure 2.17. The circuit consists of an
32
The operation is two-step procedure. The first step is the sampling phase. Unity
feedback is formed by turning on the switch ‘F’. While unity feedback is formed, the
reset voltage (VRST + VN) is sampled on C1. In this phase, charge in each capacitor can be
expressed as follows:
Q = C1{V – (V + V )}
(2.17)
Q = 0
The second step is the amplification phase. Unity feedback is removed by turning off
‘F’. After the unity feedback is removed, the signal voltage from the pixel is input to C1.
Since the ‘-‘ input of the operational amplifier tries to keep the voltage VREF due to the
feedback through C2, the charge variation from the new input VSIG + VN is totally
delivered into C2. Now the charge in each capacitor can be expressed as follows:
Q = C {V – (V + V )}
(2.18)
Q = C (V −V )
C
V =V + (V −V ) (2.19)
C
CDS circuits generate noise cancelled signal (VRST - VSIG). CDS circuits can be
programmable gain amplification) [2.3]. Using the amplification by the gain A, the input-
referred noise will be suppressed as V /A where VN2 is the noise in the signal chain
after the CDS circuit (mostly from the ADC). The variable capacitance can be
33
implemented with configurable capacitance of C2. For example, C2 can be implemented
with
A = 1: C2 = C1 (S2S1S0 = “111”)
The gain A is controlled globally, i.e., same gain is applied to entire pixel array. The
amplified signals are input to the ADC, and this amplification suppresses the input
referred noise.
The CDS circuit shown in figure 2.17 induces another FPN. The operational amplifier
has the offset voltage from the mismatch of differential pair, i.e., there is voltage
difference (VOFFSET) to make each current in the differential pair same. When unity
feedback is formed, the ‘-‘ input voltage will be VREF + VOFFSET. Since the CDS circuit is
typically implemented with column parallel, each column generates different VOFFSET and
the output image will have vertical stripe pattern. This FPN can be cancelled by 2nd CDS
in the ADC, i.e., ADC samples both VREF + VOFFSET (when CDS is in sampling phase)
and VREF + VOFFSET + VSIG and cancels each other. The result of 2nd CDS is desired signal
VSIG.
34
2.6 Analog-to-digital converter (ADC)
CMOS image sensors have integrated ADCs for the digital signal output. As shown in
figure 2.18, there are two schemes to implement ADCs: serial ADCs and column-parallel
ADCs.
to the line buffer, which is implemented with sample and hold circuit. The signals stored
in the line buffer are accessed serially by the column scanner. Therefore, high-speed
ADCs such as pipeline ADCs are required [2.12]. As the pixel array is scaled to the large
10-b image
successive approximation (SAR) ADCs and single-slope (SS) ADCs are widely used for
column-parallel ADCs [2.4]-[2.9]. Among them, SS ADCs are mostly widely used in
35
Figure 2.19 shows the architecture and the operation principle of the SS ADC. Input
voltage (VIN) is compared with the ramp signal (RAMP). When ramp signal crosses the
input voltage, the comparator output toggles and latches the counter value into the latch.
A single-slope ADC takes relatively small area because it requires only one
comparator and latches. Moreover, it provides better linearity than cyclic or SAR ADC
because the DAC output (ramp signal) is monotonously increased. The matching between
columns is the most critical factor in column parallel ADCs. The SS ADC provides good
matching because the ramp signal is globally applied to entire columns and the gain
variation of the comparator is not so critical as long as the comparator has enough gain.
The main drawback of SS ADCs is long A/D conversion time. For n-bit conversion, SS
ADCs have 2n clock cycles whereas SAR and cyclic ADC have n clock cycles.
36
Chapter 2 References
[2.1] H. I. Kwon, I. M. Kang, B.-G. Park, J. D. Lee, and S. S. Park, “The Analysis of
Dark Signals in the CMOS APS Imagers From the Characterization of Test
Structures,” IEEE Transactions on Electron Devices, vol. 51, no. 2, pp. 178-184,
Feb. 2004.
[2.2] N. V. Loukianova, H. O. Folkerts, J. P. V. Maas, D. W. E. Verbugt, A. J. Mierop,
W. Hoekstra, E. Roks, A. J. P. Theuwissen, “Leakage current modeling of test
structures for characterization of dark current in CMOS image sensors. IEEE
Transactions on Electron Devices, 50(1), pp. 77-83. 2003
[2.3] J. Nakamura, Image Sensors and Signal Processing for Digital Still Cameras,
Taylor & Francis
[2.4] K. Mabuchi et al., “CMOS image sensors comprised of floating diffusion driving
pixels with buried photodiode,” IEEE Journal of Solid-State Circuits, vol. 39, no.
12, pp. 2408-2416, Dec. 2004.
[2.5] N. Kawai and S. Kawahito, “Noise Analysis of High-Gain, Low-Noise Column
Readout Circuits for CMOS Image Sensors,” IEEE Transactions on Electron
Devices, vol. 51, no. 2, pp. 185-194, Feb. 2004.
[2.6] S. Kawahito, J.-H. Park, K. Isobe, S. Shafie, T. Iida, and T. Mizota, “A CMOS
Image Sensor Integrating Column-Parallel Cyclic ADCs with On-Chip Digital
Error Correction Circuits,” IEEE International Solid-State Circuits Conference
Digest of Technical Papers, pp. 56-595, Feb. 2008.
[2.7] M.-W. Seo et al., “An 80μV<inf>rms</inf>-temporal-noise 82dB-dynamic-range
CMOS Image Sensor with a 13-to-19b variable-resolution column-parallel
folding-integration/cyclic ADC,” IEEE International Solid-State Circuits
Conference Digest of Technical Papers, pp. 400-402, Feb. 2011.
[2.8] I. Takayanagi et al., “A 1.25-inch 60-frames/s 8.3-M-pixel digital-output CMOS
image sensor,” IEEE Journal of Solid-State Circuits, vol. 40, no. 11, pp. 2305-
2314, Nov. 2005.
[2.9] B. Pain and E. R. Fossum, “CMOS active pixel sensor with on-chip successive
approximation analog-to-digital converter,” IEEE Transactions on Electron
Devices, vol. 44, no. 10, pp. 1759-1763, 1997.
37
[2.10] M. F. Snoeij, P. Donegan, A. J. P. Theuwissen, K. A. A. Makinwa, and J. H.
Huijsing, “A CMOS Image Sensor with a Column-Level Multiple-Ramp Single-
Slope ADC,” IEEE International Solid-State Circuits Conference. Digest of
Technical Papers, pp. 506-618, Feb. 2007.
[2.11] S. Yoshihara et al., “A 1/1.8-inch 6.4MPixel 60 frames/s CMOS Image Sensor
with Seamless Mode Change,” IEEE International Solid State Circuits
Conference - Digest of Technical Papers, pp. 1984-1993, Feb. 2006..
[2.12] K. –B. Cho, C. Lee, S. Eikedal, A. Baum, J. Jiang, C. Xu, X. Fan, “A 1/2.5 inch
8.1Mpixel CMOS Image Sensor for Digital Cameras,” IEEE International Solid-
State Circuits Conference Digest of Technical Papers, pp. 508-618, Feb. 2007.
[2.13] M. Mori, M. Katsuno, S. Kasuga, T. Murata, and T. Yamaguchi, “A 1/4in 2M
pixel CMOS image sensor with 1.75 transistor/pixel,” IEEE International Solid-
State Circuits Conference Digest of Technical Papers, pp. 110-111, Feb. 2004.
[2.14] Y. C. Kim, Y. T. Kim, S. H. Choi, H. K. Kong, S. I. Hwang, J. H. Ko, B. S. Kim,
T. Asaba, S. H. Lim, J. S. Hahn, J. H. Im, T. S. Oh, D. M. Yi, J. M. Lee, W. P.
Yang, J. C. Ahn, E. S. Jung, Y. H. Lee, “1/2-inch 7.2MPixel CMOS Image
Sensor with 2.25/spl mu/m Pixels Using 4-Shared Pixel Structure for Pixel-Level
Summation,” IEEE International Solid State Circuits Conference Digest of
Technical Papers, pp. 1994-2003, Feb. 2006.
[2.15] N. Tanaka, J. Naruse, A. Mori, R. Okamoto, H. Yamashita, and M. Monoi, “A
1/2.5-inch 8Mpixel CMOS image sensor with a staggered shared-pixel
architecture and an FD-boost operation,” IEEE International Solid-State Circuits
Conference Digest of Technical Papers, pp. 44-45, Feb. 2009.
[2.16] H. Takahashi, T. Noda, T. Matsuda, T. Watanabe, M. Shinohara, T. Endo, S.
Takimoto, R. Mishima, S. Nishimura, K. Sakurai, H. Yuzurihara, S. Inoue, “A
1/2.7-in 2.96 MPixel CMOS Image Sensor With Double CDS Architecture for
Full High-Definition Camcorders,” IEEE Journal of Solid-State Circuits, vol. 42,
no. 12, pp. 2960-2967, Dec. 2007.
[2.17] R. D. McGrath, H. Fujita, R. M. Guidash, T. J. Kenney, W. Xu, “Shared pixels
for CMOS image sensor arrays,” IEEE International Image Sensor Workshop, pp.
9-12, Jun. 2005.
38
CHAPTER 3 LOW POWER CMOS IMAGE SENSORS, DYNAMIC RANGE
ENHANCEMENT AND EMBEDDED IMAGE SIGNAL PROCESSING
In this chapter, power consumption of CMOS image sensors will be estimated and
power reduction strategy will be discussed. Then three categories of CMOS image
sensors, which are essential for distributed imaging systems and biomedical sensors: (1)
low-power CMOS image sensors, (2) wide-dynamic range & high-sensitivity CMOS
image sensors and (3) CMOS image sensors with integrated image signal processing for
bandwidth reduction. This chapter will introduce principles, challenges and previous
works.
Since CMOS image sensors can be fabricated with CMOS process, power
consumption can be scaled down compared with CCD image sensors. However,
conventional CMOS image sensors still have large power consumption and they are
conventional image sensors are mostly 2nd generation imaging systems such as mobile
devices. The SNR and the image quality have been a critical factors rather than power
consumption. Moreover, voltage scaling has a limitation due to the signal range of the
pixel.
In CMOS image sensors, a variety of circuits for image enhancement and processing
range of conventional CMOS image sensors is around 60 dB, whereas the natural scene
has more than 100 dB dynamic range. Therefore, high-sensitivity readout circuits and
39
dynamic range enhancement schemes should be integrated in order to provide a reliable
monitoring.
A bandwidth is another critical factor because the image signal transmission with
constant frame rate requires huge bandwidth. For bandwidth saving, lightweight image
processing schemes such as the motion sensing or the region-of-interest (ROI) readout
has been integrated in CMOS image sensors. These schemes process the signal from
the typical approach, separate power supply between analog and digital block are used in
VDD
Pixel array
Column scanner
40
For the estimation, the analog power supply is further divided by two parts: pixel
(VDDP) and CDS/ADC circuit (VDDA). The digital power supply can be further divided by
two parts: the core including timing controllers and ADC digital circuits (VDD) and I/O
circuits (VDDIO). In this section, power consumption in each block of CMOS imager will
be estimated.
P =I ∙V (3.1)
,where IAVG is the average current and VDD is the power supply voltage.
In imaging operation, most of static power consumption comes from the analog
circuits including pixel arrays, CDS circuits, and also some part of ADCs (mostly from
the comparators). These circuits repeat the operation row by row in rolling-shutter
operation. The average current IAVG can be calculated with the bias current and the duty
cycle as follows:
# ∙T T
I =I ∙ ≅I ∙ (3.2)
T T
,where TON is the time that the circuit is activated. The row access time TROW can be
defined as the time of one-row operation including the pixel readout, CDS (TCDS), ADC
(TADC), digital signal readout (TRD) and blanking time before the next row operation.
Using Eq.3.3, the static power consumption of each circuit can be expressed as
follows:
T
P =I V (3.4)
T
The dynamic power consumption in digital circuits consists of two components: one
is from the load capacitor driving and the other is from the short circuit current during the
P =C V f+ t V I f
= C V f+ C V f (3.5)
= (C + C )V f
The first term shows the dynamic power from capacitor driving. CL is the load
capacitance of the gate and f is the frequency of transitions from ‘0’ to ‘1’. The second
term shows the dynamic power consumption from the short circuit current. tsc represents
the time of short circuit current generation, i.e., both pMOS and nMOS are turned on
during the transition. IPEAK is the peak short circuit current. In CMOS imagers, many
digital circuits have to drive long metal line connected with each column circuits. In this
case, total load capacitance (CL) consists of load capacitance of each column circuit (cl(col))
42
Eq. 3.6 C = # c( ) + Lc
(3.6)
≈# c( )+W c
,where #COL is the number of columns, L is the length of metal line, and cm is the parasitic
capacitance per unit length. The length L can be expressed as #COLWPIX where WPIX is the
pixel pitch. Some of digital circuits such as the counter in single slope ADCs and the
clock generator operate with high frequency. Therefore, dynamic power consumption in
CMOS imagers is dominated by the power consumption from the capacitor driving.
When circuits are not operating, there is power consumption from the leakage current.
The leakage current mainly comes from three components: sub-threshold (sub-VT)
current, junction leakage, and gate oxide tunneling [3.2]. In CMOS imagers, the leakage
power consumption is more severe in digital circuits compared with analog circuits
because digital circuits use thin-oxide transistor with low VT for voltage scaling whereas
analog circuits including pixels use thick-oxide transistor which allows higher voltage
I = Ae 1−e (3.7)
body effect coefficient. As long as drain-source voltage (VDS) is not zero, the sub-VT
current flows even when the gate-source voltage VGS is zero. Note that low VT
43
significantly increases the sub-VT leakage in the exponential form. In short channel
devices in digital circuits, the DIBL which lowers VT increase the sub-VT current.
The junction leakage is from the reverse-biased p-n junctions. The most significant
factor of junction leakage is from the band-to-band tunneling (BTBT). Due to high
electric field from revere biased p-n junctions, electrons tunnel from the valence band (p
The gate leakage current is from the gate oxide tunneling. High electric field across
the SiO2 makes electrons in the inverted silicon surface tunnel directly to the gate. This
Typically, leakage power consumption is much lower and is not comparable to static
low-power imagers operate in the sub-VT region with scaled VDD. In this case, leakage
power is an important factor in order to achieve minimum energy [3.3]. The low-power
CMOS imagers using sub-VT readout circuits will be introduced in section 3.2.
estimated. A signal chain of CMOS imagers is shown in figure 3.2. The operation timing
diagram is shown in figure 3.3. The main parameters for power estimation are shown in
table 3.1. In the estimation of dynamic power consumption, we consider the dynamic
power consumption only from the capacitive driving (CLVDD2f) for simplicity. Moreover,
the leakage power consumption will be ignored. This estimation causes error compared
with actual dynamic power consumption. However, main purpose of power estimation in
this section is to get power reduction strategy by inspecting the tendency of power
44
consumption according to critical parameters such as the number of pixel array, power
supply voltage and pixel pitch. In this section, we will show only estimated equation
without detailed derivation for digital circuit blocks. Detailed estimation of dynamic
Pixel Array
The circuit schematic of pixels in one column is shown in figure 3.4. The load
transistor, column line (SIG) parasitic capacitance and sample and hold capacitance (CS).
The sample and hold capacitor is a part of CDS circuits. The load capacitance of in-pixel
C =# c +# W c +C (3.8)
g 2βI
GB = f = = (3.9)
2πC 2πC
45
RST 0 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
TX 0 0 0 0 0 0 0 0 0 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
SEL 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
F 0 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
SS 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
SR 0 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
Counter CLK 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 1 0 1 0 1 0 1 0 1 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
PCH 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 1 0 1 0 1 0 1 0 1 0 1 0 0 0 0
46
SIG
Figure 3.4 Pixel circuit with load capacitance of in-pixel source follower
For 1% settling, the 3-dB bandwidth should be 4.6 times larger than the pole
frequency. Therefore, we assume that the time constant is 5 times smaller than the
required time slot (T). Then the 3-dB bandwidth of in-pixel source follower can be shown
as Eq.3.10.
5 2βI
f = = (3.10)
0.5 ∙ T 2πC
In Eq. 3.10, the sampling time is set as 0.5TCDS because two signals have to be
sampled for the correlated double sampling. The source follower needs to be turned on
during the CDS time. Considering the duty cycle (TCDS/TROW), the power consumption of
T
Eq. 3.11 P =# I V
T
(3.11)
(20πC ) T
=# V
2β ∙ T T
47
The CDS circuit consists of the programmable gain amplifier (PGA) with capacitive
feedback as shown in figure 3.5. The PGA consists of one OTA, one input sampling
capacitor (CS) and programmable feedback capacitors (CF). In figure 3.5, the gain of
PGA (APGA) can be set up to 8 as an example. The 3-dB frequency of the CDS circuit can
be described as:
5 1 2β(0.5I )
f = = ∙ (3.12)
0.5 ∙ T A 2π(2C )
T
P =# I V
T
(3.13)
(40A πC ) T
=# V
β∙T T
The single slope ADC consists of one comparator and latches. The latch can be
48
consumption of the single-slope ADC is dependent on the topology of the comparator
and the latch. In this estimation, one common topology that consists of a static
figure 3.6. The clock frequency of counter in the single-slope ADC is determined by the
number of pixels and also by the frame rate. The A/D conversion time (TADC) can be
determined as Eq.3.14.
T =T −T −T (3.14)
Note that there is an additional time budget for the digital signal readout (TREAD)
because no additional line memories to store one row of image signals. With line
memories, the ADC output will be stored in the line memory and the signal readout can
Preamplifier Comparator
VCOM
VRAMP
VDDA VDD
F
SS CS CC
10-b
VO(CDS)
latch
SR CS CC
F
IADC
VCOM
The A/D conversion time can be expressed with required clock period (TCLK) and the
49
1 2 +#
f = = (3.15)
T 1
−T
FR ∙ #
A. Comparator: Preamplifier
comparators by providing a high gain. The offset voltage of the preamplifier can be
suppressed by another CDS operation in the ADC. Therefore, the gain of the preamplifier
and the offset voltage of the dynamic comparator is 10 mV, the gain of the preamplifier
(APA) over 10 guarantees the offset error under 1 LSB. In this estimation, the differential
desirable to design the delay of a preamplifier under TCLK in order to latch the counter
signals within 1 LSB range, i.e., keep the bandwidth of the preamplifier over fCLK. The 3-
1 g
f =f = ∙
A 2πC
(3.16)
1 2β(0.5I )
= ∙
A 2πC
The load capacitance (CL) of the preamplifier includes the gate capacitance of load
transistors and the gate capacitance of the dynamic comparator. Assuming total load
capacitance (CL) is 16cg, the power consumption of the preamplifier can be calculated as:
50
32A πf c T +T
P =# V (3.17)
β T
12cg
4cg
Vinp Vinn
IADC
operates with two phases: the precharge and the comparison. Figure 3.8 shows the
comparator circuit. The first stage that consists of a latch circuit, switches (pMOS for
precharge), and inverters contributes to the power consumption in each clock cycle. The
2nd stage that consists of a RS latch and buffers has only one time switching during the
ADC time. For simplicity, we consider only the gate capacitance for total load
capacitances. The power consumption per one comparator (per column) can be expressed
as Eq.3.18.
2 1
P = 6c V + 12c V (3.18)
T T
51
Figure 3.8 Dynamic comparator
The latch for the N-bit single-slope ADC consists of N D flip-flops. The circuit
schematic of the D flip-flop is shown in figure 3.9. Note that the clock signal (CK) is
from the comparator output (Q) and the input signal (D) is from the counter. The
comparator output has the frequency 1/TROW and the counter output has different
frequency according to the bit position. The power consumption of one latch in the
1 f ( )
P _ = 13c +c V + 5c V (3.19)
T T
,where fCNT(n) is the counter output frequency of n-th bit position. As shown in Eq.3.19,
the power consumption of two inverters in the master stage of the latch is dependent on
the input frequency, i.e., counter output frequency fCNT(n). In the single-slope ADC, when
the clock signal is low, the master stage continuously changes its state from the counter
signals. After the clock signal goes high and the counter signal is latched to the slave
stage, no more switching occurs. The power consumption is dependent on the output
52
toggling from the comparator. Therefore, the power consumption is largely dependent on
the signal level (i.e., light intensity) of single slope ADCs. In this estimation, we define
the illumination level parameter ‘I’ which varies 1/2N (dark) to 1 (bright). For The power
consumption of the N-bit latch in single slope ADCs which uses a Gray counter can be
expressed as follows:
P =# N 13c +c + 5c I(1 + 1 + 2
1 (3.20)
+ ⋯+ 2 )V
T
Note that the 2nd term includes the illumination level parameter ‘I’. If the signal level is
high from high illumination, the power consumption increases because the latch has more
Compared with a binary counter, a Gray counter provides better noise immunity
because the output change occurs only in one bit position, i.e., timing jitter induces only 1
LSB error. Another advantage is low power consumption because total number of
53
switching is half of the binary counter. Detailed power estimation will be shown in
appendix A. The overall power consumption from the counter can be expressed as:
P = (39c + 0.5# W c )
(3.21)
1 + 2 + ⋯+ 2
×V
T
Figure 3.10 shows the digital signal readout circuit and the operation. The output of
the 10-b latch drives the readout buffer. The readout buffer of all columns share one
sense line (SL). The sense line is precharged with VDD first. The column selection signal
CS from the column scanner is enabled column by column. When stored digital signal is
‘1’, the readout buffer draws the current from the precharged sense line and the voltage of
the sense line is dropped. The sense amplifier senses the voltage drop of the sense line.
Sense amp
PCH
Sense line (SL)
Q
CS CPAR
qCOL(n) VREF
(b) Operation
54
The power consumption occurs from two parts: precharge switches and sense
amplifiers. The power consumption from the precharge switches is shown in Eq. 3.22.
#
P = N# W c (3.22)
T
In Eq. 3.22, we assumed VDD swing of the sense line. We also assumed the worst case
and all sense line voltage drops. Using the same dynamic comparator as in the single-
#
P = 18c V (3.23)
T
The total power consumption from the digital signal readout is:
#
P = {N# W c + 18c }V (3.24)
T
Periphery circuits
A. Column scanner
In this estimation, a simple shift register instead of a decoder will be used for both
the column and the row scanner. The column scanner requires #COL shifting operation per
one row time. The detailed power estimation will be shown in appendix A. The total
#
P = 12.5c + 0.5Nc V (3.25)
T
B. Row scanner
55
The row scanner consists of a shift register and drivers as shown in figure 3.11. The
row scanner requires #ROW shifting operation per one frame. In each row, the row scanner
consists of one flip flop, three NAND gates, three level-up converters and three inverters.
The NAND gate enables the pixel control signals (TX, R, S) according to the control
signals (TXG, RG, SG). Since the pixel power supply is higher than digital power supply
(VDD), level-up converter is required to control the pixel circuit. The total power
1
P = {7c # + 20c }V
T
(3.26)
1
+ 6c +# W c +c V
T
q(n-1) D Q TX
TXG
Q Q
CLK
R
RG
thin-ox
S
SG
C. Timing generator
Simple timing generator consists of two counters: one is the horizontal counter (H-
counter, assume N+1 bit counter) for column-wise operations and the other is the vertical
counter (V-counter) for row-wise operations. Additional logic gates and registers are
included for the signal generation. For simplicity, only the power consumption of two
56
counters will be estimated. The power consumption of binary ripple counter has shown in
1
P = 22c (1 + 2 + ⋯ + 2 ) V (3.27)
T
1
P = 22c (1 + 2 + ⋯ + 2 ) V (3.28)
T
,where X is log # .
From Eq.3.27 and 3.28, the total power consumption from the timing generator is
1
P ≈ 22c (1 + 2 + ⋯ + 2 ) V
T
(3.29)
1
+22c (1 + 2 + ⋯ + 2 ) V
T
I/O circuit
IO circuit includes the ESD protection circuit and drivers. Typically the strength of
driving can be controlled according to the register values. Accordingly, the power
consumption varies from the driving capability. In this estimation, the driver that has
simple two-stage inverter chain will be estimate. The most power-consuming I/O pad is
the clock signal pad, and the next is digital output pad. The power consumption from the
clock signal and digital output pad are shown in Eq. 3.30 and Eq. 3.31 respectively.
P/ ( ) = (16c +C )f V (3.30)
57
P/ ( ) = (16c +C )(1 + 2 + ⋯
1 (33.31)
+2 ) V
T
i Eq. 3.32.
in
P ≈ (16c +C )f V
(33.32)
1
+(16c +C )(1 + 2 + ⋯ + 2 ) V
T
3
3.1.3 Overaall Power C
Consumptioon
Using tthe power consumptioon equationns derived iin the previious sectionn, we can
e
estimate thee overall pow
wer consum
mption. Figuure 3.12 shoows the pow
wer consumpption from
e
each power supply volttage.
558
wer consum
The pow mption from
m the I/O cirrcuits takes significantt portion in the small
n
number of ppixel arrayss (320 × 2440) due to thhe large cappacitance of I/O circuiits. As the
n
number of pixel increeases, the pparasitic cappacitance of
o metal linnes and thee column-
p
parallel circuits increaase, therefoore, relativee portion oof the I/O power connsumption
d
decreases. F
Figure 3.133 and 3.14 shows detaailed powerr consumpttion from thhe analog
p
power suppply (VDDA). In the analoog power coonsumptionn, the portioon of the AD
DC power
i
increases inn the large pixel arrays. This is becaause the CD
DS time (TCDDS) is fixed according
t the chargge transfer time in the ppixel. Moreover, the looad capacitaance of CDS
to S circuit is
f
fixed with ssampling caapacitance. T
The only facctor related with the nuumber of pixxels is the
r
row TROW) as shhown in Eqq. 3.13. In the ADC ppower, the cclock frequuency fCLK
time (T
i
increases acccording to the
t number of pixels ass shown in E
Eq.3.15.
559
Figure 3.114 Digital p
power consuumption in the signal chain accorrding to thee spatial
resolution (a) 3320 × 240 (b
b) 1280 × 9660
A the num
As mber of pixeel increases,, the portionn of power consumptioon from diggital signal
r
readout incrreases. Thiss is because the power consumptioon from the digital signnal readout
(
(number of columns), load
l capacittance, and frrequency.
F
Figure mption according to th
3.155 Digital poower consum he power su
upply voltaage (320 ×
240)
660
Figgure 3.16 O
Overall pow
wer consumption accorrding to thee pixel pitch
h
(a) 320 × 240
2 (b) 12880 × 960
c
consumptio n. As expeccted, 1/2 scaling induces around 11/4 power reeduction. Fiigure 3.16
s
shows the ppower consuumption acccording to the
t pixel pittch. As pixeel pitch incrreases, the
p
pixel powerr consumption increases due to thee increase off signal line parasitic caapacitance.
3
3.1.4 Low-p
power desiggn strategyy
From thhe power estimation in previous seections, we can get sevveral concluusions and
p
power reducction strateggies for low-power CM
MOS image sensors.
s
( Voltagee scaling: obbviously wee have to usse voltage scaling in booth analog aand digital
(1)
b
blocks. In the
t case of ddigital circuuits, the volltage scalingg is effectivve because tthe power
c
consumptio n is dependdent on VDDD2. Since thhe operationn frequency is below 10 MHz in
l
low resoluttion CMOS < 320 × 2440), which is applicabble to the low-power
S imager (<
w
wireless sennsor node, we
w can easilly achieve ssub-1V pow
wer supply. However,
H thhe voltage
s
scaling of thhe pixel andd the analogg power suppply directlyy affects thee SNR due tto reduced
s
signal mplement altternative pixel circuits and CDS
rangees. Thereforre, it is desiirable to im
c
circuits whiich are less vulnerable
v tto the voltagge scaling. IIn chapter 4,
4 a low-pow
wer imager
661
employing the reconfigurable pixel architecture with reduced power supply will be
discussed.
reducing unnecessary switching activities. For example, the latch circuit in the single-
slope ADC continuously changes the state before the comparator output is toggled.
Another example is the precharging circuit. Whenever the digital signal has ‘0’ value, the
sense line voltage drops and dynamic current is drawn in the precharge operation. In
(3) Bandwidth reduction: The image sensor has high spatial-temporal resolutions and
the I/O power takes more than 50% of total power consumption as shown in figure 3.12.
(4) Small pixel size: As shown in figure 3.16, the power consumption increases
according to the pixel size due to increased parasitic capacitance. In order to achieve low-
power readout and an additional functionality, many low-power image sensors use
additional in-pixel circuits, which increase the pixel size. In chapter 4 and chapter 5, a
proposed low-power readout scheme with small pixel size will be explained.
The power consumption is not increased linearly according to the resolution because
of periphery circuits. However, the tendency is almost linear because the portion of
power consumption from the periphery circuit is not so high. One of way to show the
62
Total power consu umption pW
FOM
M= [ ]
# of pixxels × Fram
me rate p pixels fram
me
p
power FOM
M does not reflect
r any SNR degraadation, we will use addditional FO
OM, which
r
reflects SNR
R in chapterr 4.
i
image sensoors. As shoown in the ttable, the ppower consuumption of conventionnal CMOS
663
imagers is over 30 mW in most cases. The applications of these sensors are mostly digital
still cameras, cell phones, and automobiles. For the application of distributes sensors
conventional sensors is too large to guarantee long lifetime of sensor nodes. Table 3.3
shows the power delivery from the energy harvesting. According to this table, in an
overcast day (1,000 lx illumination), available power from the 3 × 3 mm2 solar cell (150
pW/mm2·lx) is only 1.35 μW. This power delivery requires 135 pW/pixel·frame FOM
with 100 × 100 image sensor (@ 1 fps). This power budget is definitely small for
However, the voltage scaling has limitations due to the SNR degradation. Figure 3.18
shows the signal ranges in the 4-T pixel architecture. The maximum and the minimum
signal in the source follower output and in the FD can be expressed as follows:
(3.33)
(a) VOUT(MAX) = (VDD – VTHR) – VTH : reset transistor’s and source follower’s VT drop
From the equation 3.33(c) and 3.33(d), the maximum signal swing in the FD is
64
Vendor Illuminance Output Power Size Power / mm2·lx
3000 mm2
200 lx 80 uW 133.3 pW
(estimated)
Cymbet [3.8]
3000 mm2
1000 lx 350 uW 116.7 pW
(estimated)
Cymbet [3.8] Sunny 0.08 uW 0.07 mm2 5.7 nW
6000 lx (direct 3.25mm2
Clare [3.9] 200 uW 10.3 nW
sunlight) (estimated)
Enocean [3.10] 200 lx 13.5 uW 448 mm2 150.6 pW
Then, the minimum power supply can be calculated from Eq. 3.35.
TX RST
∆VPPD
VPIN
PPD
∆VFD
VDD
VDD VTHR
FD
SIG
VTH
SEL
VOUT
VB VDSsat
65
As shown in Eq. 3.35 (b), the minimum power supply voltage is “signal swing +
VDSsat + two VT drop”. This means that the voltage scaling reduces the signal swing,
accordingly reduces the SNR. Assuming VT = 0.8 V, VDSsat = 0.2 V, and signal swing
ΔVFD is 1V (for 10-b resolution, 1 mV LSB), then minimum power supply voltage is 2.8
V. Note that contemporary CMOS image sensors use 2.8 V for the pixel & analog power
supply voltage.
Another factor to limit the voltage scaling is the charge injection and the clock feed-
through in the reset transistor as shown in figure 3.19. When the reset transistor is turned
off, the voltage of FD is less than VDD – VTHR due to the charge injection and the clock
feedthrough. The VFD after the reset operation is VDD - VTHR - ΔVCH. Since the
capacitance of FD is very small (< 5 fF), the resultant error ΔVCH will be large (> 100
RST
TX
SIG
FD
SEL
VDD-VTHR
VDD-VTHR-∆VCH
VFD
Figure 3.19 Charge injection and clock feedthrough in the reset transistor
66
Another side effect occurs in the source follower. Since the nMOS transistor in the
pixel is not formed on the separate p-well, the body of nMOS transistor is connected with
ground. This induces a body effect and the gain of the source follower becomes less than
1 as follows:
g 1
A ( ) = = <1 (3.37)
g +g 1+η
Typically the source follower gain is around 0.8 ~ 0.9. Therefore, we can modify the
ΔV
V > +V + 2V + ΔV (3.38)
A ( )
Moreover, the source follower has nonlinearity due to the channel length modulation
of the bias transistor. The output voltage VOUT and the current bias IBIAS has nonlinear
relationship as follows:
This non-linearity distorts the image signal, which is not desirable for the post image
processing. If the power supply voltage is scaled down and enough signal swing is not
guaranteed, the bias transistor falls into the linear region especially when the illumination
Hard reset
67
One VT drop out of two VT drop in the pixel circuit occurs in the reset transistor. In
order to remove the VT drop in the reset transistor, the signal ‘RST’ is boosted from VDD
to VDD + VT using the charge pump circuit as shown in figure 3.20 [3.12]. This operation
makes the reset transistor operates in the linear region, whereas the reset transistor
The reset operation with boosting is called hard reset and the reset operation without
boosting is called soft reset. It is known that the hard reset doubles the kTC noise [3.13].
However, the kTC noise can be suppressed in the 4-T structure. By using the reset signal
ΔV
V > +V + V + ΔV (3.40)
A ( )
Capacitive coupling
Hard reset is an effective way to increase the signal swing with lower power supply
voltage. However, additional boosting circuits that consist of charge pump are required.
The charge pump circuits need large capacitors in order to boost the reset gate (RST) that
68
has large parasitic capacitance. This is not desirable in terms of the area and also the
power consumption.
Another method to increase the reset level is using a capacitive coupling [3.14].
Figure 3.21 shows the pixel circuit with the capacitive boosting. Note that the selection
transistor is not used. For the row selection, the FD node is reset and the source follower
is turned on. Using the soft reset, the FD node is reset with VDD – VT. The source
follower output (SIG) increase slower than the FD node due to the large parasitic
capacitance. Through the gate-source capacitance of the source follower, the increase of
source follower output couples with the FD node, therefore, the reset level is boosted.
69
Another scheme to boost reset level is shown in figure 3.22 [3.7]. In this scheme,
additional metal line (FDB) has been placed. The parasitic capacitance between the FDB
and the FD node operates as a coupling capacitor. When the FDB line is driven high, the
limitations of voltage scaling. Moreover, non-linearity and gain loss further limit the
voltage scaling. Instead of using the source follower, an in-pixel comparator can be used
for the readout by performing the single-slope ADC operation, which is one kind of pulse
width modulation (PWM). Figure 3.23 shows the concept of in-pixel PWM. One input of
the comparator is connected with the detection node. The other input is driven by the
ramp signal. The comparator is toggled when the ramp signal crosses the voltage of
detection node, then the counter value is latched. The advantage of the in-pixel PWM
scheme is that there is no gain loss from the body effect. Moreover, it is less vulnerable
from the current bias variation, which induces non-linearity. However, this scheme
70
generally requires larger pixel size due to the in-pixel comparator, which requires pMOS
transistors. Another disadvantage is the input voltage variation during the ADC time
because it requires 2N cycles of operation for the N-bit conversion. In high illuminations,
Figure 3.24 shows one in-pixel PWM scheme [3.15]. In this work, nMOS-only
comparator that has common source amplifier topology is used and no fill factor loss
takes place. The operation is as follows: (1) Reset: the photodiode is reset by forming the
unity feedback. (2) Integration (3) Precharge: before the readout, the column line is
precharged (4) PWM: the ramp signal is applied to the source of the transistor. When the
ramp signal crosses the voltage VPD – VT, the charged column line starts to be discharged.
By regenerating the voltage variation of the column line, counter value is latched.
In this scheme, no static bias current is required except in the reset operation.
voltage scaling. The signal swing in the photodiode can be expressed as follows:
By adjusting the voltage of VRAMP, the maximum signal swing can be controlled.
Since the comparator has the common source topology, there is a current source
connected with VDD. In order to guarantee that current sources operate in the saturation
71
VDD
IPIX
PCH
RR
PCH
PCH
discharge
R S
SIG
VRAMP
Therefore, it is difficult to achieve sub-1V power supply. In this scheme, ramp signal
driving is another critical factor. Since the discharge current flows into the ramp signal
line, it corrupts the ramp signal if it is not buffered. Driving the ramp signal will require
Figure 3.25 shows the in-pixel PWM scheme with a sub-VT comparator [3.16]. By
operating the comparator in sub-VT region, voltage can be scaled down to < 0.5 V and the
power consumption is 1.19 μW for 128 × 128 pixels. However, the comparator uses
pMOS transistors that make the pixel size larger. The comparator requires 3 additional
transistors compared with the 4-T pixel architecture. In order to guarantee the input range
of the comparator, there is an additional diode-connected transistor in the reset path. This
reduces the reset level and suppresses the SNR. Moreover, correlated double sampling is
72
S
VRAMP
R
ENB
73
Figure 3.26 shows another in-pixel PWM scheme with sub-VT comparator [3.17].
Instead of applying voltage ramp signal, this scheme uses current DAC and changes the
threshold voltage of the transistor MCS for the PWM. The transistor MCS is biased by
the current DAC and operates in sub-VT region. The pixel includes only 5 transistors,
however, it includes pMOS transistors that make the pixel size larger.
Source follower
Scheme In-pixel pulse-width-modulation (PWM)
with voltage scaling
Pros Small pixel No gain loss, non-linearity
Sub-VTH in-pixel
Readout circuit Source follower
comparator
nMOS-only comparator
0.42 uW (pixel)
Power consumption 550 uW (total) 1.19 uW (total)
55.2 uW (total)
Power FOM
723.4 pW 8.6 pW 468 pW
[pW/pixel·frame]
SNR 49.63 dB 23.1 dB 52.9 dB
can be incorporated in one image. The DR of natural scenes is more than 100 dB,
whereas the DR of conventional CMOS image sensors is only 60 ~ 70 dB. The definition
74
Saturation level
DR = 20 ∙ log [dB] (3.43)
Minimum detectable level
The minimum detectable signal is limited by noise floor. As shown in Eq. 3.43, in
order to increase the dynamic range, we can increase the saturation level or reduce the
noise floor. When we refer to “dynamic range extension”, it typically means the dynamic
range enhancement schemes, which increase the saturation level in order to prevent the
Reducing the noise floor is common issue for both conventional CMOS image
sensors and WDR CMOS image sensors. In low illuminations, SNR is degraded due to
low signal level. If CMOS image sensors have high sensitivity, i.e., generate higher
signal level with less incident photons, SNR increases due to the increase of signal level.
However, increasing only the sensitivity reduces the dynamic range because the signal
schemes according to the illumination. The sensitivity can be enhanced by several ways.
One is the process optimization such as photodiodes, color filter arrays and microlens.
The other is by the readout circuits. The CDS circuit and the programmable amplifier
This section will explain about the sensitivity enhancement schemes. The integrated
[V] to the incident illumination level [lx] in a second. The unit of sensitivity is [V/lx·s].
75
One of way to increase the sensitivity is applying high gain in order to amplify the signal
and suppress the noise. The programmable amplification in the CDS circuit is a good
choice to increase the sensitivity. However, the PGA requires significant power
enough to guarantee the matching between columns (typically > 1 pF). Moreover, column
parallel PGAs only suppress the noise of ADCs, not the readout circuit noise from in-
pixel source followers. Therefore, there are previous works that employ in-pixel
SIG1
SIG2
Figure 3.27 In-pixel nMOS common source amplifier with p-n photodiode
One previous work used the nMOS common source amplifier with p-n photodiode
[3.18] as shown in figure 3.27. As mentioned, in the 3-T pixel structure, kTC reset noise
is not cancelled. In order to suppress the kTC noise in the 3-T structure, this work used
unity feedback using in-pixel commons amplifiers. In simple one pole amplifiers, the
gain increment by AV induces the bandwidth reduction by AV. Therefore, the noise power
76
from kTC noise is decreased by AV. Assuming the amplifier gain is infinite, the gain of
C
A = (3.44)
C
Figure 3.28 In-pixel pMOS common source amplifier with pinned photodiode
capacitance of the transistor M2. Since CS has the capacitance over 10 fF, and CI is less
than 0.5 fF, high gain > 20 is achievable. However, the gain of common source amplifiers
cannot be an infinite and it is typically around 10, which brings out the reduction of the
Another previous work used pMOS common source amplifiers with pinned
photodiode [3.19]. This work uses similar topology as a conventional 4-T pixel structure
except that they used pMOS common source amplifiers as shown in figure 3.28. In
pinned photodiodes, the reset level should be high enough in the floating diffusion in
order to guarantee complete charge transfer. In order to increase the reset level that is set
77
by the unity feedback, a pMOS amplifier is chosen. In the selection transistor, the voltage
bias VCASC is applied for the cascade configuration. It suppresses the Miller effect and
higher gain is achievable. The input referred noise is suppressed by the gain of amplifier
(gmRi) as follows:
kT 1
v = γ ∙ (3.45)
C g R
,where CS is the sampling capacitor in the column (not shown in the figure) This work
achieved 0.86 e- readout noise whereas the conventional CMOS imagers have > 2 e-
readout noise. However, pMOS transistors increase the pixel size (11 × 11 μm2 in this
work).
In summary, it is desirable to implement the in-pixel amplification for two reasons: (1)
Avoid using power-consuming column-parallel PGAs while increasing the sensitivity (2)
The in-pixel amplification suppresses the noise from pixel readout circuits.
The basic principle of nonlinear scheme is to compress the pixel output response and
cover wider ranges of illumination before the pixel output is saturated. Figure 3.29 (a)
The linear scheme can be categorized as two: sensitivity adjustment and multiple
by reducing the conversion gain. As explained before, the conversion gain is q/CD where
78
with the CD, the conversion gain will be reduced to q/(CD + COV). The additional
capacitance lowers the sensitivity. However, it increases the illumination range that is
able to be sensed. Figure 3.29 (b) shows the basic concept of sensitivity adjustment
scheme.
VA2
VA1
VB2
VB1
t
TINT(SHORT) TINT(LONG)
(c) Linear: Multiple exposure
In the multiple exposure scheme, multiple signals from multiple integration time
(from short to long) are synthesized. Figure 3.29 (c) shows the example of dual exposure.
In the figure, two cases of pixel output response are shown: one is in high illumination
(output is saturated) and the other is in low illumination. By applying the dual exposure,
VA1 and VA2 will be synthesized. Although VA2 is saturated signal in high illumination,
we can get the information from VA1. Likewise, VB1 and VB2 will be synthesized in low
illumination case. In this case, VB1 can be low and therefore has low SNR. However, VB2
79
In the next three subsections, each scheme will be explained with previous works.
implement by the transistor which operates in the sub-VT region. The logarithmic pixel is
shown in figure 3.30. Since the current is exponentially proportional to the gate-source
voltage VGS, VGS is logarithmically dependent on the current. However, since the
transistor operates in the sub-VT region, it suffers high FPN due to the variation. In order
to achieve both high SNR from the dark to moderate illumination and high DR in high
Figure 3.31 shows the linear-logarithmic pixel [3.20]. If the detection node voltage
(a) i ( )/
=i =i e =i e
( )/
=i =i e (3.46)
V +V
(b) V =
2
80
kT i
(c) V =V −2 ln
q i
response term in Eq. 3.46 (c) has doubled gain. When VPD is larger than VLOG-2VT, the
transistor M2 and M3 is turned off and the pixel operates in the linear region as
conventional 3-T pixel. In this structure, by adjusting VLOG, the trip point between the
VLOG
VRST
VX SH
R
VPD
S
The advantage of the logarithmic pixel is that it does not have much fill factor loss
and the operation simple, i.e., no additional timing control is required. However, the
logarithmic response from the sub VT region has higher FPN that is cannot be cancelled
from the CDS operation. The previous work shows 112 dB DR, 9.4 μm pixel pitch, 30 %
FF using 0.35 μm CMOS process. It shows 0.83 % FPN in the linear region and 1.37 %
81
3.3.2.2 Linear Scheme: Sensitivity Adjustment
The sensitivity is dependent on the conversion gain q/CFD. With large CFD, we can
transfer and store more charge in the FD node. However, the sensitivity is decreased due
to larger capacitance. As shown in figure 3.32, the sensitivity can be controlled using an
additional capacitor CS and a selection switch S2. In this scheme, the pixel keeps high
sensitivity in low illumination (S2 off). The pixel decreases the sensitivity in high
illumination in order to prevent the saturation from high illumination (S2 on). However,
the dynamic range extension is limited according to the capacitance of CS. Larger CS
CS
S2 R
TX
FD
S1
Another scheme using the sensitivity adjustment is shown in figure 3.33. When
illumination is high and the accumulation of photogenerated electrons exceeds the well
capacity, a photodiode is saturated and electrons are spilled over. If we collect these
spilled electrons in the additional capacitor with larger capacitance, we can read out
signals with reduced sensitivity. This lateral overflow scheme [3.21] is shown in figure
82
3.33. In figure 3.33 (a), the pixel circuit is shown. An additional overflow capacitor COV
and the floating diffusion (FD) is connected through the transistor S2.
Figure 3.33 Lateral overflow scheme (a) Pixel circuit (b) Operation procedure
The operation procedure is described in figure 3.33 (b) with electron potential
diagrams. The operation procedure is as follows: (1) Reset FD and COV (S: on) (2)
Integration, if charges are spilled over, they are collected both in the FD and COV. (3)
Reset FD: S is turned off, and FD is reset. (4) 1st signal readout: transfer charges into FD
83
and then readout signal (5) 2nd signal readout: S is turned on again and charge stored in
COV is read out together with charges in the FD. Two signals from (4) and (5) are
The advantage of this scheme is that it has linear response. Linear response is
preferable in a post image processing. It also has better SNR than nonlinear methods.
However, due to the additional capacitor and transistor, fill factor will be decreased. In
this work, the poly capacitor is implemented for COV and it is placed below the metal line.
Therefore, fill factor loss is decreased. The previous work achieved 93 dB DR, 5.6 μm
pixel pitch, 45 % fill factor using 0.18 μm CIS process. The limitation is that shared pixel
structure cannot be applicable with this scheme. Therefore, it is difficult to scale down
scheme. Image signals with multiple integration time are synthesized. Figure 3.34 shows
the architecture using triple exposure [3.22]. In the case of triple exposure, three image
signals with three different integration times from short to long have to be synthesized. In
order to synthesize three image signals, two image signals should be temporarily stored in
line memories before final image signals are read out and are synthesized together.
TROW: one row access time, T1 = L·TROW, T2 = M·TROW, T3=N·TROW (L > M > N)
The image signal from T1 is temporarily stored during (M+N)·TROW. After M·TROW,
the image signal from T2 is temporarily stored during N·TROW. After N·TROW, the final
84
image signal from T3 is read out and three image signals are synthesized into one image
signal with wide dynamic range. Therefore, total M+2N rows of line memories are
required. This takes significant portion of total die area. In the previous work, memory
Moreover, this scheme requires multiple readouts, which require additional power
consumption.
mentioned in chapter 1, the image sensor has to generate a high dynamic range image
(around 100 dB) at low-power consumption. Non-linear scheme using the logarithmic
pixel has no fill-factor loss and simple operations. However, it has low SNR and
temperature variation, which does not satisfy the requirement of reliability. Therefore, the
linear scheme is better choice in applications for distributes sensors. The sensitivity
adjustment scheme is one good solution. However, it is not applicable to the shared pixel
85
architecture, which is critical to achieve small form factor and low cost The multiple-
exposure scheme also has larger form factor due to the additional line memory. Moreover,
In the proposed adaptive image sensor, we proposed the low-power dual exposure
scheme that does not require additional circuits or memories. It will be discussed in
chapter 4.
3.4 CMOS Image Sensors with Integrated Image Signal Processing for Bandwidth
Reduction
If an image sensor generates M × N resolution 8-bit images with 30 frames per
second (fps), total bandwidth is 240·M·N, which requires huge bandwidth and power
consumption for an image signal transmission. However, the imaging with constant
frame rate and the maximum resolution has redundancy in terms of both the spatial and
temporal domain.
An image compression is one way to reduce the bandwidth [3.23, 3.24]. Many image
86
format for still images or MPEG for videos are applied to the commercial digital imaging
systems.
Even though the image compression offers the bandwidth savings, it still requires
large bandwidth if imaging systems operate with fixed resolution and fixed frame rate.
Therefore, it is not adequate for distributed imaging systems especially for applications of
surveillance systems that continuously monitor specific target objects or specific events.
Instead of generating images with maximum frame rate and resolution, several
CMOS imagers generate only low-bandwidth features from motion events [3.25-3.29] or
low resolution images that contains only target objects in order to save bandwidth [3.31,
3.32]. Another scheme generates variable resolution images for the bandwidth reduction
[3.33-3.35]. In the next three subsections, previous works with integrated image
bandwidth features from specific events. Instead of transmitting full resolution images,
the sensor delivers features only when specific events have been monitored.
One of simple feature is the temporal change from the motion sensing. The basic
operation of the motion sensing in CMOS image sensors is frame difference that has
simple operation. The difference between current frame signals and stored previous frame
signals are subtracted. For the motion detection, storage elements are required to store the
signals of previous frame. These storage elements can be placed in pixels as depicted in
figure 3.35 [3.29]. There are two sample and hold circuits and two signal readout paths
87
for direct frame difference output. However, fill factor is decreased and pixel size
becomes large due to in-pixel capacitors and additional readout paths. Pixel size is
ICOMP
R F
S
VO VC
VCOMP
ISF
88
Another previous work for the motion event generation contains two capacitors and a
comparator in pixel as shown in figure 3.36 [3.25]. It generates normal image through the
source follower as conventional 3T-pixel sensors. In addition, the comparator for the
motion detection is integrated and generates motion images (frame difference). The
analog output of frame difference is quantized into 2-bit digital signal in order to encode
three events: no-motion, darker and brighter changes. This pixel contains 6 transistors
and two capacitors. Fill factor is 17% and pixel size is 25.2 × 25.2 μm2 with 0.5 μm
technology.
(AER). Instead of frame-based imaging, the sensor generates the output asynchronously
based on the events [3.30]. The pixel responds to relative changes of light intensity and
asynchronously generates 1-b signals. The sensor outputs the address of pixels that
generated 1-b signals, i.e., have enough contrast change. Since the pixel has to monitor
the temporal contrast change, it has additional capacitors and circuits, which increase the
pixel size (8.1% fill factor, 40×40 μm2 pixel size using 0.35μm technology).
By transmitting the output only when the temporal change is detected, data
redundancy can be suppressed. However, the data rate of AER increases according to the
objects.
object out of entire image. The ROI readout scheme also suppresses the redundancy
89
because the background image is redundant. We don’t need to generate entire image
required bandwidth and power consumption are saved. Accordingly, fast and simple post
The ROI can be simply determined by external logics [3.32]. A pixel array is divided
with fixed-sized block, and several blocks for the ROI are selected by external logics.
Basic block size is 32 × 32 pixels. The ROI size can be variable by the external control.
However, this work does not include integrated ROI decision that detects target objects.
This work just selects a group of blocks and outputs ROI images.
SIG
More advanced ROI readout capability has been published [3.31]. In this work, the
sensor generates the ROI image containing brighter objects compared with the
background image. This approach is targeted to the application for the car plate (that is
bright from the headlights) monitoring at night. The pixel circuit is shown in figure 3.37.
90
In the pixel, two additional photodiodes are integrated. Each additional photodiode
detects brightness using column / row parallel charge amplifiers. Each output of charge
amplifiers generates profiles in X-Y directions. Using these profiles, ROI is decided by
external circuit. The size of additional photodiode is small and fill factor is not severely
reduced. However, this scheme is focused on the limited application and cannot detect
dark object.
variable-resolutions from the maximum to the reduced resolutions. In order to reduce the
subsampling induces the aliasing effect. Therefore, a low resolution image is generally
capacitive networks, averaged value can be calculated [3.33, 3.34]. However, additional
CCD image sensors, and is applied to CMOS image sensors by using the shared pixel
architecture [3.36, 3.37]. In this scheme, charges integrated in several photodiodes are
transferred into a floating diffusion with small capacitance, and sensitivity is enhanced.
The image signal can also be merged in digital domain [3.35]. After the analog-to-
digital conversion, signals are averaged using digital circuits. However, it has worse SNR
compared with charge domain merging because the readout noise and quantization noise
are added..
91
3 Motion
3.5 n-adaptive m
multi-resollution imagge sensor
Constannt frame-ratee images with
w maximuum spatial rresolution eentail high bbandwidth
a also large power coonsumptionn [3.38, 3.399]. This mayy limit its appplication foor portable
and
d
devices or wireless
w sennsor networrks in whichh both banddwidth and power conssumptions
a limited. Reducing the spatial resolution iis one way to reduce tthe amountt of image
are
d
data. Howevver, details in the imagee may be lost from the reduced ressolution.
r
resolutions in one focaal-plane: i.ee., to providde a high frrame rate oonly in the region-of-
i
interest (RO
OI) where thhe moving objects are located andd allow a loow frame raate for the
r
rest of stattionery bacckgrounds. The imagees of the sstationary oobjects are typically
r
redundant aat a high frrame rate annd it only ddegrades thhe SNR whhile wasting the large
b
bandwidth.
a
amount of ddata. Combiining the muultiple tempporal and sppatial resoluutions, we caan acquire
t image w
the with detailss for stationnery objectts and withh negligible motion-bluur for the
m
moving objeects at low-ppower consumption and low-bandw
width.
992
In orderr to suppresss the redunddancy of banndwidth, a m
motion-adapptive multi--resolution
C
CMOS imagger that is aable to proviide an adapttive spatial-ttemporal m
multi-resolutiion for the
s
specific reggion of interrest which is autonomoously determ
mined from motion deteection has
b
been propossed [3.40]. F
Figure 3.388 shows outpput images from the m
multi-resoluttion image
s
sensor.
f
from the tw
wo channels: one at low frame rate with the maaximum spaatial resolutiion for the
s
stationery bbackgroundss delivering the details of scenery; and the othher at high frame
f rate
993
w the redduced spatiial resolutioon for the m
with moving objeects in the ROI suppreessing the
m
motion blurr. Since the high framee readout is performed only in the ROI, banddwidth and
p
power conssumption caan be signifficantly redduced. In thhe ROI, thee spatial ressolution is
r
reduced by four times. The chargges from foour shared ppixels are ssummed in a floating
d
diffusion noode in orderr to enhance the SNR duuring the shhort integratiion time.
The RO
OI is determiined in real--time by thee integrated on-chip mootion detection circuit.
T
This circuitt has been implementted in a sm
mall area by
b sequentiial inter-pixxel switch
994
operations and does not require any additional in-pixel memory. Figure 3.39 shows the
pixel and column architecture of the multi-resolution sensor. The pixel has 4-shared pixel
architecture. The floating diffusions (FD) of two 4-shared groups are connected with
node in the pixel has been used as an analog memory. The signal from the lower FD is
transferred to upper FD through IS and the half of previous frame signal is stored in
upper FD (because the capacitance of two FDs are same). After the half of integration
time, the half of current frame signal is generated in the lower FD. The frame difference
Figure 3.40 shows the block diagram of multi-resolution sensor. In the multi-
resolution readout, the row decoder and the column selector provide independent pixel
controls for the ROI as well as the background region. The dual channel CDS/ADC
blocks will generate the normal and ROI image signals respectively. The motion
detection is performed only within the ROI. From continuous motion detection, the ROI
will be redefined and updated from the FPGA. The ROI signals are accessed at a high
frame rate during the blanking time in the one row time that is assigned for normal
readout.
3.6 Summary
In this chapter, we estimated and analyzed power consumption in CMOS image
sensor. From this estimation, we can draw power reduction strategies: voltage scaling
95
This chapter introduced three categories of previous works about low-power CMOS
image sensors, integrated dynamic range enhancement / high-sensitivity readout, and the
integrated image processing for the bandwidth savings. As mentioned in chapter 1, for
spatial and temporal bandwidth) should be satisfied together. However, these four
requirements are generally in trade-off and this trade-off is clearly shown from the
previous works.
For the low-power consumption, a voltage scaling is limited due to the SNR
consumption. However, it also has SNR degradation and the limitation of voltage scaling.
Moreover, the input voltage of in-pixel comparator is corrupted from the illumination
increase the dynamic range. Two methods for wide dynamic range are integrated
dynamic range enhancement scheme and high-sensitivity readout. Three dynamic range
enhancement schemes are introduced in this chapter. Non-linear scheme using the
logarithmic pixel offers no fill-factor loss and simple operation. However, it has low SNR
and temperature variations, which do not satisfy the requirement of reliability. The
sensitivity adjustment scheme is not applicable to the shared pixel architecture and
increases the form factor. The multiple-exposure scheme also makes large form factor
due to the additional line memory. Moreover, it requires higher power consumption than
96
For the spatial and temporal bandwidth savings, integrated lightweight image
processing is essential. The motion sensing and the ROI image readout suppress both the
spatial and temporal redundancy because they generate low bandwidth output only when
events or interested objects are detected. The variable-resolution readout also provides a
spatial bandwidth saving by merging neighboring pixels. However, these schemes have
limitations for applications of wireless imaging systems due to large pixel size and high
implement an image sensor adaptable, i.e., the image sensor changes its operation
implement with small pixel size and small form factor in order to reduce both power
consumption and cost. In this chapter, a motion-adaptive imager that is able to provide an
autonomously determined from the motion detection has been introduced. The next two
chapters will explain about the three adaptive imaging methods: energy, illumination, and
97
Chapter 3 References
98
[3.14] S. Lee, K. Lee, J. Park, H. Han, Y. Park, T. Jung, Y. Jang, “A 1/2.33-inch 14.6M
1.4μm-pixel backside-illuminated CMOS image sensor with floating diffusion
boosting,” ISSCC Dig Tech. Papers, pp. 416-418, Feb. 2011.
[3.15] K. Kagawa, S. Shishido, M. Nunoshita, and J. Ohta, “A 3.6pW/frame·pixel
1.35V PWM CMOS Imager with Dynamic Pixel Readout and no Static Bias
Current,” in 2008 IEEE International Solid-State Circuits Conference - Digest of
Technical Papers, 2008, pp. 54-595.
[3.16] S. Hanson, Z. Foo, D. Blaauw, and D. Sylvester, “A 0.5 V Sub-Microwatt CMOS
Image Sensor With Pulse-Width Modulation Read-Out,” IEEE Journal of Solid-
State Circuits, vol. 45, no. 4, pp. 759-767, Apr. 2010.
[3.17] M. –T. Chung, C. –C. Hsieh, “A 0.5V 4.95μW 11.8fps PWM CMOS imager with
82dB dynamic range and 0.055% fixed-pattern noise,” ISSCC Dig Tech. Papers,
pp. 114-116, Feb. 2012.
[3.18] J. Yang, K. G. Fife, L. Brooks, C. G. Sodini, A. Betts, and P. Mudunuru, “A
3MPixel Low-Noise Flexible Architecture CMOS Image Sensor,” in 2006 IEEE
International Solid State Circuits Conference - Digest of Technical Papers, 2006,
pp. 2004-2013.
[3.19] C. Lotto, P. Seitz, and T. Baechler, “A sub-electron readout noise CMOS image
sensor with pixel-level open-loop voltage amplification,” in 2011 IEEE
International Solid-State Circuits Conference, 2011, pp. 402-404.
[3.20] M. Vatteroni, P. Valdastri, A. Sartori, A. Menciassi, and P. Dario, “Linear–
Logarithmic CMOS Pixel With Tunable Dynamic Range,” IEEE Transactions on
Electron Devices, vol. 58, no. 4, pp. 1108-1115, Apr. 2011.
[3.21] N. Akahane, S. Sugawa, S. Adachi, K. Mori, T. Ishiuchi, and K. Mizobuchi, “A
Sensitivity and Linearity Improvement of a 100-dB Dynamic Range CMOS
Image Sensor Using a Lateral Overflow Integration Capacitor,” IEEE Journal of
Solid-State Circuits, vol. 41, no. 4, pp. 851-858, Apr. 2006.
[3.22] J. Solhusvik et al., “A 1280×960 3.75 μm pixel CMOS imager with Triple
Exposure HDR,” in International Image Sensor Workshop, 2009.
[3.23] S. Kawahito et al., “A CMOS image sensor with analog two-dimensional DCT-
based compression circuits for one-chip cameras,” IEEE Journal of Solid-State
99
Circuits, vol. 32, no. 12, pp. 2030-2041, 1997.
[3.24] W. D. León-Salas, S. Balkir, K. Sayood, N. Schemm, and M. Hoffman, “A
CMOS imager with focal plane compression using predictive coding,” IEEE
Journal of Solid-State Circuits, vol. 42, no. 11, pp. 2555–2572, 2007.\
[3.25] U. Mallik, M. Clapp, G. Cauwenberghs, and R. Etienne-Cummings, “Temporal
change threshold detection imager,” in ISSCC. 2005 IEEE International Digest of
Technical Papers. Solid-State Circuits Conference, 2005., 2005, pp. 362-364.
[3.26] V. Gruev and R. Etienne-Cummings, “A Pipelined Temporal Difference Imager,”
IEEE Journal of Solid-State Circuits, vol. 39, no. 3, pp. 538-543, Mar. 2004.
[3.27] S. Mizuno, K. Fujita, H. Yamamoto, N. Mukozaka, and H. Toyoda, “A 256 × 256
compact CMOS image sensor with on-chip motion detection function,” IEEE
Journal of Solid-State Circuits, vol. 38, no. 6, pp. 1072-1075, Jun. 2003.
[3.28] Y. Muramatsu, S. Kurosawa, M. Furumiya, H. Ohkubo, and Y. Nakashiba, “A
signal-processing CMOS image sensor using a simple analog operation,” IEEE
Journal of Solid-State Circuits, vol. 38, no. 1, pp. 101-106, Jan. 2003.
[3.29] L.-gee Chen, “A single-chip CMOS APS camera with direct frame difference
output,” IEEE Journal of Solid-State Circuits, vol. 34, no. 10, pp. 1415-1418,
1999.
[3.30] P. Lichtsteiner, C. Posch, T. Delbruck, "A 128× 128 120 dB 15 μs Latency
Asynchronous Temporal Contrast Vision Sensor," Solid-State Circuits, IEEE
Journal of , vol.43, no.2, pp.566-576, Feb. 2008
[3.31] Y. Sugiyama et al., “A high-speed CMOS image sensor with profile data
acquiring function,” IEEE Journal of Solid-State Circuits, vol. 40, no. 12, pp.
2816-2823, Dec. 2005.
[3.32] O. Schrey, J. Huppertz, G. Filimonovic, A. Bubmann, W. Brockherde, and B. J.
Hosticka, “A 1K×1K High Dynamic Range CMOS Image Sensor With On-Chip
Programmable Region-of-Interest Readout,” IEEE Journal of Solid-State Circuits,
vol. 37 no. 7 pp. 911-915, Jul. 2002
[3.33] Z. Zhou, B. Pain and E. Fossum, “A CMOS Imager with On-Chip Variable
Resolution for Light-Adaptive Imaging,” ISSCC Dig Tech. Papers, pp. 174-175,
Feb., 1998.
100
[3.34] S. E. Kemeny, R. Panicacci, B. Pain, L. Matthies, and E. R. Fossum,
“Multiresolution Image Sensor,” IEEE Transactions on Circuits and Systems for
Video Technology, vol. 7, no. 4, pp. 575-583, Aug. 1997
[3.35] E. Artyomov et al., “Morton (Z) Scan Based Real-Time Variable Resolution
CMOS Image Sensor,” IEEE T. CAS Video Tech., vol. 15, pp. 947-952, Jul.,
2005.
[3.36] K. Young Chan et al., “1/2-inch 7.2MPixel CMOS Image Sensor with 2.25/spl
mu/m Pixels Using 4-Shared Pixel Structure for Pixel-Level Summation,” in
2006 IEEE International Solid State Circuits Conference - Digest of Technical
Papers, 2006, pp. 1994-2003.
[3.37] S. Yoshihara et al., “A 1/1.8-inch 6.4 MPixel 60 frames/s CMOS Image Sensor
With Seamless Mode Change,” IEEE Journal of Solid-State Circuits, vol. 41, no.
12, pp. 2998-3006, Dec. 2006.
[3.38] S. Kleinfelder and A. El Gamal, “A 10000 frames/s CMOS digital pixel sensor,”
IEEE Journal of Solid-State Circuits, vol. 36, no. 12, pp. 2049-2059, 2001
[3.39] A I. Krymski, N. E. Bock, D. Van Blerkom, and E. R. Fossum, “A high-speed,
240-frames/s, 4.1-mpixel CMOS sensor,” IEEE Transactions on Electron Devices, vol.
50, no. 1, pp. 130-135, Jan. 2003.
[3.40] J. Choi, S. –W. Han, S. –J. Kim, S. –I. Chang, E. Yoon; , "A Spatial-Temporal
Multiresolution CMOS Image Sensor With Adaptive Frame Rates for Tracking the
Moving Objects in Region-of-Interest and Suppressing Motion Blur," Solid-State
Circuits, IEEE Journal of , vol.42, no.12, pp.2978-2989, Dec. 2007
101
CHAPTER 4 ENERGY/ILLUMINATION-ADAPTIVE IMAGE SENSOR WITH
RECONFIGURABLE MODES OF OPERATION
are important to deliver reliable information over widely changing illumination. However,
constant monitoring with maximum awareness requires large power consumption and is
scavenging wireless sensor nodes. One of ways to reduce the power is voltage scaling
[4.1, 4.2]. However, it significantly reduces the SNR and results in poor image quality
[4.2]. The signal can be easily corrupted from the noise in dark conditions or be saturated
in bright conditions. Dual power supply has been applied but still the conventional source
follower readout scheme limited the signal range [4.3]. Most imagers with high-
sensitivity and wide dynamic range reported [4.4, 4.5] consume large power > 50 mW,
low power, turns into the high sensitivity or wide dynamic range imaging when
illumination varies, or when detailed image transmission is requested from the network.
The sensor changes its operation back to monitoring mode when the energy harvesting is
poor or it is needed to save more charges in the battery. In this chapter, we report an
adaptive CMOS image sensor which employs four different imaging modes: monitoring,
102
e
enables thee reliable monitoring
m w
while signiificantly ennhancing thee battery liifetime of
w
wireless imaage sensor nnodes.
b
block diagraam of the pproposed sennsor chip. IIn the monittoring modee, the sensoor operates
e
enable pow
wer saving. For an in-ppixel ADC,, a SAR ADC is chossen becausee its short
c
conversion time (8 cyccles) saves power.
p Otheer three moodes operatee at 1.8V suupply with
s
signal from fiers, which suppresses the input
m in-pixel diifferential ccommon-souurce amplifi
r
referred noiise in dark cconditions. In normal mode,
m the siignal is readd out thoughh a source
f
follower as in conventional sensors.
1103
Figuree 4.2 Operaation and block diagraam of the ad
daptive imaage sensor with
reconfiigurable modes
In WDR
R mode, duual exposuree prevents thhe signal frrom saturatiion by provviding two
i
integration times, shorrt and long, in the sam
me frame. T
The sensor has colum
mn parallel
A
ADCs and 110-b latchess. The ADC
C is designedd to be recoonfigured too operate as either 8-b
S
SAR or 10--b SS ADCs accordingg to the moddes. The coolumn scannner accessess the latch
a image signals
and s are rread out throough sense aamplifiers.
e
each mode. Two pixelss are verticaally shared aand form onne group. Tw
wo groups can
c form a
d
differential pair for in-pixel differential aamplifiers. This differrential connfiguration
e
eliminates a gain loss ffrom the boody effect inn the sourcee follower. IIn order to rreduce the
1104
VGS drop aand increasee the signall swing, low
w-VTH transsistors are uused for ‘T
TAMP.’ The
c
control signnals for transsmission gaates, ‘TX0’ and
a ‘TX1,’ are boostedd by bootstraap circuits
o
outside the pixel. For differential readout, eaach group uuses separatte signal linnes ‘SIG0’
a ‘SIG1.’’ ‘COM’ linne, which is common to two grouups, is connnected to eiither pixel
and
p
power suppply (for souurce followeer operationn) or currennt bias (for differential amplifier
o
operation). ‘RSTC’ is connected to the capaacitive DAC
C for in-pixxel SAR AD
DC in the
m
monitoring mode or is connected tto pixel pow
wer supply ffor the reset operation. Figure
F 4.5
s
shows the tiiming diagraam of the piixel control in each mode.
1105
COM
In monitoring mode, the pixel circuit operates as a preamplifier of SAR ADC at 0.8V
as shown in figure 4.6. The signal from the capacitive DAC, which is located in the
column circuits, is provided through ‘RSTC’ and the in-pixel preamplifier generates the
amplified comparison signal in each conversion cycle. The pixel power supply (0.8 V) is
106
SAR ADC Photodiode reset
RST0
RST1
TX0
TX1
Monitoring mode
RST0 RST0
RST1 RST1
TX0 TX0
TX1 TX1
Figure 4.7 shows the capacitive DAC for the SAR ADC. Since most power
consumption comes from the driving of capacitances, we used a modified DAC with
reduced capacitance [4.6]. Conventional DAC has total 28C capacitances for 8-bit
column pitch is also a problem. However, the modified DAC has only 47C total
capacitances.
107
In order to adjust the signal swing, we used parasitic capacitance of ‘RSTC’ instead
RSTC
108
C
V = V (4.1)
C +C
,where CDAC is DAC capacitance, CPAR is the parasitic capacitance of RSTC line. In order
to suppress the FPN from the parasitic capacitance variation, we added calibration
capacitors CCAL which are configured by 4-b digital signals. However, we did not have to
use the calibration capacitors in the testing because we did not see any noticeable FPN
from signal swing variation. The measured signal swing is 0.41 V out of 0.8 V power
supply. Since the pixel operates at 0.8V, it is difficult to achieve a complete charge
transfer from pinned photodiodes due to low reset level in the floating diffusion.
Therefore, we chose to apply charge sharing. After integration, the floating diffusion
node is reset and then the integrated charge is shared with the floating diffusion by
opening the transfer gate. This operation disables true CDS like a 3-T pixel; however, we
109
In normal mode, the pixel circuit is configured as a source follower that is exactly
differential common-source amplifier and achieves a higher gain (> 6). The differential
common source amplifier generates the output signal from the two floating diffusion
nodes: one gives a signal voltage and the other gives the reset voltage. In WDR mode, the
same circuit is used as in the normal mode. In order to enhance the dynamic range, pixel
merging with dual exposure is applied to the pixel. This will be explained in the next
subsection.
Figure 4.9 Timing diagram of column-parallel ADC (a) Single ended input (normal,
WDR mode), (b) Differential input (high-sensitivity mode)
Figure 4.8 shows the column parallel ADC and figure 4.9 shows its operation. The
at 0.8V, and switches for mode configuration. In monitoring mode, only 0.8 V supply is
110
used and column-level preamplifiers are turned off for power saving. The amplified
signal from the pixel (‘SIG0’, ‘SIG1’) is directly delivered to the dynamic comparator
through ‘SM’ for SAR ADC. In the other three modes, the column parallel ADC operates
as a single-slope ADC. The coupling capacitor ‘CC’ stores the offset of the preamplifier
and suppresses the column FPN. In the normal and WDR modes, reset voltage is sampled
in CDAC, signal voltage is sampled in CS, and then VRAMP is swept. In case of high-
sensitivity mode, the operation is fully differential, i.e., the differential signals from the
111
Comp
CN
NT
CNT C
CNT
QCO
ONV
QDIFF
D
Figure 44.10 shows the hybrid latch circuiit. The latchh operates ffor both SA
AR and SS
A
ADC by m
multiplexing. Differentiial latching reduces sw
witching poower becausse it only
s
switches onnce during thhe whole SS ADC cyccles, while thhe SRAM oor transfer-ggate based
f
flip-flop alw
ways changges its state whenever the counterr signal chaanges. The switching
c
characteristi
ics of differrential and cconventionaal latches aree shown in ffigure 4.11.
4 In-equaality readou
4.3 ut
Figure 44.12 shows conventiona
c al digital siggnal readoutt scheme.
Figuree 4.12 Convventional diigital imagee signal reaadout circuiit and operation
1112
The preecharged seense line (S
SL, sense aamplifier innput) is disscharged thrrough the
r
readout celll according to
t the storedd value in thhe latch.
Figgure 4.13 Im
mage localitty
When thhe digital siggnals from the column latch are acccessed, thee prechargedd line may
W
When the sstored signaal is ‘1’, thee prechargedd line is discharged annd the line should be
c
charged agaain for the next colum
mn readout. Precharginng the sensse line that has large
c
capacitance entails hugge dynamic power
p consuumption.
Fortunattely, the im
mage signal has
h strong llocality as shown
s in figgure 4.13. IIt has high
p
probability that neighbboring pixels have sim
milar valuess. In order to reduce the
t power
c
consumptio n in the precharge
p mplement a scheme too read an inequality
node, we im
b
between neeighboring ccolumns sinnce most im
mage signalls typically have locallly similar
v
values (low spatial bandwidth).
Figure 4.14
4 shows tthe digital ssignal readoout circuit. W
We put XOR
R gates in thhe readout
c for upper 4 MSBs. One of the two XOR ggate inputs ccomes from
cell m the left collumn. The
1113
l
line keeps tthe pre-charrged voltagee if the two columns haave the sam
me value. In this case,
t
there is neggligible dynaamic currennt. We testedd with 100 images from
m the imagee database
[
[4.7]. The probability of equalityy in each bbit positionn is shown in Figure 4.15. The
p
probability of equality is defined aas Eq. 4.2. T
This plot is from the avveraged ressult of 100
i
images. Figgure 4.15 ((a) shows thhe probabillity that 1-bb digital siignals in sppecific bit
p
position aree same in thhe neighboriing columnss. Figure 4.115 (b) show
ws the probaability that
1-b digital ssignal in speecific bit poosition is ‘1’. The probaability of ‘1’’ is defined as Eq. 4.3.
The proobability of equality meeans the proobability off negligible dynamic cuurrent. As
s
shown in thhe figure, thhe probabiliity of equality is 0.73 (@ 4th bit)) and 0.94 ((@ MSB),
w
whereas thee probabilityy of ‘1’ is aaround 0.5. The power consumptioon overheadd from the
t
transmission
n gate basedd XOR is negligible.
n A
According too this analysis, we can achieve a
p
power reducction of 68%
% (only from
m the pre-ccharge) and 5% from thhe total chipp power in
1114
t monitorring mode. The
the T area ovverhead from
m XOR gattes is only 00.36 % from
m the core
s
size.
Fiigure 4.15 A
Averaged p
probability of equalityy
t sensitivvity adjustm
the ment and thee multiple eexposure scchemes provvide better SNR than
n
nonlinear scchemes. In the linear scheme,
s muultiple expossure is betteer choice beecause the
s
sensitivity a
adjustment rrequires in-ppixel capaciitors and thiis scheme iss not applicaable to the
1115
shared pixel architecture. However, multiple exposure scheme requires additional line
memories. It also entails large power consumption due to the multiple readouts.
In this work, we implemented a dual exposure scheme with pixel merging. The
operation principle is shown in figure 4.16. The signal from short integration time (TINT2)
and the signal from long integration time (TINT) are synthesized. As shown in the pixel
output response, we can get the signal without saturation from the short integration time
(TINT2) in high illumination. In conventional dual exposure scheme, line memories are
required for the synthesis as shown in figure 4.16. Moreover, two times readout doubles
proposed a dual exposure scheme based on the pixel merging. As shown in figure 4.17,
the even group has a regular integration time (TINT) and the odd group has a short
integration time (TINT2). In the readout, one row from the even group and one row from
the odd group are merged, and generate a signal with enhanced dynamic range. This
116
s
scheme sacrifices the sspatial resoolution by hhalf; howeveer, it is adeequate for low-power
a
applications
s because no
n additionaal in-pixel eelement or iimage reconnstruction is required
c
consumptio n is requiredd as the connventional m
multiple expposure schem
mes.
4 Implem
4.5 mentation an
nd experim
mental resullts
h been fabbricated usinng 0.18μm CIS processs. A chip m
A prototype chip has micrograph
i shown in figure 4.18. The perforrmance of thhe sensor is summarizeed in table 4.1.
is
Figure 44.19 shows the captureed images ffrom the faabricated deevice. Figure 4.19 (a)
s
shows the im
mage from tthe monitorring mode (@
@ 9 lx) andd figure 4.199 (b) shows the image
f
from the higgh-sensitivitty mode in a dark conddition (@ 0.44 lx).
1117
Tablee 4.1 Chip ccharacterisstics
Figu
ure 4.19 Saample imagges
1118
Figure 44.19 (c) shoows the image from thee normal moode when illlumination high. The
s
saturation frrom high illlumination iis eliminatedd in WDR m
mode as shoown in figurre 4.19 (d).
The pow
wer figure of
o merit (Poower-FOM) can be defi
fined as the power norm
malized to
f
frame rate aand the num
mber of pixeels, given bby power/fraame·pixel. T
This does not
n exactly
r
reflect a goood indicattion for thee performannce of imaage sensorss because tthe power
c
consumptio n of certainn circuit bloocks does not
n linearly increase ass the pixel array size
b
becomes laarge. More importantlly, signal-too-noise rattio (SNR) can be siggnificantly
e
extremely ppoor. Anothher figure oof merit, SN
NR-power FOM, can bee defined coonsidering
b
both SNR aand power coonsumptionn in its figurees:
SNR-Powerr FOM = ×1
10 [dB] ((4.4)
We com
mpared the SNR-Powerr FOM of oour work w
with previouus low-pow
wer CMOS
i
imagers in ttable 4.2. W FOMs of preevious workks, we used the whole
When we calculate the F
c
chip power consumptioon and the reeported SNR
R.
i
indicate a diagonal line that shoows an equuivalent SN
NR/Power-F
FOM line w
where the
1119
p
projection of achievabble SNR is estimatedd when additional poower consum
mption is
a
allowed.
In the m
monitoring mode, the ssensor shouuld be able to operate solely depeending on
h
harvested ennergy. Thenn, the powerr FOM shouuld be less tthan 100 [ppW/frame/piixel] to be
s
suitable for the operatioon of 1 μW @ 100×1000 pixels, 1 ffps. At the ssame time, iin order to
m
monitor thhe environm
ment withoout severe performannce degraddation, the effective
r
resolution oof images shhould be oveer 6 bits (SN
NR = 36.1 ddB). Furtherr image proocessing in
t sensor node or inn the base station cann enhance the SNR and providde reliable
the
m
monitoring. Our sensorr working inn the monitooring mode satisfies this condition,, while the
o
other previoous sensors cannot.
i
illumination
n conditionss (wide-dynaamic range)), the sensorr will adaptiively changee its mode
o operationn and the poower source can be swittched from eenergy harvvester to a baattery. For
of
1120
these accurate image capturing modes, at least 8-b resolution (SNR = 48.2 dB) is required.
As shown in figure 4.20, our sensor can provide adequate SNR in other modes of
operation at the minimum power consumption. The sensor, however, will be mostly
working in monitoring mode by default unless the special needs happen. During the
monitoring mode, the extra energy harvested will recharge the battery. Considering both
SNR and power FOM, this work shows the best performance as shown in the figure.
4.6 Summary
In this chapter, we explained the energy/illumination adaptive imaging. As a
characterized. It operates at 1.36 μW in the monitoring mode from harvested energy and
(100dB) at 867 μW in battery operation upon request. The chip achieved power FOM of
121
Chapter 4 References
[4.1] K. -B. Cho, A. I. Krymski and E. R. Fossum, “A 1.5-V 550-μW 176 × 144
autonomous CMOS active pixel image sensor,” IEEE Transactions on Electron
Devices, vol. 50, no. 1, pp. 96-105, Jan. 2003
[4.2] S. Hanson, Z. Foo, D. Blaauw, and D. Sylvester, “A 0.5 V Sub-Microwatt CMOS
Image Sensor With Pulse-Width Modulation Read-Out,” IEEE Journal of Solid-
State Circuits, vol. 45, no. 4, pp. 759-767, Apr. 2010
[4.3] Q. Gao, S. Member, and O. Yadid-pecht, “A Low Power Block Based CMOS
Image Sensor with Dual VDD,” IEEE Sensor Journal, pp. 1-9, 2011
[4.4] J. Yang, K. G. Fife, L. Brooks, C. G. Sodini, A. Betts, P. Mudunuru and H. –S.
Lee, “A 3MPixel Low-Noise Flexible Architecture CMOS,” ISSCC Dig Tech.
Papers, pp. 496-497, Feb. 2006
[4.5] S.-W. Han, S.-J. Kim, J. Choi, C.-K. Kim, and E. Yoon, “A High Dynamic Range
CMOS Image Sensor with In-Pixel Floating-Node Analog Memory for Pixel
Level Integration Time Control,” Dig. Symp. VLSI Circuits, pp. 31-32, June 2006
[4.6] U.S. Patent, #US6,686,865 B2, Feb. 2004
[4.7] Caltech Image Database, http://www.vision.caltech.edu/html-files/archive.html
[4.8] K. Kagawa, S. Shishido, M. Nunoshita, and J. Ohta, “A 3.6pW/frame·pixel
1.35V PWM CMOS Imager with Dynamic Pixel Readout and no Static Bias
Current,” ISSCC Dig Tech. Papers, pp. 54-55, 2008
122
CHAPTER 5 OBJECT-ADAPTIVE IMAGE SENSOR WITH EMBEDDED
FEATURE EXTRACTION FOR MOTION-TRIGGERED OBJECT-OF-
INTEREST IMAGING
5.1 Introduction
The image signal transmission requires large bandwidth to transmit the 8-b M × N
signals with 15 fps. The wireless transmission with high bandwidth requires high power
consumption > 10 mW which is not feasible in wireless sensor nodes. Using the low-
power transmission such as UWB still requires > 1 mW power consumption [5.1].
consumption of the image sensor. One way to reduce the bandwidth is to generate signals
only when event happens by monitoring temporal changes [5.2-5.4]. However, this event-
based imaging has extraneous redundancy because the sensor may also respond to
shows the operation procedure of the motion-triggered OOI imaging. In most of time the
sensor is in sleeping mode until it is triggered by motion. Then it wakes up, it generates
and transmits 8-b features for object detection. The signal processing unit which resides
either in the host or in the sensor node, performs detection of objects and feeds back 1-b
123
request signal to initiate further imaging operation if the target object-of-interest is
identified.
The feature extraction is the process that provides transformed representation sets
with less bandwidth. There is a variety of image processing for the feature extraction:
from simple low-level feature extraction that does not need shape information (for
example: edge detection, curvature detection, optical flow estimation) to the complex
high-level feature extraction that finds the shapes in the images (for example: Hough
The object detection is the task of finding presence and position of the object. The
object detection requires descriptors, which is a set of feature data describing the
properties of image. By comparing and matching the descriptors with the descriptors of
known objects, we can detect the object. The descriptor should be congruent, compact
and invariant (for example, invariant to the orientation of the object or invariant to the
illumination) [5.5]. A variety of algorithms for generating descriptor has been researched
such as Haar wavlet, sacale-invariant feature transform (SIFT), shape context and so on.
gradients (HOG) because it gives a high detection rate of objects with simple operation
[5.6]. The HOG feature is used in the human detection for applications of pedestrian
detectors. Humans have been proven to be difficult object to detect due to the variability
of appearance, cloth and articulation. While HOG feature provides powerful performance
(i.e., low miss rate of detection) compared with other features including wavelets [5.7,
5.8], it has simple operation to be implemented in wireless sensor nodes. In this work, we
implemented the HOG feature extraction algorithm using mixed-signal circuitry in order
124
t save bothh power andd area. The HOG featuure output rrequires onlyy 3.5% of bbandwidth
to
w
when compared with coonventionall 8-b image capturing.
For the object adapptive imaginng, we impplemented a prototype CMOS imaage sensor
s
size) where the numberr of blocks L × K is conntrollable. IIn the case oof 32 × 32 blocks,
b we
c achievee ×113 banddwidth savinng comparedd with geneerating imagge signals. IIn the next
can
s
section, thee operationn principles of the HO
OG featuree will be eexplained. T
Then, the
p
proposed CM
MOS imagee sensor witth OOI imagging will be described.
e
extraction cchain is as foollows:
( Spatial gradient ggeneration: apply two 1-D masks GX and GY for thee x and y
(1)
d
direction reespectively. The maskss are GX = [-1 0 1] aand GY = [-1 0 1]T. T
The result
g
generates tw
wo M × N gradient im
mages from
m the M × N image ffor x and y direction
r
respectively
y.
1125
(2) Magnitude and angle calculation: from two gradient images, calculate the magnitude
|H| = GX + GY (5.1)
GY
θ = tan (5.2)
GX
(3) Voting: each pixel is voted to 9 bin numbers (0 ~ 8), which decode the angle from 0°
to 180° as shown in figure 5.2. For example, if the angle θ is 50°, then the bin number is
2.
(4) Histogram generation: M × N image is divided by L × K blocks. Each block has M/L
× N/K pixels. Typically the block size M/L × N/K is over 8 × 8. Each block generates
one histogram. Each histogram has 9 bin numbers as an index. According to the voting,
the magnitude |H| is accumulated in each bin number. The accumulated magnitude in a
The histogram of oriented gradients shows the magnitude distributions in each angle.
For example, if there is a horizontal pillar in the scene, the histogram will have the
The block size (M/L × N/K) induces different detection result. Large cell size can
generate the information over wide area instead of losing the details. Small cell size can
generate detailed information but requires more processing. Some of object detectors
have the classification algorithms that use multiple cell sizes in one detection procedure
[5.9]. For example, the histogram from four different cell sizes 8 × 8, 16 × 16, 32 × 32
and 64 × 64 are concatenated as a feature set and the concatenated feature set is used for
the classification.
126
As an integrated feature exttractor in CMOS imaage sensorss, it is desirable to
i
implement t cell sizee programm
the mable in ordeer to be appplicable to thhe different situations
s
such as the object or thhe classificattion algorithhm. Moreovver, after thee sensor gennerates the
h
histogram w
with M/L × N/K cell size, histograams with muultiple cell ssize can be generated
b the post image proceessing, whicch can be appplicable to the algorithhm mentioneed above.
by
Figure 5.2 HO
OG featuree extraction
n proceduree
5
5.3.1 Objecct-of-interest imaging
Figure 55.3 shows thhe simplifieed block diaagram of thrree differennt modes off operation
i the propoosed sensor chip. In mootion sensinng mode, thee sensor gennerates 128 × 128 1-b
in
m
motion mapp for motionn triggeringg. The colum
mn-parallel ADCs opeerate as a 1-b motion
c
comparator.
. Once the motion
m is ddetected, thee sensor wakkes up and turns into the
t feature
e
extraction m
mode. In thiis mode, thee sensor gennerates and ttransmits 8--b feature siignals that
1127
are extracted from full 8-b 256 × 256 images. We have SRAM blocks, which temporarily
store intermediate signals during HOG feature calculation. One part of SRAM blocks are
allocated for line buffers to store 3 rows of images and the rest are allocated for feature
signal accumulation. In imaging & storing mode, the sensor generates and transmits 8-b
images upon the request from the host. The SRAM operates as a frame memory, which
single slope ADCs operate as a 1-b motion comparator in the motion sensing mode and
generates a 1-b motion map with frame-difference signals from the pixel. For HOG
feature extraction, we need to calculate the magnitude and angle of spatial gradients. The
pixel array is divided into m × m sub-blocks where the size of sub-blocks is fully
programmable. A set of histogram that maps the magnitude of spatial gradients in each
angle is generated for each block. The processing of feature signal extraction is as
follows: (1) Three rows of images are read out and are temporarily stored in SRAMs; (2)
Spatial gradients (GX and GY) are calculated with masking [-1 0 1] and [-1 0 1]T,
128
respectively; (3) Gradient-to-angle converter (GAC) generates 9 bin numbers, which map
9 different angles from 0° to 180°; and (4) Histogram generator accumulates the
calculated from GAC, and store them back to SRAMs. The accumulated magnitudes are
differential topology, which enables both source follower readout and differential
previous work [5.10]. Two vertical neighboring pixels share the COM line and connect
signals through SIG0 and SIG1. In this sensor, we used an additional in-pixel capacitor
CE/O as frame buffer memory for motion sensing as well as for low-power imaging. The
in-pixel capacitor is implemented using an MIM capacitor and it is placed on top of pixel
129
circuits. Therefore, the capacitor does not induce any fill-factor loss. Two vertically
neighboring pixels are connected with a merging switch (M) for pixel merging.
COM
Even
TXE
RE
FDE
SE
CSE CE
VCOMP(E)
M
TXO
RO
FDO
SO
CSO CO
Odd
VCOMP(O)
The equivalent circuit of the pixel is shown in figure 5.6. In the motion sensing mode,
operation sequence is as follows: (1) Two pixels are merged and one in-pixel capacitor
samples the previous frame signal (V1); (2) The other in-pixel capacitor samples the
current frame signal (V2); (3) A short pulse is applied to VCOMP(E) in the even pixel and
130
V1 is compared with V2; (4) Likewise, another pulse is applied to VCOMP(O) in the odd
pixel and two voltages are compared again. When the frame difference |V1 - V2| is higher
than the amplitude of VCOMP pulse swing, 1-b motion output is generated.
VDDP
COM
V1 V2
(3) (4)
VCOMP(E) VCOMP(O)
M
Motion sensing
VDDP
TXE TXO
COM
VSIG VRST
VCOMP(E) VCOMP(O)
Low-power imaging
Figure 5.6 Equivalent pixel circuit for motion sensing and low-power imaging
For low-power imaging in the feature extraction mode, the pixel circuit operates as a
pre-amplifier for an 8-b single slope ADC. The integrated charges in a photodiode are
transferred to an in-pixel capacitor in the even pixel whereas the odd pixel keeps the reset
voltage. A ramp signal is applied to VCOMP(E) and the potential is dropped at SIG0 node
131
due to charge transfer. The latching occurs around the reset voltage. In this scheme, VGS
drop in the amplification transistors does not induce the loss of signal swing as in the
conventional source follower readout; i.e., the power supply can be scaled down by more
than 0.5 V (nominal voltage drop in VGS). In this sensor, we used 1.3V for pixel power
The ADC consists of sample & hold circuit, preamplifier, multiplexer, comparator
and 8-b latch. Note that the sample & hold circuit and the preamplifier are used for the
readout from the in-pixel source follower only for the testing purpose. The preamplifier is
turned off in actual operation. In actual operation, the differential signal line SIG0,1 are
directly connected with comparator inputs through the multiplexer. The offset voltage of
the in-pixel differential comparator is stored in the capacitor CC in order to suppress the
FPN from the offset voltage. When F is enabled, the output from the reset voltage is
sampled on CC. After F is disabled, integrated charges are transferred and ramp signal is
132
s
swept for thhe AD convversion. Thee power suppply voltage of the collumn circuiit (VDD) is
0 V.
0.8
5
5.3.5 Gradiient-to-Anggle Convertter (GAC)
Figure 5.8 shows the votingg process. Since the angle θ ccan be exppressed as
ttan-1(GY/GX
X), we can ddecode the angle into 9 bin numbeers using thee following equation:
tanθ = ((5.3)
In orderr to decodee with 9 bin numbers using the above equaation, tanθ should be
c
compared w
with the resuult from a divider. Thiss operation rrequires a ddivider, also a look-up
t
table for thee trigonomeetric functionn. Instead of
o using the complex opperation, wee proposed
a simple m
mixed-signaal approachh in order to avoid digital im
mplementatioon, which
c
consumes hhuge area/poower. The gradient-to-a
g angle conveerter (GAC)) circuit bassed on the
m al approachh is shown iin figure 5.99. The timinng diagram of GAC is shown in
mixed-signa
f
figure m 0° to 180° in 4-b binn numbers.
5.10. The GAC ggenerates 9 angle inforrmation from
T 8-b graadient signaals (GX andd GY) are coonverted to analog signnals using tw
The wo binary
c
capacitive D
DACs. Wheen the switcch S1 is off,
f, the binaryy DAC outpput is divideed by two.
O
One capaciitor (C0,1) operates aas the part of DAC, which gennerates trigonometric
f
function, while the otther (C0s,1s) operatess as a holdd capacitor, which conntains the
w
g
gradient.
1133
R
S2 S3
C0 C20 C40
S1
(i)
C0S
|GX|
C1S
|GY|
S1 (ii)
C1 C60 C80
S2 S3
DAC
R
134
The graddient is mulltiplied by tanθ
t and 1/taanθ in each channel, reespectively, according
t capacitannce values of
to o C20, C40,, C60, and C80.
C
t3
t4
c
comparison
n is as follow
ws:
|GX| ∙ tanθ = |G
GY| (tanθ < 1) ((5.4)
|GY|
|G
GX| = (tanθ > 1)) ((5.5)
tanθ
1135
In the first comparison, |GY| is compared with |GX| ∙ tanθ. In the second comparison,
| |
|GX| is compared with . In each comparison, two-step comparisons are performed. In
the first comparison, θ is 40° in the first step and is 20° in the second step. This two-step
comparison divides the range of the angle as follows: θ > 40° or 20° < θ < 40° or θ < 20°.
Likewise, in the second comparison, θ is 60° in the first step and is 80° in the second step.
This two-step comparison divides the range of the angle as follows: θ < 60° or 60° < θ <
80° or θ > 80°. Each step generates 1-bit comparison result. Therefore, total 4-bit output
is generated from two two-step comparison. Since tanθ = - tan(180° − θ), we don’t need
to do the operation for the angle higher than 90°. In other words, the comparison only
cares about the magnitude of tanθ and detects the angle range from 0° to 90°. The sign bit
(which we already have generated from the gradient calculation) shows whether the
decoded angle is in 0° < θ < 90° or in 90° < θ < 180°. After two-step comparisons during
t3 and t4, the 4-bit output from the comparison is encoded according to the sign bit (S) in
order to generate a final bin number. Table 5.2 shows the encoding table. The sign bit
determines whether the encoded angle is in 0° < θ < 90° or in 90° < θ < 180°.
5.11 shows the chip microphotograph. The core size is 2.35 × 3.18 mm2. Figure 5.12
shows the captured images from the fabricated device including a 128 × 128 1-b motion
map and a 256×256 8-b image from the in-pixel single-slope ADC. A 256 × 256 angle
map shows the angle calculated from the GAC. A 8-b feature from 8 × 8 blocks are
136
s
shown withh the vector. In this figuure, only twoo angles thaat have the biggest
b maggnitude are
s
shown for simplicity.
The perrformance oof the sensorr is summarrized in tablle 5.3. We aachieved a normalized
n
p
power of 133.46 pW/fraame•pixel inn motion seensing and 51.94
5 pW/fframe•pixel in feature
e
extraction. IIn order to vverify the performance of the integgrated featuure extraction unit, we
t
tested the oobject detecttion from thhe extractedd features ussing 200 pedestrian im
mages from
D
DaimlerChr
rysler dataset [6.10]. F
For testing, we input ttest images serially intto the 8-b
l
latch in thee column paarallel ADC
C and generrated featuree. The test result show
ws 94.5 %
d
detection raate. The detaailed proceddure of the object
o detecction will bee described in section
5
5.4.1.
1137
Figu
ure 5.12 Saample imagges
1138
5
5.4.1 Testin
ng of the ob
bject detecttion
Figure 5.13 showss the objecct detection procedure. The objecct detectionn requires
c
classifier w
with embeddded classificcation algoriithm. The cclassifier ideentifies the object-of-
i
interest andd generates 1-b output of the detection resultt. In order tto classify tthe object,
p
pre-trained model hass to be loaaded on thee classifier.. The trainning requirees another
a
algorithm such as suppport vector machine (S
SVM), whicch is supervvised learniing model
For the training, usuually lots off images (trraining imagge set) moree than 1,0000 are used.
T
This traininng set incluudes both the
t positivee images (w
with the taarget objectt) and the
n
negative im
mages (withoout target obbject). For thhe object deetection testting, a differrent set of
i
images (tesst image sett) are used. The featurre is input to the trainned classifieer and the
d
detection raate can be m
measured. Inn this work, we used ssimple SVM
M for the traaining and
S
SVM classifier withouut any advvanced algoorithms in the
t MATLA
AB. We used 1,000
1139
training images (500 positive, 500 negative) for the training [6]. In order to measure the
detection rate, we used 200 test image sets (100 positive, 100 negative) [6.10]. The test
images are serially input to the latch in the single-slope ADC and generated HOG feature
from the fabricated chip are input to the classifier. According to the measurement, it
shows 95% detection rate for the positive image set and 94% detection rate for the
negative image set. The overall detection rate is 94.5 %. We expect improved detection
5.4.2 Power figure of merit (FOM) comparison with previous low-power imagers
The proposed CMOS imager with object-of-interest imaging consumes low power
pitch is an important factor because the parasitic capacitance of metal line (Cm) is linearly
C =N ∙W ∙c (5.6)
In Eq.5.6, NPIXEL is the number of pixels and cm is the parasitic capacitance per μm.
should have small pixel pitch. However, small pixel pitch decreases the sensitivity of the
pixel due to reduced fill factor. Therefore, it is important to keep high fill factor while
reducing the pixel pitch. Many low-power sensors use additional in-pixel circuits [6.1-6.3]
or pMOS transistors [6.4] for the in-pixel event generation or in-pixel ADC.
In this work, we use the differential pixel topology by grouping two vertically
neighboring pixels in order to achieve in-pixel ADC without additional transistors. The
additional in-pixel memory is implemented with MIM capacitor and it is located on the
top of pixel circuits in order to keep small pixel pitch. In this way, we achieved 5.9 μm
140
p
pixel pitch with 30% fill
f factor. F
For the com
mparison off power connsumption, tthe power
f
figure of m
merit (FOM
M) is typicaally used. T
The power FOM is ddefined as tthe power
n
normalized frame·pixel. Table 5.4
to frame ratte and the nnumber of piixels, given by power/fr
s
shows the ppower FOM
M comparisoon. The propposed sensoor has 13.466 pW poweer FOM in
t motion sensing, 51.94 pW in tthe feature eextraction inncluding thee imaging power,
the p and
5 Summaary
5.5
w embeddded feature extraction
In this chapter, thee object-adaaptive imagge sensor with
t
triggered by motion ssensing andd extracts features
f froom the capptured imagge for the
d
detection off object-of-iinterest. Full image cappturing operaation is perfformed onlyy when the
i
interested obbjects are foound, whichh significanttly reduces ppower consuumption at the sensor
n
node. The cchip operatees at 0.22
W/frame inn motion seensing mode and operaates at 3.4
W/frame f feature eextraction, reespectively..
for
1141
Chapter 5 References
[5.1] S. Chen; W. Tang; X. Zhang & E. Culurciello, "A 64×64 Pixels UWB Wireless
Temporal-Difference Digital Image Sensor," IEEE Transactions on Very Large
Scale Integration (VLSI) Systems, vol.20, no.12, pp.2232-2240, Dec. 2012.
[5.2] L.-gee Chen, “A single-chip CMOS APS camera with direct frame difference
output,” IEEE Journal of Solid-State Circuits, vol. 34, no. 10, pp. 1415-1418,
1999.
[5.3] U. Mallik, M. Clapp, G. Cauwenberghs, & R. Etienne-Cummings, “Temporal
change threshold detection imager,” ISSCC Dig Tech. Papers, pp. 362-364, Feb.
2005.
[5.4] N. Massari, M. Gottardi, S. Jawed, "A 100μW 64×128-Pixel Contrast-Based
Asynchronous Binary Vision Sensor for Wireless Sensor Networks," ISSCC Dig
Tech. Papers, pp. 588-638, Feb. 2008.
[5.5] Mark S. Nixon and Albert S. Aguado, “Feature Extraction and Image Processing,”
Book Chapter 7, Academic Press, 2008
[5.6] N. Dalal, B. Triggs,”Histogram of oriented gradients for human detection,”
CVPR Dig Tech. Papers, pp. 886-893, Jun. 2005.
[5.7] A. Mohan, C. Papageorgiou, and T. Poggio. Example-based object detection in
images by components. PAMI, 23(4):349-361, April 2001
[5.8] P. Viola, M. J. Jones, and D. Snow. Detecting pedestrians using patterns of
motion and appearance. The 9th ICCV, Nice, France, volume 1, pages 734–741,
2003
[5.9] S. Maji, A. C. Berg, and J. Malik, “Classification using intersection kernel
support vector machines is efficient,” in 2008 IEEE Conference on Computer
Vision and Pattern Recognition, 2008, pp. 1-8.
[5.10] J. Choi, S. Park, J. Cho & E. Yoon, “A 1.36μW adaptive CMOS image sensor
with reconfigurable modes of operation from available energy/illumination for
distributed wireless sensor network,” ISSCC Dig Tech. Papers, pp. 112-114, Feb.
2012.
[5.11] M. T. Chung & C.-C. Hsieh, “A 0.5V 4.95μW 11.8fps PWM CMOS imager with
82dB dynamic range and 0.055% fixed-pattern noise, “ISSCC Dig Tech. Papers,
142
pp. 114-116, Feb. 2012
[5.12] Daimler Pedestrian Benchmark Datasets
(http://www.gavrila.net/Research/Pedestrian_Detection/)
143
CHAPTER 6 CONCLUSION AND FUTURE WORK
efficient operation. This chapter summarizes the main contributions and future directions
of adaptive sensors.
(1) Implementation of adaptive imaging at low-power consumption (< 1.36 μW) with
high dynamic range (99 dB) and low data rate (< 3.5 %);
(2) Low power imaging at the best SNR-power FOM of 25.22 dB among the work
(3) Low power imaging with feature extraction at the best power FOM (51.94 pW)
image sensor networks and wireless biomedical imaging systems. The adaptive imaging
includes: (1) the energy-adaptive imaging for the low-power consumption, (2) the
illumination-adaptive imaging for high dynamic range, and (3) the object-adaptive
imaging for spatial-temporal bandwidth savings. In order to prove the feasibility of the
proposed adaptive imaging schemes, we designed two prototype chips: one for the
energy/illumination adaptation, and the other for the object adaptation. Both prototype
144
For the energy adaptation, we implemented the low-power monitoring operation in
the adaptive image sensor. The sensor keeps monitoring at an extremely low power <
1.36 μW/frame, and only turns into high sensitivity or wide dynamic range imaging
operation when the power delivery is enough from energy harvesting units. In-pixel SAR
ADCs enable the voltage scaling down to < 0.8 V without any significant loss of signal
swing and non-linearity. The differential latches in the single slope ADC further reduce
the switching power. The in-equality readout scheme is also introduced to reduce the
The illumination adaptation extends the dynamic range for reliable imaging in wide
high-gain amplifier, and provides high sensitivity of 23.9 V/lx·s. At high illumination, in-
pixel dynamic range extension is performed using pixel merging and dual exposure
readout. This scheme significantly saves the chip size compared with conventional
schemes, and consumes less power compared with the multiple exposure schemes. The
(OOI) from the features extracted from image signals. The sensor wakes up triggered by
motion sensing and extracts features from the captured image for detection of the object-
of-interest. Full image capturing operation is performed only when interesting objects are
found, which significantly reduces bandwidth and power consumption of the sensor node.
The fabricated chip operates at 0.22 W/frame in the motion sensing and operates at 3.4
145
In summary, the adaptability enables an energy-efficient image sensing in
and a variety of objects captured in the focal plane. By accommodating the three
adaptation schemes, the fabricated image sensor could provide high dynamic range
images at low power consumption, while maintaining low bandwidth in image signal
transmission.
processing units including power management units and control blocks for self-decision
making, and also has to be mounted on a platform suitable for the deployment in a
distributed image sensor networks and for portable biomedical imaging systems.
capability have been implemented separately in the two prototype chips. In the future
sensor platform, all adaptive imaging capabilities should be integrated in a single chip in
order to achieve low-cost and small form-factor as well as low power consumption. In the
two prototype chips that we designed, some parts of control signals has been generated
from the off-chip FPGA that is mounted on the testing board for flexibility in
characterization. In the future design, all control blocks as well as other essential circuit
blocks have to be integrated on a chip, especially power management units (PMU) and
wireless transmitters. The power management unit (PMU) selects the energy source
either from energy harvesting units or from batteries, and it generates multiple power
146
The controller is essential in order to manage the adaptive operation. The controller
frame signals, motion triggering from the motion sensor output, and object detection
algorithm. The overall architecture of the adaptive imaging SoC is illustrated in figure 6.1.
would be the integration of the sensor platform including lens, solar cells and batteries.
sensor platforms. The cubic sensor includes four adaptive imaging SoCs, two solar cell
panels, and batteries in each plane. The lens is mounted on top of each imager. The
foldable architecture using a flexible PCB enables 360° field of view. Therefore, we can
147
a
achieve higgh sensing visibility in multiple perspectiives for efffective surrveillance.
M
Moreover, tthe sensor pplatform cann be implem mall size less than 5 × 5 × 5 mm3
mented in sm
g
given area. From the iimaging cubbe systems,, we are able to collecct more accurate data
b
based on gross
g statisstics for ennvironmental monitoring, surveiillance, andd security
a
applications
s.
1148
APPENDIX
149
Appendix A
150
In this section, we describe detailed estimation of dynamic power consumption in
consumption only from the capacitive driving (CLVDD2f) for simplicity. Moreover, the
leakage power consumption will be ignored. This estimation cause error compared with
actual dynamic power consumption. However, basic purpose of power estimation in this
consumption according to important parameters such as the number of pixel array, power
The dynamic comparator performs the level decision from the pre-amplified signals.
It operates with two phases: the precharge and the comparison. Figure A.1 shows the
comparator circuit.
151
The first stage which consists of latch, switch (pMOS for precharge) and inverter
induces the power consumption in each clock cycle. The frequency is 2N/TROW. The 2nd
stage which consists of RS latch and buffer has only one time switching during the ADC.
The power consumption per one comparator (per column) can be expressed as:
P = (P +P +P ) + (P +P )
== {2c V + 2c V + 2c V }+
1 (A.1)
{2c V + 10c V }
T
2 1
= (6c )V + (12c )V
T T
,where PSW is the power consumption from the precharge switches, PLATCH is from the
latch, PINV is from the inverter, PSR is from the RS latch, and PBUFFER is from the buffer..
B. Latches
The latch for N-bit single-slope ADC consists of N D flip-flops. The circuit schematic
of D flip-flop is shown in figure A.2. The output Q is connected with read buffer (one
nMOS transistor, not shown in the figure) for the digital signal readout.
The power consumption of one latch in the single-slope ADC can be represented as
follows:
(A.2)
P _ = 2c V f + 3c V f : from 2 transmission gates & 2 inverters, master
+2c V f + 4c V f : from 2 transmission gates & 2 inverters, slave
+7c V f +c V f : from clock buffer and output buffers
152
= 13c +c V f + 5c V f
1 f ( )
= 13c +c V + 5c V
T T
fD = 1/TROW, input (D) frequency, the input signal is the counter output.
fck = fCNT(n)/TROW, clock (clk) frequency, the clock signal is the comparator output.
fCNT(n): counter output frequency of n-th bit position
The total number of switching of N-b Gray code is (1+1+2+…2N-2). Therefore, the
As shown in the equations, the power consumption of two inverters in the master
stage is dependent on the input frequency. In the single-slope ADC, when clock signal is
low, the master stage continuously changes its state from the counter signals. After clock
goes high and the counter signal is latched to the slave stage, no more switching occurs.
Therefore, the power consumption is largely dependent on the signal level (i.e., light
intensity) of single slope ADC. In this estimation, we define the illumination level
parameter ‘I’ which varies 1/2N (dark) to 1 (bright). For the power consumption of N-bit
latch in the single slope ADC which uses Gray counter can be expressed as follows:
153
1
P =N 13c +c V
T
(A.3)
1
+5c V I(1 + 1 + 2 + ⋯ + 2 )
T
Note that the 2nd term includes the illumination level parameter ‘I’. If the signal level
is high from high illumination, the power consumption increases because the latch has
C. Counter
Figure A.3 shows the Gray-code counter circuit. The 1st stage is the ripple counter
which generates the binary code. The Gray code from the XOR gates are buffered in the
D Q D Q Q<N-1>
Q Q
D Q D Q Q<1>
Q Q
D Q D Q Q<0>
CLK Q CLK Q
Figure A.3 Gray-code counter
The power consumption of the N-bit ripple counter can be expressed as:
(A.4)
154
P = 2c V F + 3c V F : from 2 transmission gates & 2 inverters, master
+2c V F + 4c V F : from transmission gates & 2 inverters, slave
+7c V F + 4c V F : from clock buffer and output buffers
= 22c V F
1 + 2 + ⋯+ 2
F=
T
The exclusive-or (XOR) gate for the Gray code conversion consists of 3 gates. The
power consumption of XOR gate is:
P = 2c V F+c V F (A.5)
The 2nd stage flip-flops for the glitch suppression has the power consumption as
follows:
(A.6)
P = 2c V F + 3c V F : from 2 transmission gates & 2 inverters, master
+2c V F + 4c V F : from transmission gates & 2 inverters, slave
+7c V F + 4c V F : from clock buffer and output buffers
= 15c V F + 7c V F
1 + 1 + ⋯+ 2 F
F = ≈
T 2
The inverters drive the metal lines with high capacitance. The power consumption
from the driver is as follows:
P =# W c V F (A.7)
The overall power consumption from the counter can be expressed as:
P =P +P +P +P
(A.8)
1 + 2 + ⋯+ 2
= (39c + 0.5# W c )×V
T
Periphery circuits
155
A. Column scanner
In this estimation, a simple shift register instead of decoder will be used. In each
stage of shift register, the output buffer drives N nMOS transistor (for N-b digital signal
readout, the column selection) in one column. The clock frequency fck is #COL/TROW, and
the input frequency fD is half of the clock frequency. The power consumption of one flip
flop is shown as:
(A.9)
P _ = 2c V f + 3c V f : from 2 transmission gates & 2 inverters, master
+2c V f + 4c V f : from 2 transmission gates & 2 inverters, slave
+7c V f + Nc V f : from clock buffer and output buffers
= 7c V f + 11c + Nc V f
f
= 7c V f + 11c + Nc V
2
#
= 12.5c + 0.5Nc V
T
#
P = 12.5c + 0.5Nc V (A.10)
T
B. Row scanner
The row scanner consists of a shift register and drivers as shown in figure A.4. In
each row, the row scanner consists of one flip flop, three NAND gates, three level-up
converters and three inverters. In the shift register, the input frequency fD is 1/TROW and
the clock frequency fck is #ROW/TROW. The power consumption of one D flip-flop in the
(A.11)
P _ = 2c V f + 3c V f : from 2 transmission gates & 2 inverters, master
+2c V f + 4c V f : from 2 transmission gates & 2 inverters, slave
156
+7c V f + 4c V f : from clock buffer and output buffers
#
= 7c V + 15c V
q(n-1) D Q TX
TXG
Q Q
CLK
R
RG
thin-ox
S
SG
# 1
P = 7c V + 15c V (A.12)
T T
The NAND gate enables the pixel control signals (TX, R, S) according to the external
control signals (TXG, RG, SG). The power consumption of three NAND gates in each
row is as follows:
1
P = 3c V (A.13)
T
Since the pixel power supply is higher than digital power supply (VDD), level-up
(A.14)
P = 2c V (1st stage)
P = 2c V (2nd stage)
157
P = 4c V +# (W c + c )V (driver)
P =P +P +P
The power consumption from level-up converters from all rows can be expressed:
1
P = 3P = 2c V
T
(A.15)
1
+ 6c +# W c +c V
T
P =P +P +P
1
= {7c # + 20c }V
T (A.16)
1
+ 6c +# W c +c V
T
158