SOLID-STATE IMAGING DEVICE AND ELECTRONIC DEVICE
20250081643 ยท 2025-03-06
Inventors
Cpc classification
H10D84/0126
ELECTRICITY
H04N25/778
ELECTRICITY
H10D84/00
ELECTRICITY
H10D84/0165
ELECTRICITY
International classification
Abstract
Solid-state imaging devices and electronic devices with reduced noise are disclosed. In one example, a solid-state imaging device includes a first substrate including a photodiode and a transfer transistor, and a second substrate including an active load circuit and a differential pair circuit for a comparator, in which the active load circuit includes a first transistor, the differential pair circuit includes a second transistor, and transconductance of the first transistor is smaller than transconductance of the second transistor.
Claims
1. A solid-state imaging device comprising: a first substrate including a photodiode and a transfer transistor; and a second substrate including an active load circuit and a differential pair circuit for a comparator, wherein the active load circuit includes a first transistor, the differential pair circuit includes a second transistor, and transconductance of the first transistor is smaller than transconductance of the second transistor.
2. The solid-state imaging device according to claim 1, wherein an area of an LDD region on a source region side and a drain region side of the first transistor is larger than an area of an LDD region on a source region side and a drain region side of the second transistor in plan view, respectively.
3. The solid-state imaging device according to claim 2, wherein an area of a source region and a drain region of the first transistor is narrower than an area of a source region and a drain region of the second transistor in plan view, respectively.
4. The solid-state imaging device according to claim 2, wherein a film thickness of a sidewall insulating film of the first transistor is thicker than a film thickness of a sidewall insulating film of the second transistor.
5. The solid-state imaging device according to claim 2, wherein a width of an LDD region on a source region side and a drain region side of the first transistor is equal to a film thickness of a sidewall insulating film of the first transistor, and a width of an LDD region on a source region side and a drain region side of the second transistor is equal to a film thickness of a sidewall insulating film of the second transistor.
6. The solid-state imaging device according to claim 2, wherein a film thickness of a sidewall insulating film of the first transistor is equal to a film thickness of a sidewall insulating film of the second transistor.
7. The solid-state imaging device according to claim 2, wherein a width of an LDD region on a source region side and a drain region side of the first transistor is different from a film thickness of a sidewall insulating film of the first transistor, and a width of an LDD region on a source region side and a drain region side of the second transistor is different from a film thickness of a sidewall insulating film of the second transistor.
8. The solid-state imaging device according to claim 1, wherein a total contact area between a source region and a drain region of the first transistor and one or more contact plugs of the first transistor is smaller than a total contact area between a source region and a drain region of the second transistor and one or more contact plugs of the second transistor, respectively.
9. The solid-state imaging device according to claim 8, wherein a part of the contact plugs of the first transistor is disposed on an insulating film.
10. The solid-state imaging device according to claim 9, wherein the insulating film is a sidewall insulating film of the first transistor.
11. The solid-state imaging device according to claim 9, wherein the insulating film is an element isolation insulating film surrounding a source region and a drain region of the first transistor in plan view.
12. The solid-state imaging device according to claim 8, wherein an area of each contact plug of the first transistor is smaller than an area of each contact plug of the second transistor in plan view.
13. The solid-state imaging device according to claim 12, wherein a shape of each contact plug of the first transistor is similar to a shape of each contact plug of the second transistor in plan view.
14. The solid-state imaging device according to claim 12, wherein a shape of each contact plug of the first transistor is not similar to a shape of each contact plug of the second transistor in plan view.
15. The solid-state imaging device according to claim 8, wherein a number of the contact plugs of the first transistor is smaller than a number of the contact plugs of the second transistor.
16. The solid-state imaging device according to claim 15, wherein an area of each contact plug of the first transistor is equal to an area of each contact plug of the second transistor in plan view.
17. The solid-state imaging device according to claim 1, wherein the first transistor is a transistor of a first conductivity type, and the second transistor is a transistor of a second conductivity type different from the first conductivity type.
18. The solid-state imaging device according to claim 1, wherein the first transistor and the second transistor are provided on a same semiconductor substrate.
19. The solid-state imaging device according to claim 1, wherein the first substrate is disposed on the second substrate, and the second substrate is disposed on a third substrate.
20. An electronic device comprising an imaging device, the imaging device including: a first substrate including a photodiode and a transfer transistor; and a second substrate including an active load circuit and a differential pair circuit for a comparator, wherein the active load circuit includes a first transistor, the differential pair circuit includes a second transistor, and transconductance of the first transistor is smaller than transconductance of the second transistor.
Description
BRIEF DESCRIPTION OF DRAWINGS
[0025]
[0026]
[0027]
[0028]
[0029]
[0030]
[0031]
[0032]
[0033]
[0034]
[0035]
[0036]
[0037]
[0038]
MODE FOR CARRYING OUT THE INVENTION
[0039] Hereinafter, embodiments of the present disclosure will be described with reference to the drawings.
First Embodiment
[0040]
[0041] As illustrated in
[0042] The pixel array unit 1 includes a plurality of pixels 11 arranged in a two-dimensional array. The pixel array unit 1 of the present embodiment includes a plurality of pixel sharing units 12, and each of the pixel sharing units 12 includes two or more pixels 11 (here, 2*2 pixels 11a to 11d).
[0043] The input unit 2 receives as input a reference clock signal, a timing control signal, characteristic data, and the like from the outside to the inside of the solid-state imaging device. The reference clock signal is a signal that serves as a reference of operation of the solid-state imaging device. The timing control signal includes a horizontal synchronization signal and a vertical synchronization signal. The characteristic data is, for example, data for storing an image signal in the solid-state imaging device. These signals are taken in by the input circuit 2b from the input terminal 2a, the amplitudes of these signals are changed to amplitudes that are easy to use by the amplitude change unit 2c, and the arrangement of data strings included in these signals is changed by the data conversion unit 2d. For example, the data conversion unit 2d converts these signals from serial signals to parallel signals.
[0044] The timing control unit 3 supplies a signal for controlling timing to the row drive unit 4 and the column signal processing unit 5 on the basis of a reference clock signal and a timing control signal.
[0045] The row drive unit 4 includes, for example, a row decoder and a row drive circuit. The row decoder determines a row to be subjected to pixel driving. The row drive circuit generates a signal for driving the pixels 11 and supplies the signal to the row (row drive signal line 13) determined by the row decoder.
[0046] The column signal processing unit 5 includes, for example, a load circuit, an amplifier circuit, a noise processing unit, an AD converter, a scanning circuit, and the like. The load circuit is electrically connected to the column read line 14, and forms a source follower circuit with the pixel sharing unit 12. The amplifier circuit amplifies a pixel signal read from the pixel supply unit 12 via the column read line 14. The noise processing unit removes noise from a pixel signal. The AD converter converts a pixel signal from an analog signal to a digital signal. The AD converter includes, for example, a comparator that compares a pixel signal with a reference signal, and a counter that counts a time until a comparison result of the comparator is inverted. The scanning circuit scans a read column (column read line 14) so as to output a pixel signal.
[0047] The image signal processing unit 6 performs various types of signal processing on data obtained by photoelectric conversion performed by the pixel array unit 1, such as an image signal output from the pixel array unit 1. The image signal processing unit 6 includes, for example, a data processing unit and a data holding unit. The data processing unit performs, for example, tone curve correction for increasing or decreasing the number of gradations of the image. The data holding unit stores, for example, data used for tone curve correction in advance.
[0048] The output unit 7 outputs image data obtained by processing an image signal output from the pixel array unit 1 by the column signal processing unit 5 or the image signal processing unit 6 from the inside to the outside of the solid-state imaging device. The image data is converted from a parallel signal to a serial signal by the data conversion unit 7a, the amplitude of this signal is changed to an amplitude that is easy to use by the amplitude change unit 7b, and this signal is output from the output terminal 7d by the output circuit 7c.
[0049]
[0050]
[0051] Furthermore, a +Z direction corresponds to an upward direction, and a Z direction corresponds to a downward direction. Note that the Z direction may strictly match the gravity direction, or does not necessarily strictly match the gravity direction.
[0052] As illustrated in
[0053] The first substrate 21 includes a semiconductor substrate 31, an element isolation insulating film 32, a gate insulating film 33 and a gate electrode 34 of each transistor Tr1, an electrode portion 35, an interlayer insulating film 36, and a photodiode PD of each of the pixels 11. The semiconductor substrate 31 includes an n-type region 31a, a p-type region 31b, and a floating diffusion portion 31c for each of the pixels 11.
[0054] The second substrate 22 includes a semiconductor substrate 41, a gate insulating film 42 and a gate electrode 43 of each transistor Tr2, an interlayer insulating film 44, an interlayer insulating film 45, a plurality of plugs 46a to 46d, a plurality of wiring layers 47a to 47c, and a plurality of pads 48. The semiconductor substrate 41 includes a plurality of diffusion regions 41a.
[0055] The third substrate 23 includes a semiconductor substrate 51, a gate insulating film 52 and a gate electrode 53 of each transistor Tr3, an interlayer insulating film 54, an interlayer insulating film 55, a plurality of plugs 56a to 56c, a plurality of wiring layers 57a and 57b, and a plurality of pads 58. The semiconductor substrate 51 includes a plurality of diffusion regions 51a.
[0056] The semiconductor substrate 31 is, for example, a silicon (Si) substrate. In
[0057] The semiconductor substrate 31 includes a photodiode PD for each of the pixels 11. The photodiode PD of each of the pixels 11 is mainly formed by pn junction between the n-type region 31a and the p-type region 31b, and functions as a photoelectric conversion unit. The photodiode PD of each of the pixels 11 receives light from the back surface side of the semiconductor substrate 31, generates a signal charge corresponding to the light amount of received light, and accumulates the generated signal charge in the floating diffusion portion 31c.
[0058] The element isolation insulating film 32 is included in the semiconductor substrate 31, and penetrates the semiconductor substrate 31 between the front surface and the back surface of the semiconductor substrate 31. The element isolation insulating film 32 is, for example, a silicon oxide film (SiO.sub.2 film). The solid-state imaging device of the present embodiment may further include a light shielding layer (for example, a tungsten (W) layer) embedded in the element isolation insulating film 32. The element isolation insulating film 32 has a mesh shape surrounding the plurality of pixels 11 described above for each of the pixels 11 in plan view.
[0059] The first substrate 21 includes a plurality of transistors Tr1. These transistors Tr1 include, for example, pixel transistors such as transfer transistors. The gate insulating film 33 and the gate electrode 34 of each of the transistors Tr1 are sequentially formed on the front surface of the semiconductor substrate 31. The gate insulating film 33 is, for example, a SiO.sub.2 film. The gate electrode 34 is, for example, a poly-Si layer.
[0060] The electrode portion 35 is formed on the front surface of the semiconductor substrate 31 and is in contact with the floating diffusion portion 31c. The electrode portion 35 is, for example, a poly-Si layer. The gate electrode 34 and the electrode portion 35 of the present embodiment are formed by the same material being processed.
[0061] The interlayer insulating film 36 is formed on the front surface of the semiconductor substrate 31 and covers the gate electrode 34 and the electrode portion 35. The interlayer insulating film 36 is, for example, a SiO.sub.2 film.
[0062] The semiconductor substrate 41 is, for example, a Si substrate. The semiconductor substrate 41 is disposed on the lower surface of the interlayer insulating film 36. In
[0063] The second substrate 22 includes a plurality of transistors Tr2. These transistors Tr2 include, for example, pixel transistors such as reset transistors, amplification transistors, and selection transistors. The gate insulating film 42 and the gate electrode 43 of each of the transistors Tr2 are sequentially formed on the front surface of the semiconductor substrate 41. As illustrated in
[0064] The interlayer insulating film 44 is formed on the front surface of the semiconductor substrate 41 and covers the gate electrode 42. The interlayer insulating film 45 is formed on the lower surface of the interlayer insulating film 44. These interlayer insulating films 44 and 45 are, for example, SiO.sub.2 films.
[0065] The plugs 46a to 46d, the wiring layers 47a to 47c, and the pads 48 are formed in the interlayer insulating films 44 and 45. Specifically, the wiring layers 47a to 47c are sequentially formed below the semiconductor substrate 41. The pads 48 are formed below the wiring layers 47a to 47c, and are located on the lower surface of the second substrate 22. Each plug 46a is a contact plug that electrically connects the diffusion region 41a or the gate electrode 42 and the wiring layer 47a. Each plug 46b is a via plug that electrically connects the wiring layer 47a and the wiring layer 47b. Each plug 46c is a via plug that electrically connects the wiring layer 47b and the wiring layer 47c. Each plug 46d is a via plug that electrically connects the wiring layer 47c and any of the pads 48.
[0066] The semiconductor substrate 51 is, for example, a Si substrate. The semiconductor substrate 51 is disposed below the interlayer insulating films 44 and 45 with the interlayer insulating films 54 and 55 interposed therebetween. In
[0067] The third substrate 23 includes a plurality of transistors Tr3. These transistors Tr3 form, for example, a logic circuit. The gate insulating film 52 and the gate electrode 53 of each of the transistors Tr3 are sequentially formed on the front surface of the semiconductor substrate 51. The gate insulating film 52 is, for example, a SiO.sub.2 film. The gate electrode 53 is, for example, a poly-Si layer. Each of the diffusion regions 51a in the semiconductor substrate 51 functions as, for example, a source region or a drain region of any of the transistors Tr3.
[0068] The interlayer insulating film 54 is formed on the front surface of the semiconductor substrate 51 and covers the gate electrode 52. The interlayer insulating film 55 is formed on the upper surface of the interlayer insulating film 54. These interlayer insulating films 54 and 55 are, for example, SiO.sub.2 films. As illustrated in
[0069] The plugs 56a to 56c, the wiring layers 57a and 57b, and the pads 58 are formed in the interlayer insulating films 54 and 55. Specifically, the wiring layers 57a and 57b are sequentially formed above the semiconductor substrate 51. The pads 58 are formed above the wiring layers 57a and 57b, and are located on the upper surface of the third substrate 23. Each plug 56a is a contact plug that electrically connects the diffusion region 51a or the gate electrode 52 and the wiring layer 57a. Each plug 56b is a via plug that electrically connects the wiring layer 57a and the wiring layer 57b. Each plug 56c is a via plug that electrically connects the wiring layer 57b and any of the pads 58. As illustrated in
[0070] The solid-state imaging device of the present embodiment has a three-layer structure including the first, second, and third substrates 21 to 23. The solid-state imaging device of the present embodiment further includes the filter layer 24 and the on-chip lens layer on the first substrate 21, and includes the through plug 26 in the first and second substrates 21 and 22.
[0071] The filter layer 24 includes a plurality of filters including an effect of transmitting light having a predetermined wavelength. For example, the filters for red (R), green (G), and blue (B) are arranged above the photodiodes PD of the red, green, and blue pixels 11, respectively. Moreover, a filter for infrared light may be arranged above the photodiode PD of the pixel 11 for infrared light.
[0072] The on-chip lens layer 25 includes a plurality of on-chip lenses including an effect of condensing incident light. In the present embodiment, light incident on each of the on-chip lens is collected by the each of the on-chip lens, transmitted through the corresponding filter, and incident on the corresponding photodiode PD. The photodiode PD converts the light into a charge by photoelectric conversion to generate a signal charge. The generated signal charge is accumulated in the floating diffusion portion 31c.
[0073] The through plug 26 is formed in the interlayer insulating film 36, the substrate 41, and the interlayer insulating film 44. The through plug 26 is a contact plug that electrically connects the electrode portion 35 and the wiring layer 47a. The first substrate 21 and the second substrate 22 of the present embodiment are electrically connected via the through plug 26. On the other hand, the second substrate 22 and the third substrate 23 of the present embodiment are electrically connected via the pads 48 and 58.
[0074]
[0075]
[0076] As illustrated in
[0077] The second substrate 22 includes a comparator 61. The comparator 61 is included in the AD converter of the above-described column signal processing unit 5 (
[0078] The transistors Tp1 and Tp2 form an active load circuit 62. The gate of the transistor Tp1 is electrically connected to a gate of the transistor Tp2. The sources of the transistors Tp1 and Tp2 are electrically connected to a power supply wiring (VDD). The drain of the transistor Tp1 is electrically connected to the drain of the transistor Tn1 and gates of the transistors Tp1 and Tp2. The drain of the transistor Tp2 is electrically connected to the drains of the transistors Tn2 and Tn4 and the pad 48. The active load circuit 62 is a current mirror circuit that causes current corresponding to a mirror ratio to flow through the transistors Tp1 and Tp2. The transistors Tp1 and Tp2 are examples of a first transistor of the present disclosure.
[0079] The transistors Tn1 and Tn2 form a differential pair circuit 63. The gate of the transistor (reference transistor) Tn1 is electrically connected to a wiring for a reference signal. The gate of the transistor (input transistor) Tn2 is electrically connected to a wiring for an image signal (through plug 26) and is electrically connected to the source of the transistor Tn4. The sources of the transistors Tn1 and Tn2 are electrically connected to the drain of the transistor Tn3. The differential pair circuit 63 outputs a comparison result (voltage difference) between an image signal and a reference signal to a node between the transistor Tp2 and the transistor Tn2, and outputs the comparison result from the node to the pad 48. The transistors Tn1 and Tn2 are examples of a second transistor of the present disclosure.
[0080] The transistor Tn3 functions as a current source. The gate of the transistor Tn3 is electrically connected to a wiring for applying predetermined voltage. The source of the transistor Tn3 is electrically connected to a ground wiring (GND). This current source maintains the entire current flowing through the transistors Tn1 and Tn2 at a predetermined value.
[0081] The transistor Tn4 is disposed between the through plug 26 and the node described above, and functions as an AZ transistor. The gate of the transistor Tn4 is electrically connected to a wiring for a reset signal. The source of the transistor Tn4 is electrically connected to the through plug 26. The drain of the transistor Tn4 is electrically connected to the node described above. The AZ transistor causes the through plug 26 (floating diffusion portion 31c) to conduct electricity to the node described above before detecting an output signal, and performs an auto-zero operation.
[0082] In the following description, each of the transistors Tp1 and Tp2 in the active load circuit 62 is also referred to as transistor Tp, and each of the transistors Tn1 and Tn2 in the differential pair circuit 63 is also referred to as transistor Tn. In the present embodiment, the transconductance of the transistor Tp is set to be smaller than the transconductance of the transistor Tn.
[0083]
[0084] A of
[0085] B of
[0086] In the present embodiment, the sizes of the transistors Tp1, Tp2, Tn1, and Tn2 are set to the same size. Specifically, the gate lengths of the transistors Tp1, Tp2, Tn1, and Tn2 are set to the same value, and the gate widths of the transistors Tp1, Tp2, Tn1, and Tn2 are set to the same value. On the other hand, the transconductance gm.sub.p of the transistor Tp is set to be smaller than the transconductance gm.sub.n of the transistor Tn (gm.sub.p<gm.sub.n). As a result, the random noise RN illustrated in A of
[0087] According to the present embodiment, it is possible to reduce noise of the solid-state imaging device without making sizes of the transistors Tp1, Tp2, Tn1, and Tn2 different from each other. Therefore, according to the present embodiment, it is possible to achieve both miniaturization of the transistors such as the transistors Tp1, Tp2, Tn1, and Tn2 and reduction of noise of the solid-state imaging device. In other words, according to the present embodiment, it is possible to miniaturize the transistors while reducing noise of the solid-state imaging device.
[0088] In the present embodiment, the transconductance of the transistor Tp1 is set to be the same as the transconductance of the transistor Tp2, the transconductance of the transistor Tn1 is set to be the same as the transconductance of the transistor Tn2, and the transconductance of the transistors Tp1 and Tp2 is set to be smaller than the transconductance of the transistors Tn1 and Tn2. However, the transconductance of the transistor Tp1 may be set to a value different from the transconductance of the transistor Tp2. Furthermore, the transconductance of the transistor Tn1 may be set to a value different from the transconductance of the transistor Tn2.
[0089]
[0090] A of
[0091] In A of
[0092] A of
[0093] A of
[0094] Next, the transistor In illustrated in A of
[0095] A of
[0096] This similarly applies to the transistor Tp illustrated in B of
[0097] In the present embodiment, the film thickness of each of the sidewall insulating films 71 of the transistor Tp is thicker than the film thickness of each of the sidewall insulating films 71 of the transistor Tn, and the width Wp of each of the diffusion regions 41b of the transistor Tp is wider than the width Wn of each of the diffusion regions 41b of the transistor Tn. As a result, the area of each of the diffusion regions 41b of the transistor Tp is larger than the area of each of the diffusion regions 41b of the transistor In in plan view. More specifically, the areas of the diffusion regions 41b on the source region side and the drain region side of the transistor Tp are larger than the areas of the diffusion regions 41b on the source region side and the drain region side of the transistor In, respectively.
[0098] A and B of
[0099] As described above, in the present embodiment, the area of each of the diffusion regions 41b of the transistor Tp is set to be larger than the area of each of the diffusion regions 41b of the transistor Tn. Therefore, according to the present embodiment, it is possible to reduce noise of the solid-state imaging device in a suitable manner, such as making it easier to achieve both miniaturization of the transistors and reduction of noise of the solid-state imaging device.
Second Embodiment
[0100]
[0101] A and B of
[0102] The structures of the transistors Tn and Tp of the present embodiment are substantially similar to the structures of the transistors Tn and Tp of the first embodiment, respectively. However, the width Wn of each of diffusion regions 41b of the transistor Tn of the present embodiment is different from the film thickness of each of sidewall insulating films 71 of the transistor Tn, and is, for example, wider than the film thickness of each of the sidewall insulating films 71 of the transistor Tn. The reason is that the diffusion regions 41b are formed by ion implantation using a gate electrode 43 as a mask, and diffusion regions 41a are then formed in the diffusion regions 41b by ion implantation using films other than the sidewall insulating films 71 as masks. Next, mask films are removed, and then the sidewall insulating films 71 are formed. Similarly, the width Wp of each of the diffusion regions 41b of the transistor Tp of the present embodiment is different from the film thickness of each of sidewall insulating films 71 of the transistor Tp, and is, for example, wider than the film thickness of each of sidewall insulating films 71 of the transistor Tp. In the present embodiment, the film thickness of each of the sidewall insulating films 71 of the transistor Tn is the same as the film thickness of each of the sidewall insulating films 71 of the transistor Tp.
[0103] Also in the present embodiment, the width Wp of each of the diffusion regions 41b of the transistor Tp is wider than the width Wn of each of the diffusion regions 41b of the transistor Tn. As a result, the area of each of the diffusion regions 41b of the transistor Tp is larger than the area of each of the diffusion regions 41b of the transistor Tn in plan view. As a result, the transconductance gm.sub.p of the transistor Tp can be made smaller, the transconductance gm.sub.n of the transistor Tn can be made larger, and the random noise RN of the solid-state imaging device can be made smaller.
[0104] According to the present embodiment, it is possible to reduce noise of the solid-state imaging device in a suitable manner, similarly to the first embodiment. However, the first embodiment has an advantage that the number of steps of manufacturing the solid-state imaging device can be reduced by using the sidewall insulating films 71 as masks. On the other hand, the present embodiment has an advantage that the film thicknesses of the sidewall insulating films 71 of the transistors Tp and Tn can be unified, and an advantage that noise of the solid-state imaging device can be reduced even in a case where the film thickness of the sidewall insulating films 71 cannot be increased.
Third Embodiment
[0105]
[0106] A and B of
[0107] A of
[0108] Next, the transistor Tn illustrated in A of
[0109] A of
[0110] B of
[0111] Furthermore, the shape and area of each of the diffusion regions 41a illustrated in A of
[0112] However, each of the plugs 46a illustrated in A of
[0113] Therefore, according to the present embodiment, the transconductance gm.sub.p of the transistor Tp can be made smaller and the transconductance gm.sub.n of the transistor Tn can be made larger, similarly to the first embodiment. As a result, the random noise RN of the solid-state imaging device of the present embodiment can be made smaller.
[0114] The structure illustrated in B of
[0115] According to the present embodiment, it is possible to reduce noise of the solid-state imaging device in a suitable manner, similarly to the first and second embodiments. Note that the solid-state imaging device of the present embodiment may include a plurality of plugs 46a on each of the diffusion regions 41a. In this case, the total contact area between each of the diffusion regions 41a illustrated in B or C of
Fourth Embodiment
[0116]
[0117] A and B of
[0118] The structures of the transistors Tn and Tp of the present embodiment illustrated A and B of
[0119] Therefore, according to the present embodiment, the transconductance gm.sub.p of the transistor Tp can be made smaller and the transconductance gm.sub.n of the transistor Tn can be made larger, similarly to the first embodiment. As a result, the random noise RN of the solid-state imaging device of the present embodiment can be made smaller.
[0120] The structure illustrated in B of
[0121] The planar shape of each of the plugs 46a may be any shape. For example, the planar shape of each of the plugs 46a of the transistors Tn and Tp may be a circle. In this case, if the diameter of the circle that is the planar shape of each of the plugs 46a of the transistor Tp is set to be smaller than the diameter of the circle that is the planar shape of each of the plugs 46a of the transistor Tn, the area of each of the plugs 46a of the transistor Tp can be set to be smaller than the area of each of the plugs 46a of the transistor Tn.
[0122] According to the present embodiment, it is possible to reduce noise of the solid-state imaging device in a suitable manner, similarly to the first to third embodiments. Note that the solid-state imaging device of the present embodiment may include a plurality of plugs 46a on each of the diffusion regions 41a. In this case, the total contact area between each of the diffusion regions 41a illustrated in B or C of
[0123] Here, the third embodiment is compared with the fourth embodiment. According to the third embodiment, since it is not necessary to form the plugs 46a each having a small area, it is possible to form the plugs 46a without using an exposure device capable of highly accurate exposure. According to the fourth embodiment, since it is not necessary to form the plugs 46a on the diffusion regions 41a and insulating films, it is possible to reduce short-circuiting of the plugs 46a.
Fifth Embodiment
[0124]
[0125] A and B of
[0126] The structures of the transistors Tn and Tp of the present embodiment illustrated A and B of
[0127] Therefore, according to the present embodiment, the transconductance gm.sub.p of the transistor Tp can be made smaller and the transconductance gm.sub.n of the transistor Tn can be made larger, similarly to the first embodiment. As a result, the random noise RN of the solid-state imaging device of the present embodiment can be made smaller.
[0128] According to the present embodiment, it is possible to reduce noise of the solid-state imaging device in a suitable manner, similarly to the first to fourth embodiments. Note that the plugs 46a illustrated in A and B of
[0129] Here, the third, fourth, and fifth embodiments are compared. According to the third or fifth embodiment, since it is not necessary to form the plugs 46a each having a small area, it is possible to form the plugs 46a without using an exposure device capable of highly accurate exposure. According to the fourth or fifth embodiment, since it is not necessary to form the plugs 46a on the diffusion regions 41a and insulating films, it is possible to reduce short-circuiting of the plugs 46a.
Application Example
[0130]
[0131] The camera 100 includes an optical unit 101 including a lens group and the like, an imaging device 102 that is the solid-state imaging device according to any of the first to fifth embodiments, a digital signal processor (DSP) circuit 103 that is a camera signal processing circuit, a frame memory 104, a display unit 105, a recording unit 106, an operation unit 107, and a power supply unit 108. Furthermore, the DSP circuit 103, the frame memory 104, the display unit 105, the recording unit 106, the operation unit 107, and the power supply unit 108 are connected to each other via a bus line 109.
[0132] The optical unit 101 captures incident light (image light) from a subject and forms an image on an imaging surface of the imaging device 102. The imaging device 102 converts a light amount of incident light formed into an image on the imaging surface by the optical unit 101 into an electric signal on a pixel-by-pixel basis and outputs the electric signal as a pixel signal.
[0133] The DSP circuit 103 performs signal processing on the pixel signal output from the imaging device 102. The frame memory 104 is a memory for storing one screen of a moving image or a still image that is obtained by imaging by the imaging device 102.
[0134] The display unit 105 includes, for example, a panel type display device such as a liquid crystal panel or an organic EL panel, and displays a moving image or a still image that is obtained by imaging by the imaging device 102. The recording unit 106 records a moving image or a still image that is obtained by imaging by the imaging device 102 on a recording medium such as a hard disk or a semiconductor memory.
[0135] The operation unit 107 issues operation commands for various functions of the camera 100 in response to an operation performed by a user. The power supply unit 108 appropriately supplies various power supplies, which are operation power supplies for the DSP circuit 103, the frame memory 104, the display unit 105, the recording unit 106, and the operation unit 107, to these power supply targets.
[0136] It can be expected to acquire a satisfactory image by using the solid-state imaging device according to any of the first to fifth embodiments as the imaging device 102.
[0137] The solid-state imaging device can be applied to various other products. For example, the solid-state imaging device may be included in any type of mobile bodies such as vehicles, electric vehicles, hybrid electric vehicles, motorcycles, bicycles, personal mobility, airplanes, drones, ships, and robots.
[0138]
[0139] The vehicle control system 200 includes a plurality of electronic control units connected to each other via a communication network 201. In the example illustrated in
[0140] The driving system control unit 210 controls the operation of devices related to a driving system of a vehicle in accordance with various types of programs. For example, the driving system control unit 210 functions as a control device for a driving force generating device for generating a driving force of a vehicle, such as an internal combustion engine, a driving motor, or the like, a driving force transmitting mechanism for transmitting the driving force to wheels, a steering mechanism for adjusting a steering angle of the vehicle, a braking device for generating a braking force of the vehicle, and the like.
[0141] The body system control unit 220 controls the operation of various types of devices provided to a vehicle body in accordance with various types of programs. For example, the body system control unit 220 functions as a control device for a smart key system, a keyless entry system, a power window device, or various types of lamps (for example, a head lamp, a backup lamp, a brake lamp, a turn signal, a fog lamp, or the like). In this case, radio waves transmitted from a mobile device as an alternative to a key or signals of various types of switches can be input to the body system control unit 220. The body system control unit 220 receives inputs of such radio waves or signals, and controls a door lock device, the power window device, the lamps, or the like of the vehicle.
[0142] The outside-vehicle information detecting unit 230 detects information of the outside of the vehicle including the vehicle control system 200. The outside-vehicle information detecting unit 230 is connected with, for example, an imaging unit 231. The outside-vehicle information detecting unit 230 makes the imaging unit 231 obtain by imaging an image of the outside of the vehicle, and receives the image that is obtained by imaging from the imaging unit 231. On the basis of the received image, the outside-vehicle information detecting unit 230 may perform processing of detecting an object such as a human, an automobile, an obstacle, a sign, a character on a road surface, or the like, or processing of detecting a distance thereto.
[0143] The imaging unit 231 is an optical sensor that receives light and that outputs an electric signal corresponding to the amount of received light. The imaging unit 231 can output the electric signal as an image, or can output the electric signal as information of a measured distance. The light received by the imaging unit 231 may be visible light, or may be invisible light such as infrared light. The imaging unit 231 includes the solid-state imaging device according to any of the first to fifth embodiments.
[0144] The in-vehicle information detecting unit 240 detects information of the inside of the vehicle including the vehicle control system 200. The in-vehicle information detecting unit 240 is, for example, connected with a driver state detecting unit 241 that detects a state of a driver. For example, the driver state detecting unit 241 includes a camera that images the driver, and on the basis of detection information input from the driver state detecting unit 241, the in-vehicle information detecting unit 240 may calculate a degree of fatigue of the driver or a degree of concentration of the driver, or may determine whether or not the driver is dozing off. The camera may include the solid-state imaging device according to any of the first to fifth embodiments, and may be, for example, the camera 100 illustrated in
[0145] The microcomputer 251 can calculate a control target value for the driving force generating device, the steering mechanism, or the braking device on the basis of the information of the inside or outside of the vehicle obtained by the outside-vehicle information detecting unit 230 or the in-vehicle information detecting unit 240, and output a control command to the driving system control unit 210. For example, the microcomputer 251 can perform cooperative control intended for implementing functions of an advanced driver assistance system (ADAS), the functions including collision avoidance or shock mitigation for the vehicle, following traveling based on a following distance, vehicle speed maintaining traveling, a warning of collision of the vehicle, a warning of deviation of the vehicle from a lane, and the like.
[0146] Furthermore, the microcomputer 251 can perform cooperative control intended for automated driving or the like in which the vehicle travels autonomously without depending on the operation of the driver by controlling the driving force generating device, the steering mechanism, or the braking device on the basis of the information of the surroundings of the vehicle obtained by the outside-vehicle information detecting unit 230 or the in-vehicle information detecting unit 240.
[0147] Furthermore, the microcomputer 251 can output a control command to the body system control unit 220 on the basis of the information of the outside of the vehicle obtained by the outside-vehicle information detecting unit 230. For example, the microcomputer 251 can perform cooperative control intended for preventing glare, such as switching from a high beam to a low beam by controlling the headlamp according to the position of a preceding vehicle or an oncoming vehicle detected by the outside-vehicle information detecting unit 230.
[0148] The sound/image output unit 252 transmits an output signal of at least one of a sound or an image to an output device that can visually or auditorily provide information to an occupant of the vehicle or the outside of the vehicle. In the example of
[0149]
[0150] A vehicle 300 illustrated in
[0151] The imaging unit 301 provided on the front nose mainly acquires an image of the front of the vehicle 300. The imaging unit 302 provided on the left side mirror and the imaging unit 303 provided on the right side mirror mainly acquire images of the sides of the vehicle 300. The imaging unit 304 provided to the rear bumper or the back door mainly acquires an image of the rear of the vehicle 300. The imaging unit 305 provided to the upper portion of the windshield in the interior of the vehicle mainly acquires an image of the front of the vehicle 300. The imaging unit 305 is used to detect, for example, a preceding vehicle, a pedestrian, an obstacle, a signal, a traffic sign, a lane, and the like.
[0152]
[0153] At least one of the imaging units 301 to 304 may include a function of acquiring distance information. For example, at least one of the imaging units 301 to 304 may be a stereo camera including a plurality of imaging devices or an imaging device including pixels for phase difference detection.
[0154] For example, the microcomputer 251 (
[0155] For example, the microcomputer 251 can classify three-dimensional object data related to three-dimensional objects into a two-wheeled vehicle, a standard-sized vehicle, a large-sized vehicle, a pedestrian, a utility pole, and other three-dimensional objects on the basis of the distance information obtained from the imaging units 301 to 304, extract the classified three-dimensional object data, and use the extracted three-dimensional object data for automatic avoidance of an obstacle. For example, the microcomputer 251 identifies obstacles around the vehicle 300 as obstacles that the driver of the vehicle 300 can recognize visually and obstacles that the driver of the vehicle 300 can hardly recognize visually. Then, the microcomputer 251 determines a collision risk indicating a risk of collision with each of the obstacles, and in a situation in which the collision risk is equal to or higher than a set value and there is thus a possibility of collision, the microcomputer 251 outputs a warning to the driver via the audio speaker 261 or the display unit 262, and performs forced deceleration or avoidance steering via the driving system control unit 210 to assist in driving to avoid collision.
[0156] At least one of the imaging units 301 to 304 may be an infrared camera that detects infrared light. The microcomputer 251 can, for example, recognize a pedestrian by determining whether or not there is a pedestrian in images obtained by imaging by the imaging units 301 to 304. Such recognition of a pedestrian is, for example, performed by a procedure of extracting characteristic points in the images obtained by imaging by the imaging units 301 to 304 as infrared cameras and a procedure of determining whether or not an object is a pedestrian by performing pattern matching processing on a series of characteristic points representing the contour of the object. In a case where the microcomputer 251 determines that there is a pedestrian in the images obtained by imaging by the imaging units 301 to 304 and recognizes the pedestrian, the sound/image output unit 252 controls the display unit 262 so that a square contour line for emphasis is displayed in a superimposed manner on the recognized pedestrian. Furthermore, the sound/image output unit 252 may also control the display unit 262 so that an icon or the like representing the pedestrian is displayed at a desired position.
[0157]
[0158]
[0159] The endoscope 500 includes a lens barrel 501 including a region of a predetermined length from a distal end thereof to be inserted into a body cavity of the patient 532, and a camera head 502 connected to a proximal end of the lens barrel 501. Although the illustrated example illustrates the endoscope 500 is configured as a so-called rigid endoscope including a rigid lens barrel 501, the endoscope 500 may be a so-called flexible endoscope including a flexible lens barrel.
[0160] An opening in which an objective lens is fitted is provided at the distal end of the lens barrel 501. A light source device 603 is connected to the endoscope 500, and light generated by the light source device 603 is guided to the distal end of the lens barrel by a light guide extending in the lens barrel 501 and is emitted to an observation target in the body cavity of the patient 532 through the objective lens. Note that the endoscope 500 may be a forward-viewing endoscope, an oblique-viewing endoscope, or a side-viewing endoscope.
[0161] An optical system and an imaging element are provided in the camera head 502, and light reflected by the observation target (observation light) is collected on the imaging element by the optical system. The imaging element photoelectrically converts the observation light and generates an electric signal corresponding to the observation light, that is, an image signal corresponding to an observation image. The image signal is transmitted to a camera control unit (CCU) 601 as RAW data.
[0162] The CCU 601 includes a central processing unit (CPU), a graphics processing unit (GPU), and the like, and integrally controls the operation of the endoscope 500 and a display device 602. Moreover, the CCU 601 receives the image signal from the camera head 502 and applies various types of image processing for displaying an image based on the image signal, for example, development processing (demosaicing processing) and the like on the image signal.
[0163] The display device 602 displays the image based on the image signal which has been subjected to the image processing by the CCU 601 under the control of the CCU 601.
[0164] The light source device 603 includes, for example, a light source such as a light emitting diode (LED), and supplies irradiation light for capturing an image of a surgical site or the like to the endoscope 500.
[0165] An input device 604 is an input interface for the endoscopic surgery system 11000. A user may input various types of information and instructions to the endoscopic surgery system 400 via the input device 604. For example, the user inputs an instruction and the like to change an imaging condition (type of irradiation light, magnification, focal length and the like) by the endoscope 500.
[0166] A treatment tool control device 605 controls driving of the energy treatment tool 512 for tissue cauterization, incision, blood vessel sealing, and the like. A pneumoperitoneum device 606 sends gas into the body cavity of the patient 532 via the pneumoperitoneum tube 511 in order to inflate the body cavity for a purpose of securing a field of view by the endoscope 500 and securing work space for the operator. A recorder 607 is a device that can record various types of information regarding surgery. A printer 608 is a device that can print various types of information regarding surgery in various formats such as a text, an image, or a graph.
[0167] Note that, the light source device 603 which supplies the irradiation light for capturing an image of the surgical site to the endoscope 500 may include, for example, an LED, a laser light source, or a white light source obtained by combining these. In a case where the white light source includes a combination of RGB laser light sources, because an output intensity and an output timing of each color (each wavelength) can be controlled with high accuracy, the light source device 603 can adjust white balance of an image obtained by imaging. Furthermore, in this case, by irradiating the observation target with the laser light from each of the R, G, and B laser light sources in time division and controlling driving of the imaging element of the camera head 502 in synchronism with the irradiation timing, images corresponding to R, G, and B can be obtained by imaging in time division. According to this method, a color image can be obtained even if color filters are not provided for the imaging element.
[0168] Furthermore, the driving of the light source device 603 may be controlled such that the intensity of light to be output is changed every predetermined time. By controlling driving of the imaging element of the camera head 502 in synchronism with the timing of the change of the intensity of light to acquire images time-divisionally and synthesizing the images, an image of a high dynamic range free from underexposed blocked up shadows and overexposed highlights can be created.
[0169] Furthermore, the light source device 603 may be able to supply light in a predetermined wavelength band adapted to special light observation. In the special light observation, for example, by emitting light in a narrower band than irradiation light (in other words, white light) at the time of normal observation using wavelength dependency of a body tissue to absorb light, so-called narrow band imaging is performed in which an image of a predetermined tissue, such as a blood vessel in a mucosal surface layer, is captured with high contrast. Alternatively, in special light observation, fluorescent observation for obtaining an image from fluorescent light generated by irradiation of excitation light may be performed. In fluorescent observation, for example, it is possible to perform observation of fluorescent light from a body tissue by irradiating excitation light on the body tissue (autofluorescence observation) or to obtain a fluorescent light image by locally injecting a reagent such as indocyanine green (ICG) into a body tissue and irradiating excitation light corresponding to a fluorescent light wavelength of the reagent upon the body tissue. The light source device 603 can be configured to supply narrow band light and/or excitation light adapted to such special light observation.
[0170]
[0171] The camera head 502 includes a lens unit 701, an imaging unit 702, a drive unit 703, a communication unit 704, and a camera head control unit 705. The CCU 601 includes a communication unit 711, an image processing unit 712, and a control unit 713. The camera head 502 and the CCU 601 are connected to each other communicably by a transmission cable 700.
[0172] The lens unit 701 is an optical system provided at a connection portion with the lens barrel 501. The observation light captured from the distal end of the lens barrel 501 is guided to the camera head 502 and enters the lens unit 701. The lens unit 701 is configured by combining a plurality of lenses including a zoom lens and a focus lens.
[0173] The imaging unit 702 includes an imaging element. The number of imaging elements included in the imaging unit 702 may be one (so-called single plate type) or two or more (so-called multiple plate type). In a case where the imaging unit 702 is configured as the multiple plate type, for example, image signals corresponding to R, G, and B may be generated by the respective imaging elements, and a color image may be obtained by combining the generated image signals. Alternatively, the imaging unit 702 may include a pair of imaging elements for obtaining right-eye and left-eye image signals corresponding to three-dimensional (3D) display. By performing the 3D display, the operator 531 can grasp a depth of a living body tissue in a surgical site more accurately. Note that, in a case where the imaging unit 702 is configured as the multiple plate type, a plurality of systems of lens units 701 may be provided so as to correspond to the respective imaging elements. The imaging unit 702 is, for example, the solid-state imaging device according to any of the first to fifth embodiments.
[0174] Furthermore, the imaging unit 702 is not necessarily provided in the camera head 502. For example, the imaging unit 702 may be provided inside the lens barrel 501 immediately behind the objective lens.
[0175] The drive unit 703 includes an actuator and moves the zoom lens and the focusing lens of the lens unit 701 by a predetermined distance along an optical axis under the control of the camera head control unit 705. Thus, the magnification and focal point of the image obtained by imaging by the imaging unit 702 may be appropriately adjusted.
[0176] The communication unit 704 includes a communication device for transmitting and receiving various types of information to and from the CCU 601. The communication unit 704 transmits the image signal obtained from the imaging unit 702 as the RAW data to the CCU 601 via the transmission cable 700.
[0177] Furthermore, the communication unit 704 receives a control signal for controlling driving of the camera head 502 from the CCU 601 and supplies the control signal to the camera head control unit 705. The control signal includes, for example, information regarding an imaging condition such as information specifying a frame rate of an image obtained by imaging, information specifying an exposure value at the time of imaging, and/or information specifying the magnification and focal point of the image obtained by imaging.
[0178] Note that the imaging condition such as the frame rate, exposure value, magnification, and focus described above may be appropriately specified by the user, or may be automatically set by the control unit 713 of the CCU 601 on the basis of the acquired image signal. In the latter case, the endoscope 500 includes a so-called auto exposure (AE) function, an auto focus (AF) function, and an auto white balance (AWB) function.
[0179] The camera head control unit 705 controls the driving of the camera head 502 on the basis of the control signal from the CCU 601 received via the communication unit 704.
[0180] The communication unit 711 includes a communication device for transmitting and receiving various types of information to and from the camera head 502. The communication unit 711 receives the image signal transmitted from the camera head 502 via the transmission cable 700.
[0181] Furthermore, the communication unit 711 transmits the control signal for controlling the driving of the camera head 502 to the camera head 502. The image signal and the control signal can be transmitted by electrical communication, optical communication or the like.
[0182] The image processing unit 712 performs various types of image processing on the image signal which is the RAW data transmitted from the camera head 502.
[0183] The control unit 713 performs various types of control regarding imaging of the surgical site and the like by the endoscope 500 and display of the image obtained by imaging that is obtained by the imaging of the surgical site and the like. For example, the control unit 713 generates the control signal for controlling the driving of the camera head 502.
[0184] Furthermore, the control unit 713 causes the display device 602 to display the image obtained by imaging including the surgical site and the like on the basis of the image signal subjected to the image processing by the image processing unit 712. At this time, the control unit 713 may recognize various objects in the image obtained by imaging using various image recognition technologies. For example, the control unit 713 may detect edge shapes, colors, and the like of the objects included in the image obtained by imaging, to recognize the surgical tool such as forceps, a specific living body site, bleeding, mist when the energy treatment tool 512 is used, and the like. At the time of causing the display device 602 to display the image obtained by imaging, the control unit 713 may display various types of surgery assistance information in a superimposed manner on the image of the surgical site using the recognition result. The surgery assistance information is displayed in a superimposed manner, and presented to the operator 531, so that it is possible to reduce the burden on the operator 531 and enable the operator 531 to reliably proceed with surgery.
[0185] The transmission cable 700 that connects the camera head 502 and the CCU 601 is an electric signal cable compatible with communication of electric signals, an optical fiber compatible with optical communication, or a composite cable thereof.
[0186] Here, the communication is performed by wire using the transmission cable 700 in the illustrated example, but the communication between the camera head 502 and the CCU 601 may be performed wirelessly.
[0187] Although the embodiments of the present disclosure have been described above, these embodiments may be implemented with various modifications within a scope not departing from the gist of the present disclosure. For example, two or more embodiments may be implemented in combination.
[0188] Note that the present disclosure can also have the following configurations. [0189] (1) [0190] A solid-state imaging device including: [0191] a first substrate including a photodiode and a transfer transistor; and [0192] a second substrate including an active load circuit and a differential pair circuit for a comparator, [0193] in which the active load circuit includes a first transistor, [0194] the differential pair circuit includes a second transistor, and [0195] transconductance of the first transistor is smaller than transconductance of the second transistor. [0196] (2) [0197] The solid-state imaging device according to (1), in which an area of an LDD region on a source region side and a drain region side of the first transistor is larger than an area of an LDD region on a source region side and a drain region side of the second transistor in plan view, respectively. [0198] (3) [0199] The solid-state imaging device according to (2), in which an area of a source region and a drain region of the first transistor is narrower than an area of a source region and a drain region of the second transistor in plan view, respectively. [0200] (4) [0201] The solid-state imaging device according to (2), in which a film thickness of a sidewall insulating film of the first transistor is thicker than a film thickness of a sidewall insulating film of the second transistor. [0202] (5) [0203] The solid-state imaging device according to (2), in which [0204] a width of an LDD region on a source region side and a drain region side of the first transistor is equal to a film thickness of a sidewall insulating film of the first transistor, and [0205] a width of an LDD region on a source region side and a drain region side of the second transistor is equal to a film thickness of a sidewall insulating film of the second transistor. [0206] (6) [0207] The solid-state imaging device according to (2), in which a film thickness of a sidewall insulating film of the first transistor is equal to a film thickness of a sidewall insulating film of the second transistor. [0208] (7) [0209] The solid-state imaging device according to (2), in which [0210] a width of an LDD region on a source region side and a drain region side of the first transistor is different from a film thickness of a sidewall insulating film of the first transistor, and [0211] a width of an LDD region on a source region side and a drain region side of the second transistor is different from a film thickness of a sidewall insulating film of the second transistor. [0212] (8) [0213] The solid-state imaging device according to (1), in which a total contact area between a source region and a drain region of the first transistor and one or more contact plugs of the first transistor is smaller than a total contact area between a source region and a drain region of the second transistor and one or more contact plugs of the second transistor, respectively. [0214] (9) [0215] The solid-state imaging device according to (8), in which a part of the contact plugs of the first transistor is disposed on an insulating film. [0216] (10) [0217] The solid-state imaging device according to (9), in which the insulating film is a sidewall insulating film of the first transistor. [0218] (11) [0219] The solid-state imaging device according to (9), in which the insulating film is an element isolation insulating film surrounding a source region and a drain region of the first transistor in plan view. [0220] (12) [0221] The solid-state imaging device according to (8), in which an area of each contact plug of the first transistor is smaller than an area of each contact plug of the second transistor in plan view. [0222] (13) [0223] The solid-state imaging device according to (12), in which a shape of each contact plug of the first transistor is similar to a shape of each contact plug of the second transistor in plan view. [0224] (14) [0225] The solid-state imaging device according to (12), in which a shape of each contact plug of the first transistor is not similar to a shape of each contact plug of the second transistor in plan view. [0226] (15) [0227] The solid-state imaging device according to (8), in which a number of the contact plugs of the first transistor is smaller than a number of the contact plugs of the second transistor. [0228] (16) [0229] The solid-state imaging device according to (15), in which an area of each contact plug of the first transistor is equal to an area of each contact plug of the second transistor in plan view. [0230] (17) [0231] The solid-state imaging device according to (1), in which [0232] the first transistor is a transistor of a first conductivity type, and [0233] the second transistor is a transistor of a second conductivity type different from the first conductivity type. [0234] (18) [0235] The solid-state imaging device according to (1), in which the first transistor and the second transistor are provided on a same semiconductor substrate. [0236] (19) [0237] The solid-state imaging device according to (1), in which the first substrate is disposed on the second substrate, and the second substrate is disposed on a third substrate. [0238] (20) [0239] An electronic device including an imaging device, the imaging device including: [0240] a first substrate including a photodiode and a transfer transistor; and [0241] a second substrate including an active load circuit and a differential pair circuit for a comparator, [0242] in which the active load circuit includes a first transistor, [0243] the differential pair circuit includes a second transistor, and [0244] transconductance of the first transistor is smaller than transconductance of the second transistor.
REFERENCE SIGNS LIST
[0245] 1 Pixel array unit [0246] 2 Input unit [0247] 2a Input terminal [0248] 2b Input circuit [0249] 2c Amplitude change unit [0250] 2d Data conversion unit [0251] 3 Timing control unit [0252] 4 Row drive unit [0253] 5 Column signal processing unit [0254] 6 Image signal processing unit [0255] 7 Output unit [0256] 7a Data conversion unit [0257] 7b Amplitude change unit [0258] 7c Output circuit [0259] 7d Output terminal [0260] 11 Pixel [0261] 11a Pixel [0262] 11b Pixel [0263] 11c Pixel [0264] 11d Pixel [0265] 12 Pixel sharing unit [0266] 13 Row drive signal line [0267] 14 Column read line [0268] 21 First substrate [0269] 22 Second substrate [0270] 23 Third substrate [0271] 24 Filter layer [0272] 25 On-chip lens layer [0273] 26 Through plug [0274] 31 Semiconductor substrate [0275] 31a n-type region [0276] 31b p-type region [0277] 31c Floating diffusion portion [0278] 32 Element isolation insulating film [0279] 33 Gate insulating film [0280] 34 Gate electrode [0281] 35 Electrode portion [0282] 36 Interlayer insulating film [0283] 41 Semiconductor substrate [0284] 41a Diffusion region [0285] 41b Diffusion region [0286] 41c Diffusion region [0287] 41d Diffusion region [0288] 42 Gate insulating film [0289] 43 Gate electrode [0290] 44 Interlayer insulating film [0291] 45 Interlayer insulating film [0292] 46a Plug [0293] 46b Plug [0294] 46c Plug [0295] 46d Plug [0296] 47a Wiring layer [0297] 47b Wiring layer [0298] 47c Wiring layer [0299] 48 Pad [0300] 51 Semiconductor substrate [0301] 51a Diffusion region [0302] 52 Gate insulating film [0303] 53 Gate electrode [0304] 54 Interlayer insulating film [0305] 55 Interlayer insulating film [0306] 56a Plug [0307] 56b Plug [0308] 56c Plug [0309] 57a Wiring layer [0310] 57b Wiring layer [0311] 58 Pad [0312] 61 Comparator [0313] 62 Active load circuit [0314] 63 Differential pair circuit [0315] 71 Sidewall insulating film [0316] 73 Element isolation Insulating film