US20190260951A1 - Image-capturing device and image sensor - Google Patents
Image-capturing device and image sensor Download PDFInfo
- Publication number
- US20190260951A1 US20190260951A1 US16/401,436 US201916401436A US2019260951A1 US 20190260951 A1 US20190260951 A1 US 20190260951A1 US 201916401436 A US201916401436 A US 201916401436A US 2019260951 A1 US2019260951 A1 US 2019260951A1
- Authority
- US
- United States
- Prior art keywords
- pixels
- pixel
- image
- pixel group
- control unit
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- H04N5/3696—
-
- H—ELECTRICITY
- H10—SEMICONDUCTOR DEVICES; ELECTRIC SOLID-STATE DEVICES NOT OTHERWISE PROVIDED FOR
- H10F—INORGANIC SEMICONDUCTOR DEVICES SENSITIVE TO INFRARED RADIATION, LIGHT, ELECTROMAGNETIC RADIATION OF SHORTER WAVELENGTH OR CORPUSCULAR RADIATION
- H10F39/00—Integrated devices, or assemblies of multiple devices, comprising at least one element covered by group H10F30/00, e.g. radiation detectors comprising photodiode arrays
- H10F39/80—Constructional details of image sensors
- H10F39/802—Geometry or disposition of elements in pixels, e.g. address-lines or gate electrodes
- H10F39/8027—Geometry of the photosensitive area
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B7/00—Mountings, adjusting means, or light-tight connections, for optical elements
- G02B7/28—Systems for automatic generation of focusing signals
- G02B7/34—Systems for automatic generation of focusing signals using different areas in a pupil plane
-
- H01L27/14607—
-
- H01L27/14621—
-
- H01L27/14627—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/67—Focus control based on electronic image sensor signals
- H04N23/672—Focus control based on electronic image sensor signals based on the phase difference signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/80—Camera processing pipelines; Components thereof
- H04N23/84—Camera processing pipelines; Components thereof for processing colour signals
- H04N23/843—Demosaicing, e.g. interpolating colour pixel values
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/10—Circuitry of solid-state image sensors [SSIS]; Control thereof for transforming different wavelengths into image signals
- H04N25/11—Arrangement of colour filter arrays [CFA]; Filter mosaics
- H04N25/13—Arrangement of colour filter arrays [CFA]; Filter mosaics characterised by the spectral characteristics of the filter elements
- H04N25/134—Arrangement of colour filter arrays [CFA]; Filter mosaics characterised by the spectral characteristics of the filter elements based on three different wavelength filter elements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/70—SSIS architectures; Circuits associated therewith
- H04N25/703—SSIS architectures incorporating pixels for producing signals other than image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/70—SSIS architectures; Circuits associated therewith
- H04N25/703—SSIS architectures incorporating pixels for producing signals other than image signals
- H04N25/704—Pixels specially adapted for focusing, e.g. phase difference pixel sets
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/70—SSIS architectures; Circuits associated therewith
- H04N25/76—Addressed sensors, e.g. MOS or CMOS sensors
- H04N25/78—Readout circuits for addressed sensors, e.g. output amplifiers or A/D converters
-
- H04N5/23212—
-
- H04N5/378—
-
- H—ELECTRICITY
- H10—SEMICONDUCTOR DEVICES; ELECTRIC SOLID-STATE DEVICES NOT OTHERWISE PROVIDED FOR
- H10F—INORGANIC SEMICONDUCTOR DEVICES SENSITIVE TO INFRARED RADIATION, LIGHT, ELECTROMAGNETIC RADIATION OF SHORTER WAVELENGTH OR CORPUSCULAR RADIATION
- H10F39/00—Integrated devices, or assemblies of multiple devices, comprising at least one element covered by group H10F30/00, e.g. radiation detectors comprising photodiode arrays
- H10F39/10—Integrated devices
- H10F39/12—Image sensors
- H10F39/18—Complementary metal-oxide-semiconductor [CMOS] image sensors; Photodiode array image sensors
- H10F39/182—Colour image sensors
-
- H—ELECTRICITY
- H10—SEMICONDUCTOR DEVICES; ELECTRIC SOLID-STATE DEVICES NOT OTHERWISE PROVIDED FOR
- H10F—INORGANIC SEMICONDUCTOR DEVICES SENSITIVE TO INFRARED RADIATION, LIGHT, ELECTROMAGNETIC RADIATION OF SHORTER WAVELENGTH OR CORPUSCULAR RADIATION
- H10F39/00—Integrated devices, or assemblies of multiple devices, comprising at least one element covered by group H10F30/00, e.g. radiation detectors comprising photodiode arrays
- H10F39/80—Constructional details of image sensors
- H10F39/805—Coatings
- H10F39/8053—Colour filters
-
- H—ELECTRICITY
- H10—SEMICONDUCTOR DEVICES; ELECTRIC SOLID-STATE DEVICES NOT OTHERWISE PROVIDED FOR
- H10F—INORGANIC SEMICONDUCTOR DEVICES SENSITIVE TO INFRARED RADIATION, LIGHT, ELECTROMAGNETIC RADIATION OF SHORTER WAVELENGTH OR CORPUSCULAR RADIATION
- H10F39/00—Integrated devices, or assemblies of multiple devices, comprising at least one element covered by group H10F30/00, e.g. radiation detectors comprising photodiode arrays
- H10F39/80—Constructional details of image sensors
- H10F39/806—Optical elements or arrangements associated with the image sensors
- H10F39/8063—Microlenses
-
- H04N9/045—
Definitions
- the present invention relates to an image-capturing device and an image sensor.
- An image-capturing device which performs focus detection by a split-pupil phase detection method on the basis of output signals from a plurality of pixels dedicated for focus detection arranged on a part of an image sensor (see Patent Literature 1).
- Patent Literature 1 Japanese Patent Application Laid-open No. 2007-282109.
- the focus detection is possible only at positions where pixels dedicated for focus detection are arranged.
- an increased number of pixels dedicated for focus detection results in a decrease in image quality since no image signal is obtained from the positions where the pixels dedicated for focus detection are arranged.
- an image-capturing device comprises: an image sensor that captures an image of a subject with light fluxes from the subject that have passed through an imaging optical system; an image generation unit that generates an image signal based upon an output signal from the image sensor; and a focus detection unit that detects a focusing condition of the imaging optical system according to a phase detection method based upon an output signal from the image sensor.
- the image sensor includes an upper layer pixel group and a lower layer pixel group that receives the light fluxes from the subject that have passed through each pixel in the upper layer pixel group, and a microlens group arranged so as to guide the light fluxes from the subject to the upper layer pixel group.
- the upper layer pixel group includes first, second and third pixels having first, second and third spectral sensitivities, respectively, differing from each other, and being arranged in a two-dimensional pattern, with one of the first pixels, one of the second pixels and two of the third pixels being arranged in a two-by-two matrix behind each of microlenses in the microlens group, and the four pixels receive four light fluxes, respectively, that pass through four pupil areas, respectively, of an exit pupil of the imaging optical system.
- the lower layer pixel group includes fourth, fifth and sixth pixels having fourth, fifth and sixth spectral sensitivities, respectively, that are complementary to the first, second, and third spectral sensitivities, respectively, of the upper layer pixel group, being arranged in a two-dimensional pattern. Positions of the first, second and third pixels in the upper layer pixel group and positions of the fourth, fifth and sixth pixels in the lower layer pixel group are determined such that the fourth, fifth and sixth pixels receive light fluxes that pass through the first, second and third pixels, respectively.
- the image generation unit generates the image signal based upon an output signal from one of the upper layer pixel group and the lower layer pixel group.
- the focus detection unit detects the focusing condition based upon an output signal from other of the upper layer pixel group and the lower layer pixel group.
- the first, second, and third pixels are arranged such that pixels having substantially same spectral sensitivities are adjacently arranged in a two-by-two matrix, and four pixels adjacent to the two-by-two matrix are arranged behind four different microlenses in the microlens group, respectively, and at different positions with respect to the microlenses.
- the first, second and third pixels provide output signals relating to cyan, yellow, and magenta, respectively
- the fourth, fifth and sixth pixels provide output signals relating to colors complementary to the cyan, yellow and magenta, respectively.
- the upper layer pixel group and the lower layer pixel group each include an array of a plurality of sets of pixels arranged in a two-dimensional pattern, each of the plurality of sets of pixels having four pixels arranged in a two-by-two matrix behind any particular one of the microlenses and the sets include first through fourth sets having different arrangements of pixels
- the upper layer pixel group is configured such that in the first set, the first pixel and the third pixel are adjacently arranged in a predetermined array direction and the third pixel and the second pixel are arranged adjacent to the first pixel and the third pixel, respectively, in a direction perpendicular to the predetermined array direction, in the second set, the third pixel and the first pixel are adjacently arranged in the predetermined array direction and the second pixel and the third pixel are arranged adjacent to the third pixel and the first pixel, respectively, in the direction perpendicular to the predetermined array direction, in the second set, the third pixel and the first pixel are adjacently arranged in the predetermined array direction and
- the image generation unit adds output signals from four of the fourth pixels that are adjacent to each other in a form of a two-by-two matrix, adds output signals from four of the fifth pixels that are adjacent to each other in a form of a two-by-two matrix, and adds output signals from four of the sixth pixels that are adjacent to each other in a form of a two-by-two matrix to generate an image signal of a Bayer arrangement.
- the image generation unit obtains three color signals at a position corresponding to each microlens based upon output signals from the fourth, fifth and sixth pixels positioned behind each microlens.
- the image generation unit executes, at respective positions of the fourth through sixth pixels, color interpolation processing for generating signals of other two spectral components to obtain three color signals and generates a luminance signal and color difference signals based upon the three color signals thus obtained.
- the focus detection unit detects the focusing condition of the imaging optical system based upon output signals from a pair of pixels having substantially the same spectral sensitivities and located at positions differing from each other with respect to the microlens, out of the upper layer pixel group or the lower layer pixel group.
- the focus detection unit detects the focusing condition of the imaging optical system in the predetermined array direction based upon output signals from at least one plurality of the third pixels of a plurality of the third pixels included in the first set and the second set, respectively, and a plurality of the third pixels included in the third set and the fourth set, respectively, in the upper layer pixel group.
- the focus detection unit detects the focusing condition of the imaging optical system in the direction perpendicular to the predetermined array direction based upon output signals from at least one plurality of the third pixels of a plurality of the third pixels included in the first set and the third set, respectively, and a plurality of the third pixels included in the second set and the fourth set, respectively, in the upper layer pixel group.
- the focus detection unit detects the focusing condition of the imaging optical system in a direction oblique to the predetermined array direction based upon output signals from at least one plurality of the third pixels of a plurality of the third pixels included in the first set and the fourth set, respectively, and a plurality of the third pixels included in the second set and the third set, respectively, in the upper layer pixel group.
- an image sensor comprises: a first image-capturing unit that includes a plurality of microlenses arranged in a two-dimensional pattern; and a plurality of light reception units that are provided in correspondence to each of the microlenses, respectively, and that receive light having a predetermined wavelength and transmit lights having wavelengths different from the predetermined wavelength; and a second image-capturing unit that receives the lights that are transmitted through the first image-capturing units.
- the first imaging unit includes light reception units that are arranged adjacent to each other so as to be provided in correspondence to adjacent two of the microlenses and that absorb light having same wavelength corresponding to the predetermined wavelength.
- image signal generation and focus detection according to a phase detection method can be performed based upon output signals from the image sensor without providing the image sensor with pixels dedicated for focus 3 o detection.
- FIG. 1 A diagram illustrating by an example the configuration of a digital camera system according to an embodiment of the present invention.
- FIG. 2 A plan view illustrating by an example the layout of pixels in an upper photoelectric conversion layer.
- FIG. 3 A plan view illustrating by an example the layout of pixels in a lower photoelectric conversion layer.
- FIG. 4 A diagram illustrating by an example a cross-section of an image sensor.
- FIG. 5 A diagram illustrating by an example a circuit configuration per pixel in the image sensor.
- FIG. 6 A diagram illustrating by an example an exit pupil of an interchangeable lens.
- FIG. 7 A diagram illustrating pixel rows for obtaining a defocus amount.
- FIG. 8 A diagram illustrating light fluxes passing through the exit pupil.
- FIG. 9 A diagram illustrating pixel rows used for obtaining a defocus amount.
- FIG. 10 A diagram illustrating light fluxes passing through the exit pupil.
- FIG. 11 A diagram illustrating pixel rows used for obtaining a defocus amount.
- FIG. 12 A diagram illustrating light fluxes passing through the exit pupil.
- FIG. 13 A diagram illustrating first image signal generation processing.
- FIG. 14 A diagram illustrating second image signal generation processing.
- FIG. 15 A diagram illustrating third image signal generation processing.
- FIG. 16 A diagram illustrating third image signal generation processing.
- FIG. 17 A diagram illustrating third image signal generation processing.
- FIG. 18 A flowchart illustrating the flow of imaging processing.
- FIG. 1 presents a diagram illustrating by an example the configuration of a digital camera system according to an embodiment of the present invention.
- a digital camera system 1 includes an interchangeable lens 2 and a camera body 3 .
- the interchangeable lens 2 is mounted on the camera body 3 via a mount unit 4 .
- the interchangeable lens 2 includes a lens control unit 5 , a main lens 9 , a zoom lens 8 , a focusing lens 7 , and a diaphragm 6 .
- the lens control unit 5 includes a microcomputer, a memory and so on and performs drive control of the focusing lens 7 and the diaphragm 6 , detection of an open state of the diaphragm 6 , detection of positions of the zoom lens 8 and the focusing lens 7 , transmission of information about lens to a body control unit 14 on the side of the camera body 3 described later, reception of information about camera from the body control unit 14 , and so on.
- the camera body 3 includes an image sensor 12 , an image sensor drive control unit 19 , the body control unit 14 , a liquid crystal display element drive circuit 15 , a liquid crystal display element 16 , an ocular lens 17 , an operating member 18 and so on.
- a detachable memory card 20 is attached to the camera body 3 .
- the image sensor 12 is arranged on a predetermined imaging plane of the interchangeable lens 2 and captures an image of a photographic subject that is formed by the interchangeable lens 2 .
- the body control unit 14 includes a microcomputer, a memory and so on.
- the body control unit 14 controls operations of the digital camera system in whole.
- the body control unit 14 and the lens control unit 5 are configured to communicate with each other via an electric contact unit 13 of the mount unit 4 .
- the image sensor drive control unit 19 generates a control signal that is necessary for the image sensor 12 in response to a command from the body control unit 14 .
- the liquid crystal display element drive circuit 15 drives the liquid crystal display element 16 that constitutes a liquid crystal view finder (EVF: electric view finder) in response to a command from the body control unit 14 .
- EMF electric view finder
- the photographer observes an image displayed on the liquid crystal display element 16 through the ocular lens 17 .
- the memory card 20 is a storage medium in which image signals and the like are stored.
- the image of the photographic subject formed on the image sensor 12 by the interchangeable lens 2 is subjected to photoelectric conversion by the image sensor 12 .
- the image sensor 12 is controlled by a control signal from the image sensor drive control unit 19 with respect to timing (frame rate) of storage of photoelectric conversion signals and of reading of signals.
- the image signals read out from the image sensor 12 are converted in an A/D conversion unit (not shown) into digital data, which then is transmitted to the body control unit 14 .
- the body control unit 14 calculates a defocus amount on the basis of image signals from the image sensor 12 corresponding to predetermined focus detection areas, respectively, and transmits the calculated defocus amount to the lens control unit 5 .
- the lens control unit 5 calculates a focusing lens drive amount on the basis of the defocus amount received from the body control unit 14 and drives the focusing lens 7 by a motor or the like, which is not shown, to move the focusing lens 7 to a focusing position on the basis of the calculated lens drive amount.
- the body control unit 14 generates image data to be recorded on the basis of a signal that is output from the image sensor 12 after shooting is commanded.
- the body control unit 14 stores the generated image data in the memory card 20 and at the same time transmits the generated image data to the liquid crystal display element drive circuit and controls it to be reproduced and displayed on the liquid crystal display element 16 .
- the camera body 3 is provided with the operating member 18 that includes a shutter button and a setting member for setting a focus detection area or areas and so on.
- the body control unit 14 detects an operating signal from the operating member 18 and controls the operations (photographing processing, setting of a focus detection area and so on) in response to the result of the detection.
- the image sensor 12 has a stacked structure, in which an upper photoelectric conversion layer 41 ( FIG. 4 ) and a lower photoelectric conversion layer 43 ( FIG. 4 ) are stacked one above another.
- the upper photoelectric conversion layer 41 is constituted by a photoconductive film that absorbs (photoelectrically converts) light having wavelength components described later and light having wavelength components not absorbed (photoelectrically converted) by the upper photoelectric conversion layer 41 is transmitted to the lower photoelectric conversion layer 43 , where the transmitted light is photoelectrically converted.
- FIGS. 2( a ) and ( b ) presents plan views each illustrate by an example the layout of pixels in the upper photoelectric conversion layer 41 of the image sensor 12 .
- 10 ⁇ 10 pixels which are extracted as representatives, are shown.
- the extracted pixels are laid out in a substantially square form and arranged in a two-dimensional pattern.
- the extracted pixels include three types of pixels, i.e., a pixel that photoelectrically converts light of a cyan (Cy) color component (Cy pixel), a pixel that photoelectrically converts light of a magenta (Mg) color component (Mg pixel), and a pixel that photoelectrically converts light of a yellow (Ye) color component (Ye pixel).
- the Cy pixel is constituted by a photoelectric conversion unit that absorbs (photoelectrically converts) the light of the cyan color component.
- the Mg pixel is constituted by a photoelectric conversion unit that absorbs (photoelectrically converts) the light of the magenta color component.
- the Ye pixel is constituted by a photoelectric conversion unit that absorbs (photoelectrically converts) the light of the yellow color component.
- the image sensor 12 is formed of a plurality of microlenses 40 each of which is to efficiently guide light fluxes from the interchangeable lens 2 to a set of four pixels.
- the microlenses 40 each constituted by a spherical lens of axial symmetry whose center substantially coincides with its optical axis or a nonspherical lens, are arranged in a two-dimensional pattern, with the light incident side thereof having a convex shape.
- each of the microlenses 40 one Cy pixel, two Mg pixels and one Ye pixel are arranged in a two-by-two matrix.
- a plurality of sets of four pixels positioned behind the corresponding microlenses 40 are classified into four types (P 1 through P 4 ) according to differences in their arrangements as shown in FIG. 2( a ) .
- first sets P 1 each include a Cy pixel at a left upper position, an Mg pixel at a right upper position, an Mg pixel at a left lower position, and a Ye pixel at a right lower position.
- Second sets P 2 each include an Mg pixel at a left upper position, a Cy pixel at a right upper position, a Ye pixel at a left lower position, and an Mg pixel at a right lower position.
- Third sets P 3 each include an Mg pixel at a left upper position, a Ye pixel at a right upper position, a Cy pixel at a left lower position, and an Mg pixel at a right lower position.
- Fourth sets P 4 each include a Ye pixel at a left upper position, an Mg pixel at a right upper position, an Mg pixel at a left lower position, and a Cy pixel at a right lower position.
- the first sets P 1 and the second sets P 2 are adjacent to each other in a horizontal direction (X direction) and arranged alternately in a repeated manner in the horizontal direction.
- a line formed by the first sets P 1 and the second sets P 2 is called a “first line L 1 ”.
- the third sets P 3 and the fourth sets P 4 are adjacent to each other in the horizontal direction and are arranged alternately in a repeated manner in the horizontal direction.
- a line formed by the third sets P 3 and the fourth sets P 4 is called a “second line L 2 ”.
- the first line L 1 and the second line L 2 described above are adjacent to each other in a vertical direction (Y direction) and are alternately arranged in a repeated manner in the vertical direction.
- Y direction vertical direction
- each of the first sets P 1 and each of the third sets P 3 are adjacent to each other in the vertical direction
- each of the second sets P 2 and each of the fourth sets P 4 are adjacent to each other in the vertical direction.
- the microlenses 40 and the Cy pixels, the Mg pixels, and the Ye pixels have the following positional relationships.
- the Cy pixels, behind four microlenses 40 adjacent to each other in the horizontal direction and in the vertical direction are arranged at a left upper position, a right upper position, a left lower position, and a right lower position, respectively.
- the Mg pixels, behind four microlenses 40 adjacent to each other in the horizontal direction and in the vertical direction are arranged at right upper and left lower positions, left upper and right lower positions, left upper and right lower positions, and right upper and left lower positions, respectively.
- the Ye pixels, behind four microlenses 40 adjacent to each other in the horizontal direction and in the vertical direction are arranged at a right lower position, a left lower position, a right upper position, and a left upper position, respectively. In this manner, the Cy pixels, the Mg pixels, and the Ye pixels are uniformly arranged behind the microlenses 40 without being arranged disproportionately to specific positions.
- FIG. 2( b ) is a diagram showing a part extracted from the upper photoelectric conversion layer of the image sensor, which part is similar to that shown in FIG. 2( a ) .
- the Cy. Mg and Ye pixels each are arranged such that adjacent four pixels in any particular two-by-two matrix have the same color as shown in FIG. 2( b ) .
- the four pixels in any particular two-by-two matrix having the same color are arranged behind different microlenses 40 , respectively, so that they assume different positions with respect to the microlenses 40 .
- the Cy, Mg and Ye pixels arranged behind the four microlenses 40 are arranged such that they are adjacent to each other in a two-by-two matrix for each color.
- FIGS. 3( a ) and ( b ) presents plan views each illustrate by an example the layout of pixels in the lower photoelectric conversion layer 43 ( FIG. 4 ) of the image sensor 12 .
- 10 ⁇ 10 pixels which correspond to the positions of the pixels exemplified in FIG. 2 .
- These pixels are laid out in a substantially square form and arranged in a two-dimensional pattern. They include three types of pixels, i.e., a pixel that photoelectrically converts light of a red (R) color component (R pixel), a pixel that photoelectrically converts light of a green (G) color component (G pixel), and a pixel that photoelectrically converts light of a blue (B) color component (B pixel).
- RGB red
- G green
- B blue
- the R pixel is constituted by a photoelectric conversion unit that photoelectrically converts light of the red color component (i.e., a complementary color of Cy) that has not been absorbed (photoelectrically converted) by the Cy pixel lying above.
- the G pixel is constituted by a photoelectric conversion unit that photoelectrically converts light of the green color component (i.e., a complementary color of Mg) that has not been absorbed (photoelectrically converted) by the Mg pixel lying above.
- the B pixel is constituted by a photoelectric conversion unit that photoelectrically converts light of the blue color component (i.e., complementary color of Ye) that has not been absorbed (photoelectrically converted) by the Ye pixel lying above.
- the Cy, Mg and Ye pixels in the upper photoelectric conversion layer 43 constitute R, G and B light reception units, respectively, that serve as color filters.
- each of sets (referred to as Q 1 ) underlying the first sets P 1 includes an R pixel at a left upper position, a G pixel at a right upper position, a G pixel at a left lower position, and a B pixel at a right lower position.
- Each of sets (referred to as Q 2 ) underlying the second sets P 2 respectively, includes a G pixel at a left upper position, an R pixel at a right upper position, a B pixel at a left lower position, and a G pixel at a right lower position.
- Each of sets (referred to as Q 3 ) underlying the third sets P 3 respectively, includes a G pixel at a left upper position, a B pixel at a right upper position, an R pixel at a left lower position, and a G pixel at a right lower position.
- Each of sets (referred to as Q 4 ) underlying the fourth sets P 4 respectively, includes a B pixel at a left upper position, a G pixel at a right upper position, a G pixel at a left lower position, and an R pixel at a right lower position.
- the microlenses 40 and the R pixels, the G pixels, and the B pixels have the following positional relationships.
- the R pixels, behind four microlenses 40 that are adjacent to each other in the horizontal direction and in the vertical direction are arranged at a left upper position, a right upper position, a left lower position, and a right lower position, respectively.
- the G pixels, behind four microlenses 40 that are adjacent to each other in the horizontal direction and in the vertical direction are arranged at right upper and left lower positions, left upper and right lower positions, left upper and right lower positions, and right upper and left lower positions, respectively.
- the B pixels, behind four microlenses 40 that are adjacent to each other in the horizontal direction and in the vertical direction are arranged at a right lower position, a left lower position, a right upper position, and a left upper position, respectively. In this manner, the R pixels, the G pixels, and the B pixels are uniformly arranged behind the microlenses 40 without being arranged disproportionately to specific positions.
- FIG. 3( b ) is a diagram showing a part extracted from the lower photoelectric conversion layer of the image sensor, which part is similar to that shown in FIG. 3( a ) .
- the R, G and B pixels each are arranged such that adjacent four pixels in any particular two-by-two matrix have the same color as shown in FIG. 3( b ) .
- the four pixels in any particular two-by-two matrix having the same color are arranged behind different microlenses 40 , respectively, so that they assume different positions with respect to the microlenses 40 .
- the R, G and B pixels arranged behind the four microlenses 40 are arranged such that they are adjacent to each other in a two-by-two matrix for each color.
- the sets constituted by four pixels in a two-by-two matrix of the same color i.e., a set 50 r constituted by four R pixels, a set 50 g constituted by four G pixels, and a set 50 b constituted by four B pixels, when the four pixels are viewed as one set, each form a Bayer arrangement.
- FIG. 4 presents a diagram illustrating by an example a cross-section of the image sensor 12 .
- the image sensor 12 includes the lower photoelectric conversion layer 43 formed on a silicone substrate and the upper photoelectric conversion layer 41 stacked on the lower photoelectric conversion layer 43 via a wiring layer 42 .
- Above the upper photoelectric conversion layer 41 are formed the microlenses 40 .
- the upper photoelectric conversion layer 41 includes electrodes and photoconductive films each sandwiched by the electrodes to constitute the Cy, Mg and Ye pixels. For instance, an upper electrode a and a lower electrode k-Cy sandwich therebetween a photoconductive film P 1 -Cy to constitute a Cy pixel in the first set P 1 . In addition, for instance, the upper electrode a and a lower electrode k-Mg sandwich therebetween a photoconductive film P 2 -Mg to constitute an Mg pixel in the second set P 2 .
- the lower photoelectric conversion layer 43 is constituted by the R pixel, the G pixel, and the B pixel on a silicone substrate and photoelectrically converts light that enters each of the pixels.
- the R pixel in the first set Q 1 receives complementary light (R) that is transmitted through the Cy pixel in the first set P 1 of the upper photoelectric conversion layer 41 .
- the G pixel in the second set Q 2 receives complementary light (G) that is transmitted through the Mg pixel in the second set P 2 of the upper photoelectric conversion layer 41 .
- FIG. 5 illustrates by an example a circuit configuration per pixel in the image sensor 12 .
- a reference voltage source Vref is provided to reference voltage source terminals t 32 a and t 32 b .
- a voltage source Vcc is provided to the voltage source terminals t 31 a and t 31 b .
- a voltage source Vpc is provided from a terminal t 33 to a PC (photoconductor) 20 .
- a signal detection unit of the upper photoelectric conversion layer 41 has the following configuration.
- the PC 20 constitutes a photoelectric conversion unit for one pixel of the upper photoelectric conversion layer 41 .
- incident light is photoelectrically converted into a charge, so that charges are accumulated therein.
- a source follower amplifier MOS transistor Tr 6 amplifies a voltage signal based upon the accumulated charges.
- a transfer switch MOS transistor Tr 5 constitutes a switch for selecting a target pixel to be read out. When a control pulse signal ⁇ SEL#1 that turns on/off the transfer switch MOS transistor Tr 5 is provided to a terminal t 11 , an amplified signal is read out from the terminal t 11 via the transfer switch MOS transistor Tr 5 .
- a reset MOS transistor Tr 7 discharges unnecessary charges in response to a reset pulse signal ⁇ R#1 provided to a terminal t 13 (in other words, resets to a predetermined potential).
- a signal detection unit of the lower photoelectric conversion layer 43 has the following configuration.
- a photodiode PD constitutes a photoelectric conversion unit for one pixel of the lower photoelectric conversion layer 43 .
- the photodiode PD photoelectrically converts light that is transmitted through the PC 20 to generate charges.
- the photodiode PD and a floating diffusion (FD) unit are connected with each other through the transfer MOS transistor Tr 4 .
- a control pulse signal ⁇ Tx#2 that turns on/off the transfer switch MOS transistor Tr 4 is provided to a terminal t 24 , charges are transferred to the floating diffusion unit via the transfer MOS transistor Tr 4 .
- a source follower amplifier MOS transistor Tr 2 amplifies a voltage signal based upon the accumulated charges.
- a transfer switch MOS transistor Tr 1 constitutes a switch for selecting a target pixel to be read out.
- a control pulse signal ⁇ SEL#2 that turns on/off the transfer switch MOS transistor Tr 1 is provided to a terminal t 22
- an amplified signal is read out from a terminal t 21 via the transfer switch MOS transistor Tr 1 .
- a reset MOS transistor Tr 3 discharges unnecessary charges in the floating diffusion unit in response to a reset pulse signal+R#1 provided to a terminal t 13 (in other words, resets it to a predetermined potential).
- FIG. 6 presents a diagram illustrating by an example an exit pupil 80 of the interchangeable lens 2 in a state in which the diaphragm is open.
- Light fluxes that have passed through four regions 81 through 84 of the exit pupil 80 enter pixels located at positions corresponding to a left upper part, a right upper part, a left lower part, and a right lower part, respectively, of each of the microlenses 40 in FIG. 2 .
- correspondence relationship between the light fluxes that enter the pixels located at positions corresponding to the left upper, right upper, left lower, and right lower parts of the microlens and the first region 81 , the second region 82 , the third region 83 , and the fourth region 84 , respectively, of the exit pupil 80 may be considered such that the up and down relation as well as the left and right relation are inverted with respect to the light axis Ax of the interchangeable lens 2 as an axis of symmetry.
- the pixel row 90 is constituted by an Mg pixel (Mg-a) that is included in each second set P 2 and located at the left upper position of each microlens 40 and an Mg pixel (Mg-b) that is included in each first set P 1 and located at the right upper position of each microlens 40 .
- Mg-a Mg pixel
- Mg-b Mg pixel
- a light flux A that pass through a first region 81 on the exit pupil 80 and a light flux B that pass through a second region 82 on the exit pupil 80 enter the pixels that constitute the pixel row 90 .
- the light flux A enters the Mg pixel (Mg-a) located at the left upper position of each microlens 40 .
- the light flux B enters the Mg pixel (Mg-b) located at the right upper position of each microlens 40 .
- the image sensor 12 Upon focusing, the image sensor 12 is in a state in which a sharp image is formed, so that as described above, a pair of images formed by light fluxes through different positions of the pupil as a result of pupil-splitting coincide with each other on the image sensor 12 .
- a signal waveform (signal sequence a 1 , a 2 , a 3 , a 4 , . . . ) obtained from the Mg pixels (Mg-a) that receive the light fluxes A
- a signal waveform (signal sequence b 1 , b 2 , b 3 , b 4 , . . . ) obtained from the Mg pixels (Mg-b) that receive the light fluxes B overlap in their shape.
- the signal waveform (signal sequence a 1 , a 2 , a 3 , a 4 . . . ) by the light fluxes A and the signal waveform (signal sequence b 1 , b 2 , b 3 , b 4 , . . . ) by the light fluxes B have different positional relationships (deviation direction and deviation amount) therebetween according to a deviation (defocus amount) from the focused state.
- the body control unit 14 calculates the focusing condition (defocus amount) of the focus position by the interchangeable lens 2 on the basis of the positional relationship between the signal waveform (signal sequence a 1 , a 2 , a 3 , a 4 . . . ) by the light fluxes A and the signal waveform (signal sequence b 1 , b 2 , b 3 , b 4 , . . . ) by the light fluxes B and transmits the result of calculation that serves as camera information to the lens control unit 5 .
- the lens control unit 5 moves the focusing lens 7 back and forth along the optical axis direction on the basis of the camera information, the focus is adjusted so that a sharp image can be formed on the image sensor 12 .
- the pixel row 120 is constituted by an Mg pixel (Mg-a) that is included in each second set P 2 and located at the left upper position of the each microlens 40 and an Mg pixel (Mg-b) that is included in each fourth set P 4 and located at the left lower position of each microlens 40 . As shown in FIG.
- the light flux A that passes through the first region 81 on the exit pupil 80 and a light flux C that passes through a third region 83 on the exit pupil 80 enter the pixels that constitute the pixel row 120 .
- the light flux A enters the Mg pixel (Mg-a) located at the left upper position of each microlens 40 .
- the light flux C enters the Mg pixels (Mg-c) located at the left lower position of each microlens 40 .
- the image sensor 12 Upon focusing, the image sensor 12 is in a state in which a sharp image is formed therein, so that in the pixel row 120 as described above, a signal waveform (signal sequence a 1 , a 2 , a 3 , a 4 , . . . ) obtained from the Mg pixels (Mg-a) that receive the light fluxes A and a signal waveform (signal sequence c 1 , c 2 , c 3 , c 4 , . . . ) obtained from the Mg pixels (Mg-c) that receive the light fluxes C overlap in their shape.
- a signal waveform signal sequence a 1 , a 2 , a 3 , a 4 , . . .
- the signal waveform (signal sequence a 1 , a 2 , a 3 , a 4 . . . ) provided by the light fluxes A and the signal waveform (signal sequence c 1 , c 2 , c 3 , c 4 , . . . ) provided by the light fluxes C have positional relationships (deviation direction and deviation amount) therebetween which are different from each other according to a deviation (defocus amount) from the focused state.
- the body control unit 14 calculates the focusing condition (defocus amount) at the focus position achieved by the interchangeable lens 2 on the basis of the positional relationship between the signal waveform (signal sequence a 1 , a 2 , a 3 , a 4 . . . ) provided by the light fluxes A and the signal waveform (signal sequence c 1 , c 2 , c 3 , c 4 , . . . ) provided by the light fluxes C and transmits the result of calculation that serves as camera information to the lens control unit 5 .
- the lens control unit 5 moves the focusing lens 7 back and forth along the optical axis direction on the basis of the camera information, the focus is adjusted so that a sharp image can be formed on the image sensor 12 .
- the pixel row 150 is constituted by an Mg pixel (Mg-a) included in each second set P 2 and located at the left upper position of each microlens 40 , an Mg pixel (Mg-d) included in each second set P 2 and located at the right lower position of each microlens 40 , an Mg pixel (Mg-a) included in each third set P 3 and located at the left upper position of each microlens 40 , and an Mg pixel (Mg-d) included in each third set P 3 and located at the right lower position of each microlens 40 . As shown in FIG.
- the light flux A that passes s 15 through the first region 81 on the exit pupil 80 and a light flux D that passes through a fourth region 84 on the exit pupil 80 enter the pixels that constitute the pixel row 150 .
- the light flux A enters the Mg pixel (Mg-a) located at the left upper position of each microlens 40 .
- the light flux D enters the Mg pixel (Mg-d) located at the right lower part of each microlens 40 .
- the image sensor 12 Upon focusing, the image sensor 12 is in a state in which a sharp image can be formed therein, so that in the pixel row 150 as described above, a signal waveform (signal sequence a 1 , a 2 , a 3 , a 4 , . . . ) obtained from the Mg pixels (Mg-a) that receive the light fluxes A and a signal waveform (signal sequence c 1 , c 2 , c 3 , c 4 , . . . ) obtained from the Mg pixels (Mg-c) that receive the light fluxes C overlap in their shape.
- a signal waveform signal sequence a 1 , a 2 , a 3 , a 4 , . . .
- the signal waveform (signal sequence a 1 , a 2 , a 3 , a 4 . . . ) provided by the light fluxes A and the signal waveform (signal sequence d 1 , d 2 , d 3 , d 4 , . . . ) provided by the light fluxes D have positional relationships (deviation direction and deviation amount) therebetween, which are different from each other, according to the deviation (defocus amount) from the focused state.
- the body control unit 14 calculates the focusing condition (defocus amount) of the focus position by the interchangeable lens 2 on the basis of the positional relationship between the signal waveform (signal sequence a 1 , a 2 , a 3 , a 4 . . . ) provided by the light fluxes A and the signal waveform (signal sequence d 1 , d 2 , d 3 , d 4 , . . . ) provided by the light fluxes D and transmits the result of calculation that serves as camera information to the lens control unit 5 .
- the lens control unit 5 moves the focusing lens 7 back and forth along the optical axis direction on the basis of the camera information, the focus is adjusted so that a sharp image can be formed on the image sensor 12 .
- any of the following three methods is used as image signal generation processing for generating a color image signal on the basis of output signals from the lower photoelectric conversion layer 43 .
- the body control unit 14 executes image signal generation processing by a method indicated by an initial setting in advance.
- FIG. 13 presents a diagram illustrating first image signal generation processing.
- the body control unit 14 which executes the first image signal generation processing, treats four pixels that receive light fluxes through the one and the same microlens 40 as one set 200 , as shown in FIG. 13( a ) .
- Each of the sets 200 includes two G pixels, one B pixel and one R pixel.
- the body control unit 14 treats, for any particular one of the sets 200 , an output signal from the R pixel as an R image signal of the particular one set 200 , an output signal from the B pixel as a B image signal of the particular one set 200 , and an average value of output signals from the two G pixels as a G image signal of the particular one set 200 .
- the body control unit 14 can obtain color image signals (RGB) in a number that is 1 ⁇ 4 times the number of the pixels included in the lower photoelectric conversion layer 43 of the image sensor 12 , as shown in FIG. 13( b ) .
- the body control unit 14 generates an image file for recording by using the thus-obtained color image signals.
- color image signals can be obtained without executing color interpolation processing for interpolating color signals.
- FIG. 14 presents a diagram illustrating second image signal generation processing.
- the body control unit 14 which executes the second image signal generation processing, treats adjacent four pixels in a two-by-two matrix having the same color as one set 210 as shown in FIG. 14( a ) .
- the body control unit 14 treats a signal obtained by adding output signals from the four pixels included in any particular one of the sets 120 as an image signal of the particular set 210 . Specifically, in the case of any particular one set 210 that is all constituted by R pixels, the body control unit 14 treats a signal obtained by adding output signals from the four R pixels as an R image signal of the particular one set 210 . In the case of any particular one set 210 that is all constituted by G pixels, the body control unit 14 treats a signal obtained by adding output signals from the four G pixels as a G image signal of the particular one set 210 .
- the body control unit 14 treats a signal obtained by adding output signals from the four B pixels as a B image signal of the particular one set 210 .
- the body control unit 14 can obtain color image signals of a Bayer arrangement in a number that is 1 ⁇ 4 times the number of the pixels included in the lower photoelectric conversion layer 43 of the image sensor 12 , as shown in FIG. 14( b ) .
- each microlens 40 receive uneven amounts of light. For instance, at a certain incident angle ⁇ 1 , the amount of light received by the pixel located at the left upper position of each microlens 40 is relatively large while the amount of light received by the pixel located at the right lower position of each microlens 40 is relatively small. At another incident angle ⁇ 2 , the amount of light received by the pixel located at the left upper position of each microlens 40 is relatively small while the amount of light received by the pixel located at the right lower position of each microlens 40 is relatively large.
- an optimal image signal can be generated independently of the incident angles of the light fluxes that enter the microlens 40 .
- the body control unit 14 generates, in an image signal of a Bayer arrangement in any particular one set 210 , a color component that is in short by interpolation processing using signals from a plurality of other sets 210 that are adjacent to the particular one set 210 .
- a color component that is in short by interpolation processing using signals from a plurality of other sets 210 that are adjacent to the particular one set 210 .
- color interpolation processing is executed by using signals from a plurality of other sets 210 that are circumjacent to the particular one set 210 . Since such color interpolation processing in the Bayer arrangement is known in the art, detailed description thereof is omitted herein.
- the body control unit 14 generates a file for recording by using color image signals (RGB) obtained by executing this color interpolation processing.
- the body control unit 14 which executes third signal generation processing, first executes color interpolation processing for interpolating a color component that is in short in each pixel.
- FIG. 15 presents a diagram illustrating processing for interpolating a G image signal.
- the body control unit 14 generates, at a position of any particular one pixel out of the R pixels and the B pixels, a G image signal by using output signals from four G pixels located near the particular one pixel by interpolation processing. For instance, in case that a G image signal is to be interpolated at the position of the R pixel in a thick-frame in FIG. 15( a ) , output signals from four G pixels (G 1 through G 4 ) that are located near the particular R pixel are used.
- the body control unit 14 defines ( ⁇ G1+ ⁇ G2+ ⁇ G3+ ⁇ G4)/4 as a G image signal of the particular R pixel.
- the body control unit 14 executes processing for interpolating G image signals at positions of R pixels and B pixels, so that a G image signal can be obtained at a position of each pixel 30 as shown in FIG. 15( b ) .
- FIG. 16 presents a diagram illustrating processing for interpolating R image signals.
- the body control unit 14 treats four pixels constituted by adjacent four pixels in a two-by-two matrix having the same color as one set 220 .
- the body control unit 14 defines a signal obtained by adding output signals from the four pixels in a particular one set 220 of R pixels to be an R image signal of the particular one set 220 .
- the body control unit 14 interpolates R image signals in a particular one set 220 of G pixels and a particular one set 220 of B pixels using R image signals from a plurality of sets of R pixels that are circumjacent to the particular one set 220 of G pixels and the particular one set 220 of B pixels. Since the sets 220 each form a Bayer arrangement as shown in FIG. 16( b ) , the body control unit 14 can execute this interpolation processing by using color interpolation processing for Bayer arrangements that is known in the art.
- the body control unit 14 defines an R image signal that is interpolated in a particular one set 220 of B pixels and divided by 4 (R/4) to be an R image signal for all the four G pixels that constitute the particular one set 220 of G pixels. Similarly, the body control unit 14 defines an R image signal that is interpolated in a particular one set 220 of B pixels and divided by 4 (R/4) to be an R image signal for all the four B pixels that constitute the particular one set 220 of B pixels. In this manner, the body control unit 14 can obtain an R image signal at the position of each of the pixels 30 by executing interpolation processing for interpolating R image signals at the positions of the G pixels and of the B pixels, as shown in FIG. 16( c ) .
- interpolation processing for interpolating B image signals is similar to the interpolation processing for interpolating R image signals, so that detailed description thereof is omitted here.
- the body control unit 14 can obtain a B image signal at the position of each of the pixels 30 by executing interpolation processing for interpolating B image signals at the positions of the R pixels and of the G pixels.
- the body control unit 14 executes the color interpolation processing as described above to obtain an image signal of RGB at the position of each of the pixels 30 as shown in FIG. 17( a ) .
- the body control unit 14 obtains a luminance (brightness) signal Y at the position of each of the pixels 30 by using the image signal of RGB at the position of each of the pixels 30 .
- the body control unit 14 defines 0.299R+0.587G+0.114B as a luminance signal Y.
- the body control unit 14 defines a signal (R-Y) obtained by deducing the luminance signal Y from the R image signal at the position of each of the pixels 30 as a color difference (chrominance) signal Cr.
- the body control unit 14 defines a signal (B-Y) obtained by deducing the luminance signal Y from the B image signal at the position of each of the pixels 30 as a color difference (chrominance) signal Cb.
- the body control unit 14 can obtain the luminance signal Y and the color difference signals Cr and Cb at the position of each of the pixels 30 as shown in FIG. 17( b ) .
- the body control unit 14 uses the color image signals (YCrCb) thus obtained, the body control unit 14 generates an image file for recording having a resolution higher than those generated according to the first image signal generation processing and the second image signal generation processing, respectively.
- FIG. 18 presents a flowchart illustrating the flow of imaging processing to be executed by the body control unit 14 .
- the body control unit 14 when a main switch (not shown) that constitutes the operating member 18 is turned ON, starts up a program that executes the processing exemplified in FIG. 18 .
- step S 11 in FIG. 18 the body control unit 14 controls the image sensor 12 to start photoelectric conversion at a predetermined frame rate.
- the body control unit 14 makes a decision as to whether or not a command for shooting is issued while it controls the liquid crystal display element 16 to successively reproduce and display a through image based on the image signals from the lower photoelectric conversion layer 43 .
- the through image is an image for monitoring that is obtained before shooting is commanded.
- step S 18 the body control unit 14 makes a decision as to whether or not time is up.
- the body control unit 14 measures a predetermined time (for instance, 5 seconds)
- it makes a positive decision in step S 18 and the processing in FIG. 15 is terminated.
- the body control unit 14 makes a negative decision in step S 18 and the program returns to step S 11 .
- step S 12 the body control unit 14 executes AE processing and AF processing.
- AE processing exposure is calculated on the basis of the level of image signal for the through image in order to decide aperture value AV and shutter speed TV so that optimal exposure can be obtained.
- AF processing the focus detection processing described above is executed based on an output signal sequence from the pixel row included in the focus detection area that is set in the upper photoelectric conversion layer 41 .
- step S 13 the body control unit 14 executes shooting processing and the program proceeds to step S 14 .
- the body control unit 14 controls the diaphragm 6 based on the AV and controls the image sensor 12 to perform photoelectric conversion for recording for a storage time on the basis of the TV.
- the body control unit 14 executes the image signal generation processing by using output signals from the lower photoelectric conversion layer 43 and the obtained image signal is subjected to predetermined image processing (gradation processing, contour enhancement, white balance adjustment processing and so on).
- predetermined image processing grade processing, contour enhancement, white balance adjustment processing and so on.
- step S 15 the body control unit 14 controls the liquid crystal display element 16 to display the captured image thereon and the program proceeds to step S 16 .
- step S 16 the body control unit 14 generates an image file for recording and the program proceeds to step S 17 .
- step S 17 the body control unit 14 records the generated image file in a memory card 20 and terminates the processing in FIG. 18 .
- the digital camera system 1 is configured as follows. That is, the digital camera system 1 includes the image sensor 12 that captures an image of a subject with a light flux from a subject that passes through the interchangeable lens 2 ; the body control unit 14 that generates an image signal on the basis of output signals from the image sensor 12 ; and the body control unit 14 detecting a focusing condition of the interchangeable lens 2 according to a phase detection method.
- the image sensor 12 has a pixel group in the upper photoelectric conversion layer 41 and a pixel group in the lower photoelectric conversion layer 43 that receives light fluxes from the subject that have passed through each pixel in the upper photoelectric conversion layer 41 and a microlens group arranged such that the light fluxes from the subject are guided to the pixel group in the upper photoelectric conversion layer 41 ;
- the pixel group in the upper photoelectric conversion layer 41 includes Cy pixels, Ye pixels, and Mg pixels having first, second, and third spectral sensitivities, respectively, differing from each other, arranged in a two-dimensional pattern such that behind each microlens 40 , there are arranged one Cy pixel, one Ye pixel and two Mg pixels in a two-by-two matrix, with these four pixels receiving four light fluxes A through D, respectively, that pass through four pupil regions 81 through 84 , respectively, of the exit pupil 80 .
- the pixel group in the lower photoelectric conversion layer 43 includes R pixels, B pixels, and G pixels having fourth, fifth and sixth spectral sensitivities, respectively, which are complementary to the first, second and third spectral sensitivities, respectively, of the pixel group in the upper photoelectric conversion layer 41 , arranged in a two dimensional pattern, with the positions of the Cy pixel, the Ye pixel and the Mg pixel in the upper photoelectric conversion layer 41 and the positions of the R pixel, the B pixel and the G pixel in the lower photoelectric conversion layer 43 are set such that the R pixel, the B pixel and the G pixel receive light fluxes that have passed through the Cy pixel, the Ye pixel and the Mg pixel, respectively.
- the body control unit 14 generates image signals on the basis of output signals from one of the pixel group in the upper photoelectric conversion layer 41 and the pixel group in the lower photoelectric conversion layer 43 whereas the body control unit 14 detects the focusing condition on the basis of output signals from the other of the pixel group in the upper photoelectric conversion layer 41 and the pixel group in the lower photoelectric conversion layer 43 .
- the generation of image signals and the focus detection according to the phase detection method on the basis of output signals from the image sensor 12 can be executed without providing the image sensor 12 with any pixels that are dedicated for focus detection.
- the digital camera system 1 in (1) above it is further configured such that among the Cy pixels, the Ye pixels and the Mg pixels in the pixel group of the upper photoelectric conversion layer 41 , those pixels having substantially the same spectral sensitivities (in other words, pixels having the same color) are adjacently arranged so as to form two-by-two matrices and four other pixels that are adjacent to each of the two-by-two matrices are arranged behind four different microlenses 40 , respectively, such that their positions with respect to the corresponding microlenses 40 are different from each other.
- the incident light fluxes can be optimally photoelectrically converted regardless of the angle of incidence of light to the microlens 40 .
- the pixel groups of the upper photoelectric conversion layer 41 and the lower photoelectric conversion layer 43 each are formed by arranging, in a two-dimensional pattern, a plurality of sets of pixels, each of which sets includes four pixels arrayed in a two-by-two matrix behind one microlens 40 , the sets include first through fourth sets P 1 through P 4 in which arrangements of pixels are different from each other.
- the pixel group in the upper photoelectric conversion layer 41 is configured such that in the first set P 1 , a Cy pixel and an Mg pixel are arrayed adjacent to each other in the horizontal direction and an Mg pixel and a Ye pixel are arrayed adjacent to the horizontally arrayed Cy and Mg pixels, respectively, in the vertical direction.
- an Mg pixel and a Cy pixel are arrayed adjacent to each other in the horizontal direction and a Ye pixel and an Mg pixel are arrayed adjacent to the horizontally arrayed Mg and Cy pixels, respectively, in the vertical direction.
- an Mg pixel and a Ye pixel are arrayed adjacent to each other in the horizontal direction and a Cy pixel and an Mg pixel are arrayed adjacent to the horizontally arrayed Mg and Ye pixels, respectively, in the vertical direction.
- a Ye pixel and an Mg pixel are arrayed adjacent to each other in the horizontal direction and an Mg pixel and a Cy pixel are arrayed adjacent to the horizontally arrayed Mg and Ye pixels, respectively, in the vertical direction.
- the first set P 1 and the second set P 2 are adjacent to each other in the horizontal direction and alternately arrayed in a repeated manner in the horizontal direction.
- the third set P 3 and the fourth set P 4 are adjacent to each other in the horizontal direction and alternately arrayed in a repeated manner in the horizontal direction.
- a first line L 1 formed by the first set P 1 and the second set P 2 and a second line L 2 formed by the third set P 3 and the fourth set P 4 are adjacent to each other in the vertical direction and alternately arrayed in a repeated manner in the vertical direction.
- a configuration is adopted, in which the body control unit 14 adds output signals from four adjacent R pixels in a two-by-two matrix, adds output signals from four adjacent B pixels in a two-by-two matrix, and adds output signals from four adjacent G pixels in a two-by-two matrix, thereby forming an image signal of a Bayer arrangement (that is, executing the second image signal generation processing).
- the body control unit 14 adds output signals from four adjacent R pixels in a two-by-two matrix, adds output signals from four adjacent B pixels in a two-by-two matrix, and adds output signals from four adjacent G pixels in a two-by-two matrix, thereby forming an image signal of a Bayer arrangement (that is, executing the second image signal generation processing).
- optimal image signals can be generated regardless of the incident angle of light into the microlenses 40 .
- a conventional image processing engine that executes color interpolation of a Bayer arrangement can be used in the color interpolation processing.
- a configuration is adopted, in which the body control unit 14 obtains three color signals at a position of each of the microlenses on the basis of output signals from the R, B and G pixels positioned behind each of the microlenses 40 (that is, executes the first image signal generation processing).
- the body control unit 14 obtains three color signals at a position of each of the microlenses on the basis of output signals from the R, B and G pixels positioned behind each of the microlenses 40 (that is, executes the first image signal generation processing).
- a configuration is adopted, in which the body control unit 14 executes color interpolation processing for generating signals for two other spectral components at each position of R, B and G pixels to obtain three color signals and generates a luminance signal and color difference signals on the basis of the three color signals (that is, executes the third image signal generation processing).
- the body control unit 14 executes color interpolation processing for generating signals for two other spectral components at each position of R, B and G pixels to obtain three color signals and generates a luminance signal and color difference signals on the basis of the three color signals (that is, executes the third image signal generation processing).
- a configuration is adopted, in which the body control unit 14 detects the focusing condition of the interchangeable lens 2 on the basis of outputs signals from a pair of pixels out of the pixels in the pixel group of the upper photoelectric conversion layer 41 , the pair of pixels having substantially the same spectral sensitivities and having different positions from each other with respect to the microlens 40 .
- the focusing condition can be detected according to the phase detection method appropriately based on the output signals from the image sensor 12 .
- the body control unit 14 detects the focusing condition of the interchangeable lens 2 in the horizontal direction on the basis of output signals from Mg pixels included in the first set P 1 and the second set P 2 , respectively, among the pixels in the pixel groups in the upper photoelectric conversion layer 41 .
- the focusing condition can be detected in the horizontal direction of the image sensor 12 according to the phase detection method appropriately based on the output signals from the image sensor 12 .
- the body control unit 14 detects the focusing condition of the interchangeable lens 2 in a direction oblique to the horizontal direction on the basis of output signals from the Mg pixels included in the second set P 2 and the third set P 3 , respectively, among the pixels in the pixel group of the upper photoelectric conversion layer 41 .
- the focusing condition can be appropriately detected according to the phase detection method.
- the focus detection processing is executed by using output signals from the Mg pixels in the upper photoelectric conversion layer 41 .
- the focus detection processing may also be executed by using output signals from the Cy pixels or the Ye pixels.
- the body control unit 14 according to Variation Example 1 is configured to obtain evaluation values by using output signals from the upper photoelectric conversion layer 41 .
- the evaluation values are cumulative values of output signals for each of; for instance, Cy, Mg or Ye pixels.
- the body control unit 14 according to Variation Example 1 executes, when the cumulative value for the Mg pixels is equal to or lower than a predetermined threshold value, the above-mentioned focus detection processing by using either one of Cy pixel or Ye pixel that has a greater cumulative value than that of the rest. With this configuration, appropriate focus detection processing can be executed even when a subject having a small amount of Mg components is shot.
- the processing that is indicated by initial setting is used to generate image signals for recording.
- the present invention is not limited thereto.
- the body control unit 14 according to Variation Example 2 when a through image is to be displayed, selects the first image signal generation processing in which image signals can be generated without executing color interpolation processing and generates image signals by using the selected first image signal generation processing.
- the third image signal generation processing capable of generating image signals having high resolutions is selected and image signals are generated by using the selected third image signal generation processing.
- the body control unit 14 according to Variation Example 2 is configured to select, upon image signal generation, any of the first, second and third image signal generation processing. With this configuration, image signal generation processing that is suitable for uses of images to be generated can be selected.
- the first image signal generation processing which does not require any color interpolation processing is selected in a scene where it is desired to display images on a real-time basis, whereas the third image signal generation processing is selected in a scene where it is desired to record images with high image quality.
- the body control unit 14 may be configured to generate image signals by the first or second image signal generation processing for video images or by the third image signal generation processing for still images.
- the body control unit 14 may be configured to generate image signals by using, for instance, both of the first image signal generation processing and the second image signal generation processing.
- the body control unit 14 controls, for instance, both the image generated by the first image signal generation processing and the image generated by the second image signal generation processing to be displayed on a display device (not shown) on the rear side.
- the body control unit 14 records one of the two displayed images, which one is selected by the user through the operating member 18 into the memory card 20 .
- the configuration is adopted in which the defocus amount in the horizontal direction is obtained on the basis of output signal from the pixel row 90 constituted by the Mg pixel (Mg-b) included in each of the first sets P 1 and the Mg pixel (Mg-a) included in each of the second sets P 2 among the pixels in the upper photoelectric conversion layer 41 .
- the present invention is not limited thereto.
- a configuration may be adopted, in which the defocus amount in the horizontal direction is obtained on the basis of output signals from a pixel row constituted by the Mg pixel (Mg-d) included in each of the third sets P 3 and the Mg pixel (Mg-c) included in each of the fourth sets P 4 .
- the defocus amount in the horizontal direction may be obtained on the basis of both the above pixel row and the pixel row 90 .
- the configuration is adopted in which the defocus amount in the vertical direction is obtained on the basis of output signals from the pixel row 120 constituted by the Mg pixel (Mg-a) included in each of the second sets P 2 and the Mg pixel (Mg-c) included in each of the fourth sets P 4 .
- the present invention is not limited thereto.
- a configuration may be adopted in which the defocus amount in the vertical direction is obtained on the basis of output signals from a pixel row constituted by the Mg pixel (Mg-b) included in each of the first sets P 1 and the Mg pixel (Mg-d) included in each of the third sets P 3 .
- the defocus amount in the horizontal direction may be obtained on the basis of both the above pixel row and the pixel row 120 .
- the configuration is adopted, in which the defocus amount in an oblique direction is obtained on the basis of output signals from the pixel row 150 constituted by the Mg pixels (Mg-a) and (Mg-d) included in each of the second set P 2 and the Mg pixels (Mg-a) and (Mg-d) included in each of the third set P 3 .
- the present invention is not limited thereto.
- a configuration may be adopted, in which the defocus amount in the oblique direction is obtained on the basis of the pixel row constituted by the Mg pixels (Mg-b) and (Mg-c) included in each of the first set P 1 and the Mg pixels (Mg-b) and (Mg-c) included in each of the fourth set P 4 .
- the defocus amount in the oblique direction may be obtained on the basis of both the above pixel row and the pixel row 150 .
- a configuration is adopted, in which Mg pixels, Cy pixels and Ye pixels are provided in the upper photoelectric conversion layer 41 and G pixels, R pixels and B pixels are provided in the lower photoelectric conversion layer 43 .
- a configuration may be adopted, in which G pixels, R pixels and B pixels are provided in the upper photoelectric conversion layer and Mg pixels, Cy pixels and Ye pixels are provided in the lower photoelectric conversion layer
- the present invention is applied to the digital camera system 1 having a configuration such that the interchangeable lens 2 is mounted to the camera body 3 .
- the present invention is not limited thereto.
- the present invention may also be applied to a lens-integrated digital camera.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- Optics & Photonics (AREA)
- General Physics & Mathematics (AREA)
- Spectroscopy & Molecular Physics (AREA)
- Automatic Focus Adjustment (AREA)
- Transforming Light Signals Into Electric Signals (AREA)
- Focusing (AREA)
- Color Television Image Signal Generators (AREA)
- Solid State Image Pick-Up Elements (AREA)
- Studio Devices (AREA)
Abstract
Description
- This application is a continuation application of U.S. patent application Ser. No. 15/794,710 filed Oct. 26, 2017, which is a continuation application of U.S. patent application Ser. No. 14/389,626 filed Apr. 8, 2015, which is a national stage application of PCT/JP2013/059621 filed Mar. 29, 2013, which in turn claims priority to Japanese Patent Application No. 2012-081165 filed Mar. 30, 2012. The disclosures of these applications are incorporated herein by reference in their entirety.
- The present invention relates to an image-capturing device and an image sensor.
- An image-capturing device is known which performs focus detection by a split-pupil phase detection method on the basis of output signals from a plurality of pixels dedicated for focus detection arranged on a part of an image sensor (see Patent Literature 1).
- Patent Literature 1: Japanese Patent Application Laid-open No. 2007-282109.
- To perform focus detection by the conventional art, the focus detection is possible only at positions where pixels dedicated for focus detection are arranged. However, an increased number of pixels dedicated for focus detection results in a decrease in image quality since no image signal is obtained from the positions where the pixels dedicated for focus detection are arranged. As described above, according to the conventional technology, it is made possible to perform focus detection according to a phase detection method as well as generation of image signals based upon output signals from image sensor. On the other hand, there occurs an adverse effect due to provision of the pixels dedicated for focus detection at a part of the image sensor.
- According to the 1st aspect of the present invention, an image-capturing device comprises: an image sensor that captures an image of a subject with light fluxes from the subject that have passed through an imaging optical system; an image generation unit that generates an image signal based upon an output signal from the image sensor; and a focus detection unit that detects a focusing condition of the imaging optical system according to a phase detection method based upon an output signal from the image sensor. The image sensor includes an upper layer pixel group and a lower layer pixel group that receives the light fluxes from the subject that have passed through each pixel in the upper layer pixel group, and a microlens group arranged so as to guide the light fluxes from the subject to the upper layer pixel group. The upper layer pixel group includes first, second and third pixels having first, second and third spectral sensitivities, respectively, differing from each other, and being arranged in a two-dimensional pattern, with one of the first pixels, one of the second pixels and two of the third pixels being arranged in a two-by-two matrix behind each of microlenses in the microlens group, and the four pixels receive four light fluxes, respectively, that pass through four pupil areas, respectively, of an exit pupil of the imaging optical system. The lower layer pixel group includes fourth, fifth and sixth pixels having fourth, fifth and sixth spectral sensitivities, respectively, that are complementary to the first, second, and third spectral sensitivities, respectively, of the upper layer pixel group, being arranged in a two-dimensional pattern. Positions of the first, second and third pixels in the upper layer pixel group and positions of the fourth, fifth and sixth pixels in the lower layer pixel group are determined such that the fourth, fifth and sixth pixels receive light fluxes that pass through the first, second and third pixels, respectively. The image generation unit generates the image signal based upon an output signal from one of the upper layer pixel group and the lower layer pixel group. The focus detection unit detects the focusing condition based upon an output signal from other of the upper layer pixel group and the lower layer pixel group.
- According to the 2nd aspect of the present invention, in the image-capturing device according to the 1st aspect, it is preferred that, in the upper layer pixel group, the first, second, and third pixels are arranged such that pixels having substantially same spectral sensitivities are adjacently arranged in a two-by-two matrix, and four pixels adjacent to the two-by-two matrix are arranged behind four different microlenses in the microlens group, respectively, and at different positions with respect to the microlenses.
- According to the 3rd aspect of the present invention, in the image-capturing device according to the 2nd aspect, it is preferred that, in the upper layer pixel group, the first, second and third pixels provide output signals relating to cyan, yellow, and magenta, respectively, and, in the lower layer pixel group, the fourth, fifth and sixth pixels provide output signals relating to colors complementary to the cyan, yellow and magenta, respectively.
- According to the 4th aspect of the present invention, in the image-capturing device according to the 3rd aspect, it is preferred that, the upper layer pixel group and the lower layer pixel group each include an array of a plurality of sets of pixels arranged in a two-dimensional pattern, each of the plurality of sets of pixels having four pixels arranged in a two-by-two matrix behind any particular one of the microlenses and the sets include first through fourth sets having different arrangements of pixels, and the upper layer pixel group is configured such that in the first set, the first pixel and the third pixel are adjacently arranged in a predetermined array direction and the third pixel and the second pixel are arranged adjacent to the first pixel and the third pixel, respectively, in a direction perpendicular to the predetermined array direction, in the second set, the third pixel and the first pixel are adjacently arranged in the predetermined array direction and the second pixel and the third pixel are arranged adjacent to the third pixel and the first pixel, respectively, in the direction perpendicular to the predetermined array direction, in the third set, the third pixel and the second pixel are adjacently arranged in the predetermined array direction and the first pixel and the third pixel are arranged adjacent to the third pixel and the second pixel, respectively, in the direction perpendicular to the predetermined array direction, in the fourth set, the second pixel and the third pixel are adjacently arranged in the predetermined array direction and the third pixel and the first pixel are arranged adjacent to the second pixel and the third pixel, respectively, in the direction perpendicular to the predetermined array direction, and the first set and the second set are adjacent to each other in the predetermined array direction and alternately arranged in a repeated manner in the predetermined array direction, the third set and the fourth set are adjacent to each other in the predetermined array direction and alternately arranged in a repeated manner in the predetermined array direction, and a first row formed by the first set and the second set and a second row formed by the third set and the fourth set are adjacent to each other in the direction perpendicular to the predetermined array direction and alternately arranged in a repeated manner in the direction perpendicular to the predetermined array direction.
- According to the 5th aspect of the present invention, in the image-capturing device according to any one of the 2nd to 4th aspects, it is preferred that the image generation unit adds output signals from four of the fourth pixels that are adjacent to each other in a form of a two-by-two matrix, adds output signals from four of the fifth pixels that are adjacent to each other in a form of a two-by-two matrix, and adds output signals from four of the sixth pixels that are adjacent to each other in a form of a two-by-two matrix to generate an image signal of a Bayer arrangement.
- According to the 6th aspect of the present invention, in the image-capturing device according to any one of the 1st to 4th aspects, it is preferred that the image generation unit obtains three color signals at a position corresponding to each microlens based upon output signals from the fourth, fifth and sixth pixels positioned behind each microlens.
- According to the 7th aspect of the present invention, in the image-capturing device according to any one of the 1st to 4th aspects, it is preferred that the image generation unit executes, at respective positions of the fourth through sixth pixels, color interpolation processing for generating signals of other two spectral components to obtain three color signals and generates a luminance signal and color difference signals based upon the three color signals thus obtained.
- According to the 8th aspect of the present invention, in the image-capturing device according to any one of the 1st to 7th aspects, it is preferred that the focus detection unit detects the focusing condition of the imaging optical system based upon output signals from a pair of pixels having substantially the same spectral sensitivities and located at positions differing from each other with respect to the microlens, out of the upper layer pixel group or the lower layer pixel group.
- According to the 9th aspect of the present invention, in the image-capturing device according to the 4th aspect, it is preferred that the focus detection unit detects the focusing condition of the imaging optical system in the predetermined array direction based upon output signals from at least one plurality of the third pixels of a plurality of the third pixels included in the first set and the second set, respectively, and a plurality of the third pixels included in the third set and the fourth set, respectively, in the upper layer pixel group.
- According to the 10th aspect of the present invention, in the image-capturing device according to the 4th aspect, it is preferred that the focus detection unit detects the focusing condition of the imaging optical system in the direction perpendicular to the predetermined array direction based upon output signals from at least one plurality of the third pixels of a plurality of the third pixels included in the first set and the third set, respectively, and a plurality of the third pixels included in the second set and the fourth set, respectively, in the upper layer pixel group.
- According to the 11th aspect of the present invention, in the image-capturing device according to the 4th aspect, it is preferred that the focus detection unit detects the focusing condition of the imaging optical system in a direction oblique to the predetermined array direction based upon output signals from at least one plurality of the third pixels of a plurality of the third pixels included in the first set and the fourth set, respectively, and a plurality of the third pixels included in the second set and the third set, respectively, in the upper layer pixel group.
- According to the 12th aspect of the present invention, an image sensor comprises: a first image-capturing unit that includes a plurality of microlenses arranged in a two-dimensional pattern; and a plurality of light reception units that are provided in correspondence to each of the microlenses, respectively, and that receive light having a predetermined wavelength and transmit lights having wavelengths different from the predetermined wavelength; and a second image-capturing unit that receives the lights that are transmitted through the first image-capturing units.
- According to the 13th aspect of the present invention, in the image sensor according to the 12th aspect, it is preferred that the first imaging unit includes light reception units that are arranged adjacent to each other so as to be provided in correspondence to adjacent two of the microlenses and that absorb light having same wavelength corresponding to the predetermined wavelength.
- According to the present invention, image signal generation and focus detection according to a phase detection method can be performed based upon output signals from the image sensor without providing the image sensor with pixels dedicated for focus 3 o detection.
-
FIG. 1 A diagram illustrating by an example the configuration of a digital camera system according to an embodiment of the present invention. -
FIG. 2 A plan view illustrating by an example the layout of pixels in an upper photoelectric conversion layer. -
FIG. 3 A plan view illustrating by an example the layout of pixels in a lower photoelectric conversion layer. -
FIG. 4 A diagram illustrating by an example a cross-section of an image sensor. -
FIG. 5 A diagram illustrating by an example a circuit configuration per pixel in the image sensor. -
FIG. 6 A diagram illustrating by an example an exit pupil of an interchangeable lens. -
FIG. 7 A diagram illustrating pixel rows for obtaining a defocus amount. -
FIG. 8 A diagram illustrating light fluxes passing through the exit pupil. -
FIG. 9 A diagram illustrating pixel rows used for obtaining a defocus amount. -
FIG. 10 A diagram illustrating light fluxes passing through the exit pupil. -
FIG. 11 A diagram illustrating pixel rows used for obtaining a defocus amount. -
FIG. 12 A diagram illustrating light fluxes passing through the exit pupil. -
FIG. 13 A diagram illustrating first image signal generation processing. -
FIG. 14 A diagram illustrating second image signal generation processing. -
FIG. 15 A diagram illustrating third image signal generation processing. -
FIG. 16 A diagram illustrating third image signal generation processing. -
FIG. 17 A diagram illustrating third image signal generation processing. -
FIG. 18 A flowchart illustrating the flow of imaging processing. - Hereinafter, an embodiment of the present invention is described with reference to the attached drawings.
FIG. 1 presents a diagram illustrating by an example the configuration of a digital camera system according to an embodiment of the present invention. Adigital camera system 1 includes aninterchangeable lens 2 and acamera body 3. Theinterchangeable lens 2 is mounted on thecamera body 3 via a mount unit 4. - The
interchangeable lens 2 includes alens control unit 5, amain lens 9, azoom lens 8, a focusinglens 7, and a diaphragm 6. Thelens control unit 5 includes a microcomputer, a memory and so on and performs drive control of the focusinglens 7 and the diaphragm 6, detection of an open state of the diaphragm 6, detection of positions of thezoom lens 8 and the focusinglens 7, transmission of information about lens to abody control unit 14 on the side of thecamera body 3 described later, reception of information about camera from thebody control unit 14, and so on. - The
camera body 3 includes animage sensor 12, an image sensor drive control unit 19, thebody control unit 14, a liquid crystal displayelement drive circuit 15, a liquidcrystal display element 16, anocular lens 17, anoperating member 18 and so on. Adetachable memory card 20 is attached to thecamera body 3. Theimage sensor 12 is arranged on a predetermined imaging plane of theinterchangeable lens 2 and captures an image of a photographic subject that is formed by theinterchangeable lens 2. - The
body control unit 14 includes a microcomputer, a memory and so on. Thebody control unit 14 controls operations of the digital camera system in whole. Thebody control unit 14 and thelens control unit 5 are configured to communicate with each other via anelectric contact unit 13 of the mount unit 4. - The image sensor drive control unit 19 generates a control signal that is necessary for the
image sensor 12 in response to a command from thebody control unit 14. The liquid crystal displayelement drive circuit 15 drives the liquidcrystal display element 16 that constitutes a liquid crystal view finder (EVF: electric view finder) in response to a command from thebody control unit 14. The photographer observes an image displayed on the liquidcrystal display element 16 through theocular lens 17. Thememory card 20 is a storage medium in which image signals and the like are stored. - The image of the photographic subject formed on the
image sensor 12 by theinterchangeable lens 2 is subjected to photoelectric conversion by theimage sensor 12. Theimage sensor 12 is controlled by a control signal from the image sensor drive control unit 19 with respect to timing (frame rate) of storage of photoelectric conversion signals and of reading of signals. The image signals read out from theimage sensor 12 are converted in an A/D conversion unit (not shown) into digital data, which then is transmitted to thebody control unit 14. - The
body control unit 14 calculates a defocus amount on the basis of image signals from theimage sensor 12 corresponding to predetermined focus detection areas, respectively, and transmits the calculated defocus amount to thelens control unit 5. Thelens control unit 5 calculates a focusing lens drive amount on the basis of the defocus amount received from thebody control unit 14 and drives the focusinglens 7 by a motor or the like, which is not shown, to move the focusinglens 7 to a focusing position on the basis of the calculated lens drive amount. - The
body control unit 14 generates image data to be recorded on the basis of a signal that is output from theimage sensor 12 after shooting is commanded. Thebody control unit 14 stores the generated image data in thememory card 20 and at the same time transmits the generated image data to the liquid crystal display element drive circuit and controls it to be reproduced and displayed on the liquidcrystal display element 16. - It is to be noted that the
camera body 3 is provided with the operatingmember 18 that includes a shutter button and a setting member for setting a focus detection area or areas and so on. Thebody control unit 14 detects an operating signal from the operatingmember 18 and controls the operations (photographing processing, setting of a focus detection area and so on) in response to the result of the detection. - <Description of Image Sensor>
- Since the present embodiment is featured by the
image sensor 12, further description is focused on theimage sensor 12. Theimage sensor 12 has a stacked structure, in which an upper photoelectric conversion layer 41 (FIG. 4 ) and a lower photoelectric conversion layer 43 (FIG. 4 ) are stacked one above another. The upperphotoelectric conversion layer 41 is constituted by a photoconductive film that absorbs (photoelectrically converts) light having wavelength components described later and light having wavelength components not absorbed (photoelectrically converted) by the upperphotoelectric conversion layer 41 is transmitted to the lowerphotoelectric conversion layer 43, where the transmitted light is photoelectrically converted. - (Upper Photoelectric Conversion Layer)
-
FIGS. 2(a) and (b) presents plan views each illustrate by an example the layout of pixels in the upperphotoelectric conversion layer 41 of theimage sensor 12. In this case, 10×10 pixels, which are extracted as representatives, are shown. The extracted pixels are laid out in a substantially square form and arranged in a two-dimensional pattern. The extracted pixels include three types of pixels, i.e., a pixel that photoelectrically converts light of a cyan (Cy) color component (Cy pixel), a pixel that photoelectrically converts light of a magenta (Mg) color component (Mg pixel), and a pixel that photoelectrically converts light of a yellow (Ye) color component (Ye pixel). - The Cy pixel is constituted by a photoelectric conversion unit that absorbs (photoelectrically converts) the light of the cyan color component. The Mg pixel is constituted by a photoelectric conversion unit that absorbs (photoelectrically converts) the light of the magenta color component. The Ye pixel is constituted by a photoelectric conversion unit that absorbs (photoelectrically converts) the light of the yellow color component.
- In addition, the
image sensor 12 is formed of a plurality ofmicrolenses 40 each of which is to efficiently guide light fluxes from theinterchangeable lens 2 to a set of four pixels. InFIG. 2 , 5×5=25 circles correspond to themicrolenses 40. Themicrolenses 40, each constituted by a spherical lens of axial symmetry whose center substantially coincides with its optical axis or a nonspherical lens, are arranged in a two-dimensional pattern, with the light incident side thereof having a convex shape. - Behind each of the
microlenses 40, one Cy pixel, two Mg pixels and one Ye pixel are arranged in a two-by-two matrix. In the description of the present invention herein, a plurality of sets of four pixels positioned behind the correspondingmicrolenses 40, respectively, are classified into four types (P1 through P4) according to differences in their arrangements as shown inFIG. 2(a) . - Behind the
microlenses 40, first sets P1 each include a Cy pixel at a left upper position, an Mg pixel at a right upper position, an Mg pixel at a left lower position, and a Ye pixel at a right lower position. Second sets P2 each include an Mg pixel at a left upper position, a Cy pixel at a right upper position, a Ye pixel at a left lower position, and an Mg pixel at a right lower position. Third sets P3 each include an Mg pixel at a left upper position, a Ye pixel at a right upper position, a Cy pixel at a left lower position, and an Mg pixel at a right lower position. Fourth sets P4 each include a Ye pixel at a left upper position, an Mg pixel at a right upper position, an Mg pixel at a left lower position, and a Cy pixel at a right lower position. - The first sets P1 and the second sets P2 are adjacent to each other in a horizontal direction (X direction) and arranged alternately in a repeated manner in the horizontal direction. A line formed by the first sets P1 and the second sets P2 is called a “first line L1”. The third sets P3 and the fourth sets P4 are adjacent to each other in the horizontal direction and are arranged alternately in a repeated manner in the horizontal direction. A line formed by the third sets P3 and the fourth sets P4 is called a “second line L2”.
- The first line L1 and the second line L2 described above are adjacent to each other in a vertical direction (Y direction) and are alternately arranged in a repeated manner in the vertical direction. With this configuration, each of the first sets P1 and each of the third sets P3 are adjacent to each other in the vertical direction, whereas each of the second sets P2 and each of the fourth sets P4 are adjacent to each other in the vertical direction.
- With such an arrangement, the
microlenses 40 and the Cy pixels, the Mg pixels, and the Ye pixels have the following positional relationships. - First, the Cy pixels, behind four
microlenses 40 adjacent to each other in the horizontal direction and in the vertical direction, are arranged at a left upper position, a right upper position, a left lower position, and a right lower position, respectively. The Mg pixels, behind fourmicrolenses 40 adjacent to each other in the horizontal direction and in the vertical direction, are arranged at right upper and left lower positions, left upper and right lower positions, left upper and right lower positions, and right upper and left lower positions, respectively. The Ye pixels, behind fourmicrolenses 40 adjacent to each other in the horizontal direction and in the vertical direction, are arranged at a right lower position, a left lower position, a right upper position, and a left upper position, respectively. In this manner, the Cy pixels, the Mg pixels, and the Ye pixels are uniformly arranged behind themicrolenses 40 without being arranged disproportionately to specific positions. -
FIG. 2(b) is a diagram showing a part extracted from the upper photoelectric conversion layer of the image sensor, which part is similar to that shown inFIG. 2(a) . When four sets of pixels (P1 through P4) shown inFIG. 2(a) are viewed by shifting them by 1 pixel both in the horizontal direction and in the vertical direction, the Cy. Mg and Ye pixels each are arranged such that adjacent four pixels in any particular two-by-two matrix have the same color as shown inFIG. 2(b) . - In addition, the four pixels in any particular two-by-two matrix having the same color are arranged behind
different microlenses 40, respectively, so that they assume different positions with respect to themicrolenses 40. In other words, the Cy, Mg and Ye pixels arranged behind the fourmicrolenses 40, respectively, are arranged such that they are adjacent to each other in a two-by-two matrix for each color. - (Lower Photoelectric Conversion Layer)
-
FIGS. 3(a) and (b) presents plan views each illustrate by an example the layout of pixels in the lower photoelectric conversion layer 43 (FIG. 4 ) of theimage sensor 12. In this case, 10×10 pixels, which correspond to the positions of the pixels exemplified inFIG. 2 , are shown. These pixels are laid out in a substantially square form and arranged in a two-dimensional pattern. They include three types of pixels, i.e., a pixel that photoelectrically converts light of a red (R) color component (R pixel), a pixel that photoelectrically converts light of a green (G) color component (G pixel), and a pixel that photoelectrically converts light of a blue (B) color component (B pixel). - The R pixel is constituted by a photoelectric conversion unit that photoelectrically converts light of the red color component (i.e., a complementary color of Cy) that has not been absorbed (photoelectrically converted) by the Cy pixel lying above. The G pixel is constituted by a photoelectric conversion unit that photoelectrically converts light of the green color component (i.e., a complementary color of Mg) that has not been absorbed (photoelectrically converted) by the Mg pixel lying above. The B pixel is constituted by a photoelectric conversion unit that photoelectrically converts light of the blue color component (i.e., complementary color of Ye) that has not been absorbed (photoelectrically converted) by the Ye pixel lying above. In other words, the Cy, Mg and Ye pixels in the upper
photoelectric conversion layer 43 constitute R, G and B light reception units, respectively, that serve as color filters. - As a result, each of sets (referred to as Q1) underlying the first sets P1, respectively, includes an R pixel at a left upper position, a G pixel at a right upper position, a G pixel at a left lower position, and a B pixel at a right lower position. Each of sets (referred to as Q2) underlying the second sets P2, respectively, includes a G pixel at a left upper position, an R pixel at a right upper position, a B pixel at a left lower position, and a G pixel at a right lower position. Each of sets (referred to as Q3) underlying the third sets P3, respectively, includes a G pixel at a left upper position, a B pixel at a right upper position, an R pixel at a left lower position, and a G pixel at a right lower position. Each of sets (referred to as Q4) underlying the fourth sets P4, respectively, includes a B pixel at a left upper position, a G pixel at a right upper position, a G pixel at a left lower position, and an R pixel at a right lower position.
- With such an arrangement, the
microlenses 40 and the R pixels, the G pixels, and the B pixels have the following positional relationships. - First, the R pixels, behind four
microlenses 40 that are adjacent to each other in the horizontal direction and in the vertical direction, are arranged at a left upper position, a right upper position, a left lower position, and a right lower position, respectively. The G pixels, behind fourmicrolenses 40 that are adjacent to each other in the horizontal direction and in the vertical direction, are arranged at right upper and left lower positions, left upper and right lower positions, left upper and right lower positions, and right upper and left lower positions, respectively. The B pixels, behind fourmicrolenses 40 that are adjacent to each other in the horizontal direction and in the vertical direction, are arranged at a right lower position, a left lower position, a right upper position, and a left upper position, respectively. In this manner, the R pixels, the G pixels, and the B pixels are uniformly arranged behind themicrolenses 40 without being arranged disproportionately to specific positions. -
FIG. 3(b) is a diagram showing a part extracted from the lower photoelectric conversion layer of the image sensor, which part is similar to that shown inFIG. 3(a) . When four sets of pixels (Q1 through Q4) shown inFIG. 3(a) are viewed by shifting them by 1 pixel both in the horizontal direction and in the vertical direction, the R, G and B pixels each are arranged such that adjacent four pixels in any particular two-by-two matrix have the same color as shown inFIG. 3(b) . - In addition, the four pixels in any particular two-by-two matrix having the same color are arranged behind
different microlenses 40, respectively, so that they assume different positions with respect to themicrolenses 40. In other words, the R, G and B pixels arranged behind the fourmicrolenses 40, respectively, are arranged such that they are adjacent to each other in a two-by-two matrix for each color. - The sets constituted by four pixels in a two-by-two matrix of the same color, i.e., a
set 50 r constituted by four R pixels, a set 50 g constituted by four G pixels, and aset 50 b constituted by four B pixels, when the four pixels are viewed as one set, each form a Bayer arrangement. -
FIG. 4 presents a diagram illustrating by an example a cross-section of theimage sensor 12. InFIG. 4 , theimage sensor 12 includes the lowerphotoelectric conversion layer 43 formed on a silicone substrate and the upperphotoelectric conversion layer 41 stacked on the lowerphotoelectric conversion layer 43 via awiring layer 42. Above the upperphotoelectric conversion layer 41 are formed themicrolenses 40. - The upper
photoelectric conversion layer 41 includes electrodes and photoconductive films each sandwiched by the electrodes to constitute the Cy, Mg and Ye pixels. For instance, an upper electrode a and a lower electrode k-Cy sandwich therebetween a photoconductive film P1-Cy to constitute a Cy pixel in the first set P1. In addition, for instance, the upper electrode a and a lower electrode k-Mg sandwich therebetween a photoconductive film P2-Mg to constitute an Mg pixel in the second set P2. - The lower
photoelectric conversion layer 43 is constituted by the R pixel, the G pixel, and the B pixel on a silicone substrate and photoelectrically converts light that enters each of the pixels. InFIG. 4 , the R pixel in the first set Q1 receives complementary light (R) that is transmitted through the Cy pixel in the first set P1 of the upperphotoelectric conversion layer 41. In addition, the G pixel in the second set Q2 receives complementary light (G) that is transmitted through the Mg pixel in the second set P2 of the upperphotoelectric conversion layer 41. -
FIG. 5 illustrates by an example a circuit configuration per pixel in theimage sensor 12. InFIG. 5 , a reference voltage source Vref is provided to reference voltage source terminals t32 a and t32 b. In addition, a voltage source Vcc is provided to the voltage source terminals t31 a and t31 b. Furthermore, a voltage source Vpc is provided from a terminal t33 to a PC (photoconductor) 20. - A signal detection unit of the upper
photoelectric conversion layer 41 has the following configuration. ThePC 20 constitutes a photoelectric conversion unit for one pixel of the upperphotoelectric conversion layer 41. In thePC 20, incident light is photoelectrically converted into a charge, so that charges are accumulated therein. A source follower amplifier MOS transistor Tr6 amplifies a voltage signal based upon the accumulated charges. A transfer switch MOS transistor Tr5 constitutes a switch for selecting a target pixel to be read out. When a control pulsesignal ϕSEL# 1 that turns on/off the transfer switch MOS transistor Tr5 is provided to a terminal t11, an amplified signal is read out from the terminal t11 via the transfer switch MOS transistor Tr5. A reset MOS transistor Tr7 discharges unnecessary charges in response to a reset pulsesignal ϕR# 1 provided to a terminal t13 (in other words, resets to a predetermined potential). - A signal detection unit of the lower
photoelectric conversion layer 43 has the following configuration. A photodiode PD constitutes a photoelectric conversion unit for one pixel of the lowerphotoelectric conversion layer 43. The photodiode PD photoelectrically converts light that is transmitted through thePC 20 to generate charges. The photodiode PD and a floating diffusion (FD) unit are connected with each other through the transfer MOS transistor Tr4. When a control pulsesignal ϕTx# 2 that turns on/off the transfer switch MOS transistor Tr4 is provided to a terminal t24, charges are transferred to the floating diffusion unit via the transfer MOS transistor Tr4. - A source follower amplifier MOS transistor Tr2 amplifies a voltage signal based upon the accumulated charges. A transfer switch MOS transistor Tr1 constitutes a switch for selecting a target pixel to be read out. When a control pulse
signal ϕSEL# 2 that turns on/off the transfer switch MOS transistor Tr1 is provided to a terminal t22, an amplified signal is read out from a terminal t21 via the transfer switch MOS transistor Tr1. A reset MOS transistor Tr3 discharges unnecessary charges in the floating diffusion unit in response to a reset pulse signal+R# 1 provided to a terminal t13 (in other words, resets it to a predetermined potential). - <Focus Detection Processing>
- Next, an example in which signals for focus detection are obtained from the
image sensor 12 having the configuration described above is described referring toFIG. 6 throughFIG. 12 .FIG. 6 presents a diagram illustrating by an example anexit pupil 80 of theinterchangeable lens 2 in a state in which the diaphragm is open. Light fluxes that have passed through fourregions 81 through 84 of theexit pupil 80 enter pixels located at positions corresponding to a left upper part, a right upper part, a left lower part, and a right lower part, respectively, of each of themicrolenses 40 inFIG. 2 . For each of themicrolenses 40, correspondence relationship between the light fluxes that enter the pixels located at positions corresponding to the left upper, right upper, left lower, and right lower parts of the microlens and thefirst region 81, thesecond region 82, thethird region 83, and thefourth region 84, respectively, of theexit pupil 80 may be considered such that the up and down relation as well as the left and right relation are inverted with respect to the light axis Ax of theinterchangeable lens 2 as an axis of symmetry. - First, as exemplified in
FIG. 7 , explanation is made on a case in which the defocus amount is obtained based on apixel row 90, in which Mg pixels out of the pixels in theimage sensor 12 are arranged in the horizontal direction (X axis direction). Thepixel row 90 is constituted by an Mg pixel (Mg-a) that is included in each second set P2 and located at the left upper position of eachmicrolens 40 and an Mg pixel (Mg-b) that is included in each first set P1 and located at the right upper position of eachmicrolens 40. As exemplified inFIG. 8 , a light flux A that pass through afirst region 81 on theexit pupil 80 and a light flux B that pass through asecond region 82 on theexit pupil 80 enter the pixels that constitute thepixel row 90. The light flux A enters the Mg pixel (Mg-a) located at the left upper position of eachmicrolens 40. The light flux B enters the Mg pixel (Mg-b) located at the right upper position of eachmicrolens 40. - Upon focusing, the
image sensor 12 is in a state in which a sharp image is formed, so that as described above, a pair of images formed by light fluxes through different positions of the pupil as a result of pupil-splitting coincide with each other on theimage sensor 12. In other words, in thepixel row 90, a signal waveform (signal sequence a1, a2, a3, a4, . . . ) obtained from the Mg pixels (Mg-a) that receive the light fluxes A and a signal waveform (signal sequence b1, b2, b3, b4, . . . ) obtained from the Mg pixels (Mg-b) that receive the light fluxes B overlap in their shape. - On the other hand, upon non-focusing, i.e., in a state in which a sharp image is formed on the front side or on the rear side of the
image sensor 12, a pair of images formed by the light fluxes subjected to the pupil-splitting do not coincide with each other on theimage sensor 12. In this case, the signal waveform (signal sequence a1, a2, a3, a4 . . . ) by the light fluxes A and the signal waveform (signal sequence b1, b2, b3, b4, . . . ) by the light fluxes B have different positional relationships (deviation direction and deviation amount) therebetween according to a deviation (defocus amount) from the focused state. - The
body control unit 14 calculates the focusing condition (defocus amount) of the focus position by theinterchangeable lens 2 on the basis of the positional relationship between the signal waveform (signal sequence a1, a2, a3, a4 . . . ) by the light fluxes A and the signal waveform (signal sequence b1, b2, b3, b4, . . . ) by the light fluxes B and transmits the result of calculation that serves as camera information to thelens control unit 5. As thelens control unit 5 moves the focusinglens 7 back and forth along the optical axis direction on the basis of the camera information, the focus is adjusted so that a sharp image can be formed on theimage sensor 12. - Next, explanation is made on a case in which the defocus amount is obtained based upon a
pixel row 120 in which Mg pixels out of the pixels in theimage sensor 12 are arranged in the vertical direction (Y axis direction) as exemplified inFIG. 9 . Thepixel row 120 is constituted by an Mg pixel (Mg-a) that is included in each second set P2 and located at the left upper position of the eachmicrolens 40 and an Mg pixel (Mg-b) that is included in each fourth set P4 and located at the left lower position of eachmicrolens 40. As shown inFIG. 10 , the light flux A that passes through thefirst region 81 on theexit pupil 80 and a light flux C that passes through athird region 83 on theexit pupil 80 enter the pixels that constitute thepixel row 120. The light flux A enters the Mg pixel (Mg-a) located at the left upper position of eachmicrolens 40. The light flux C enters the Mg pixels (Mg-c) located at the left lower position of eachmicrolens 40. - Upon focusing, the
image sensor 12 is in a state in which a sharp image is formed therein, so that in thepixel row 120 as described above, a signal waveform (signal sequence a1, a2, a3, a4, . . . ) obtained from the Mg pixels (Mg-a) that receive the light fluxes A and a signal waveform (signal sequence c1, c2, c3, c4, . . . ) obtained from the Mg pixels (Mg-c) that receive the light fluxes C overlap in their shape. - On the other hand, upon non-focusing, the signal waveform (signal sequence a1, a2, a3, a4 . . . ) provided by the light fluxes A and the signal waveform (signal sequence c1, c2, c3, c4, . . . ) provided by the light fluxes C have positional relationships (deviation direction and deviation amount) therebetween which are different from each other according to a deviation (defocus amount) from the focused state.
- The
body control unit 14 calculates the focusing condition (defocus amount) at the focus position achieved by theinterchangeable lens 2 on the basis of the positional relationship between the signal waveform (signal sequence a1, a2, a3, a4 . . . ) provided by the light fluxes A and the signal waveform (signal sequence c1, c2, c3, c4, . . . ) provided by the light fluxes C and transmits the result of calculation that serves as camera information to thelens control unit 5. As thelens control unit 5 moves the focusinglens 7 back and forth along the optical axis direction on the basis of the camera information, the focus is adjusted so that a sharp image can be formed on theimage sensor 12. - In addition, explanation is made on a case in which the defocus amount is obtained on the basis of a
pixel row 150 constituted by Mg pixels out of the pixels in theimage sensor 12 that are arranged in an oblique direction as exemplified inFIG. 11 . Thepixel row 150 is constituted by an Mg pixel (Mg-a) included in each second set P2 and located at the left upper position of eachmicrolens 40, an Mg pixel (Mg-d) included in each second set P2 and located at the right lower position of eachmicrolens 40, an Mg pixel (Mg-a) included in each third set P3 and located at the left upper position of eachmicrolens 40, and an Mg pixel (Mg-d) included in each third set P3 and located at the right lower position of eachmicrolens 40. As shown inFIG. 12 , the light flux A that passes s15 through thefirst region 81 on theexit pupil 80 and a light flux D that passes through afourth region 84 on theexit pupil 80 enter the pixels that constitute thepixel row 150. The light flux A enters the Mg pixel (Mg-a) located at the left upper position of eachmicrolens 40. The light flux D enters the Mg pixel (Mg-d) located at the right lower part of eachmicrolens 40. - Upon focusing, the
image sensor 12 is in a state in which a sharp image can be formed therein, so that in thepixel row 150 as described above, a signal waveform (signal sequence a1, a2, a3, a4, . . . ) obtained from the Mg pixels (Mg-a) that receive the light fluxes A and a signal waveform (signal sequence c1, c2, c3, c4, . . . ) obtained from the Mg pixels (Mg-c) that receive the light fluxes C overlap in their shape. - On the other hand, upon non-focusing, the signal waveform (signal sequence a1, a2, a3, a4 . . . ) provided by the light fluxes A and the signal waveform (signal sequence d1, d2, d3, d4, . . . ) provided by the light fluxes D have positional relationships (deviation direction and deviation amount) therebetween, which are different from each other, according to the deviation (defocus amount) from the focused state.
- The
body control unit 14 calculates the focusing condition (defocus amount) of the focus position by theinterchangeable lens 2 on the basis of the positional relationship between the signal waveform (signal sequence a1, a2, a3, a4 . . . ) provided by the light fluxes A and the signal waveform (signal sequence d1, d2, d3, d4, . . . ) provided by the light fluxes D and transmits the result of calculation that serves as camera information to thelens control unit 5. As thelens control unit 5 moves the focusinglens 7 back and forth along the optical axis direction on the basis of the camera information, the focus is adjusted so that a sharp image can be formed on theimage sensor 12. - <Image Signal Generation Processing>
- Next, explanation is made on an example in which image signals are obtained from the
image sensor 12 referring toFIG. 13 throughFIG. 17 . In the present embodiment, any of the following three methods is used as image signal generation processing for generating a color image signal on the basis of output signals from the lowerphotoelectric conversion layer 43. Thebody control unit 14 executes image signal generation processing by a method indicated by an initial setting in advance. - (First Image Signal Generation Processing)
-
FIG. 13 presents a diagram illustrating first image signal generation processing. Thebody control unit 14, which executes the first image signal generation processing, treats four pixels that receive light fluxes through the one and thesame microlens 40 as oneset 200, as shown inFIG. 13(a) . Each of thesets 200 includes two G pixels, one B pixel and one R pixel. - The
body control unit 14 treats, for any particular one of thesets 200, an output signal from the R pixel as an R image signal of the particular one set 200, an output signal from the B pixel as a B image signal of the particular one set 200, and an average value of output signals from the two G pixels as a G image signal of the particular one set 200. As a result, thebody control unit 14 can obtain color image signals (RGB) in a number that is ¼ times the number of the pixels included in the lowerphotoelectric conversion layer 43 of theimage sensor 12, as shown inFIG. 13(b) . Thebody control unit 14 generates an image file for recording by using the thus-obtained color image signals. - As described above, in the first image signal generation processing, color image signals can be obtained without executing color interpolation processing for interpolating color signals.
- (Second Image Signal Generation Processing)
-
FIG. 14 presents a diagram illustrating second image signal generation processing. Thebody control unit 14, which executes the second image signal generation processing, treats adjacent four pixels in a two-by-two matrix having the same color as oneset 210 as shown inFIG. 14(a) . - The
body control unit 14 treats a signal obtained by adding output signals from the four pixels included in any particular one of thesets 120 as an image signal of theparticular set 210. Specifically, in the case of any particular one set 210 that is all constituted by R pixels, thebody control unit 14 treats a signal obtained by adding output signals from the four R pixels as an R image signal of the particular one set 210. In the case of any particular one set 210 that is all constituted by G pixels, thebody control unit 14 treats a signal obtained by adding output signals from the four G pixels as a G image signal of the particular one set 210. In the case of any particular one set 210 that is all constituted by B pixels, thebody control unit 14 treats a signal obtained by adding output signals from the four B pixels as a B image signal of the particular one set 210. As a result, thebody control unit 14 can obtain color image signals of a Bayer arrangement in a number that is ¼ times the number of the pixels included in the lowerphotoelectric conversion layer 43 of theimage sensor 12, as shown inFIG. 14(b) . - And now, depending on the angle of incidence of the light fluxes that enter each microlens 40 it may happen that the four pixels arranged behind each
microlens 40 receive uneven amounts of light. For instance, at a certain incident angle θ1, the amount of light received by the pixel located at the left upper position of eachmicrolens 40 is relatively large while the amount of light received by the pixel located at the right lower position of eachmicrolens 40 is relatively small. At another incident angle θ2, the amount of light received by the pixel located at the left upper position of eachmicrolens 40 is relatively small while the amount of light received by the pixel located at the right lower position of eachmicrolens 40 is relatively large. - In the second image signal generation processing, as a signal obtained by adding output signals from four pixels located at positions corresponding to different positions (left upper, right upper, left lower, and right lower positions) of each microlens 40 (that is, four pixels included in any particular one set 210) is treated as an image signal of the particular one set 210, an optimal image signal can be generated independently of the incident angles of the light fluxes that enter the
microlens 40. - In addition, the
body control unit 14 generates, in an image signal of a Bayer arrangement in any particular one set 210, a color component that is in short by interpolation processing using signals from a plurality ofother sets 210 that are adjacent to the particular one set 210. For instance, in the case of any particular one set 210 that is all constituted by G pixels, as there is present neither R image signal nor B image signal therefrom, color interpolation processing is executed by using signals from a plurality ofother sets 210 that are circumjacent to the particular one set 210. Since such color interpolation processing in the Bayer arrangement is known in the art, detailed description thereof is omitted herein. Thebody control unit 14 generates a file for recording by using color image signals (RGB) obtained by executing this color interpolation processing. - (Third Image Signal Generation Processing)
- The
body control unit 14, which executes third signal generation processing, first executes color interpolation processing for interpolating a color component that is in short in each pixel. -
FIG. 15 presents a diagram illustrating processing for interpolating a G image signal. Thebody control unit 14 generates, at a position of any particular one pixel out of the R pixels and the B pixels, a G image signal by using output signals from four G pixels located near the particular one pixel by interpolation processing. For instance, in case that a G image signal is to be interpolated at the position of the R pixel in a thick-frame inFIG. 15(a) , output signals from four G pixels (G1 through G4) that are located near the particular R pixel are used. Thebody control unit 14 defines (αG1+βG2+γG3+δG4)/4 as a G image signal of the particular R pixel. It is to be noted that a through 8 each are coefficients depending on the distances from the particular R pixel. The smaller the distance from the target pixel to be interpolated, the greater the coefficient is. In this case, as the G pixels G1 and G2 are closer to the particular R pixel than the G pixels G3 and G4 are, it is set that α=β>γ=δ. - In this manner, the
body control unit 14 executes processing for interpolating G image signals at positions of R pixels and B pixels, so that a G image signal can be obtained at a position of eachpixel 30 as shown inFIG. 15(b) . -
FIG. 16 presents a diagram illustrating processing for interpolating R image signals. As shown inFIG. 16 , thebody control unit 14 treats four pixels constituted by adjacent four pixels in a two-by-two matrix having the same color as oneset 220. Thebody control unit 14 defines a signal obtained by adding output signals from the four pixels in a particular one set 220 of R pixels to be an R image signal of the particular one set 220. Thebody control unit 14 interpolates R image signals in a particular one set 220 of G pixels and a particular one set 220 of B pixels using R image signals from a plurality of sets of R pixels that are circumjacent to the particular one set 220 of G pixels and the particular one set 220 of B pixels. Since thesets 220 each form a Bayer arrangement as shown inFIG. 16(b) , thebody control unit 14 can execute this interpolation processing by using color interpolation processing for Bayer arrangements that is known in the art. - The
body control unit 14 defines an R image signal that is interpolated in a particular one set 220 of B pixels and divided by 4 (R/4) to be an R image signal for all the four G pixels that constitute the particular one set 220 of G pixels. Similarly, thebody control unit 14 defines an R image signal that is interpolated in a particular one set 220 of B pixels and divided by 4 (R/4) to be an R image signal for all the four B pixels that constitute the particular one set 220 of B pixels. In this manner, thebody control unit 14 can obtain an R image signal at the position of each of thepixels 30 by executing interpolation processing for interpolating R image signals at the positions of the G pixels and of the B pixels, as shown inFIG. 16(c) . - It is to be noted that interpolation processing for interpolating B image signals is similar to the interpolation processing for interpolating R image signals, so that detailed description thereof is omitted here. The
body control unit 14 can obtain a B image signal at the position of each of thepixels 30 by executing interpolation processing for interpolating B image signals at the positions of the R pixels and of the G pixels. - The
body control unit 14 executes the color interpolation processing as described above to obtain an image signal of RGB at the position of each of thepixels 30 as shown inFIG. 17(a) . In addition, thebody control unit 14 obtains a luminance (brightness) signal Y at the position of each of thepixels 30 by using the image signal of RGB at the position of each of thepixels 30. For instance, thebody control unit 14 defines 0.299R+0.587G+0.114B as a luminance signal Y. - In addition, the
body control unit 14 defines a signal (R-Y) obtained by deducing the luminance signal Y from the R image signal at the position of each of thepixels 30 as a color difference (chrominance) signal Cr. Thebody control unit 14 defines a signal (B-Y) obtained by deducing the luminance signal Y from the B image signal at the position of each of thepixels 30 as a color difference (chrominance) signal Cb. - As a result, the
body control unit 14 can obtain the luminance signal Y and the color difference signals Cr and Cb at the position of each of thepixels 30 as shown inFIG. 17(b) . By using the color image signals (YCrCb) thus obtained, thebody control unit 14 generates an image file for recording having a resolution higher than those generated according to the first image signal generation processing and the second image signal generation processing, respectively. - <Shooting Processing>
-
FIG. 18 presents a flowchart illustrating the flow of imaging processing to be executed by thebody control unit 14. Thebody control unit 14, when a main switch (not shown) that constitutes the operatingmember 18 is turned ON, starts up a program that executes the processing exemplified inFIG. 18 . - In step S11 in
FIG. 18 , thebody control unit 14 controls theimage sensor 12 to start photoelectric conversion at a predetermined frame rate. Thebody control unit 14 makes a decision as to whether or not a command for shooting is issued while it controls the liquidcrystal display element 16 to successively reproduce and display a through image based on the image signals from the lowerphotoelectric conversion layer 43. The through image is an image for monitoring that is obtained before shooting is commanded. When a release button that constitutes the operatingmember 18 is pushed down, thebody control unit 14 makes a positive decision in step S11 and the program proceeds to step S12. When the release button is not pushed down, thebody control unit 14 makes a negative decision in step S11 and the program proceeds to step 818. - In step S18, the
body control unit 14 makes a decision as to whether or not time is up. When thebody control unit 14 measures a predetermined time (for instance, 5 seconds), it makes a positive decision in step S18 and the processing inFIG. 15 is terminated. When the measured time is shorter than the predetermined time, thebody control unit 14 makes a negative decision in step S18 and the program returns to step S11. - In step S12, the
body control unit 14 executes AE processing and AF processing. In the AE processing, exposure is calculated on the basis of the level of image signal for the through image in order to decide aperture value AV and shutter speed TV so that optimal exposure can be obtained. In the AF processing, the focus detection processing described above is executed based on an output signal sequence from the pixel row included in the focus detection area that is set in the upperphotoelectric conversion layer 41. When thebody control unit 14 completed execution of the AE and AF processing described above, the program proceeds to step S13. - In step S13, the
body control unit 14 executes shooting processing and the program proceeds to step S14. Specifically, thebody control unit 14 controls the diaphragm 6 based on the AV and controls theimage sensor 12 to perform photoelectric conversion for recording for a storage time on the basis of the TV. In step S14, thebody control unit 14 executes the image signal generation processing by using output signals from the lowerphotoelectric conversion layer 43 and the obtained image signal is subjected to predetermined image processing (gradation processing, contour enhancement, white balance adjustment processing and so on). When thebody control unit 14 has executed the image processing, the program proceeds to step S15. - In step S15, the
body control unit 14 controls the liquidcrystal display element 16 to display the captured image thereon and the program proceeds to step S16. In step S16, thebody control unit 14 generates an image file for recording and the program proceeds to step S17. In step S17, thebody control unit 14 records the generated image file in amemory card 20 and terminates the processing inFIG. 18 . - According to the embodiment described above, the following operations and advantageous effects can be obtained.
- (1) The
digital camera system 1 is configured as follows. That is, thedigital camera system 1 includes theimage sensor 12 that captures an image of a subject with a light flux from a subject that passes through theinterchangeable lens 2; thebody control unit 14 that generates an image signal on the basis of output signals from theimage sensor 12; and thebody control unit 14 detecting a focusing condition of theinterchangeable lens 2 according to a phase detection method. Theimage sensor 12 has a pixel group in the upperphotoelectric conversion layer 41 and a pixel group in the lowerphotoelectric conversion layer 43 that receives light fluxes from the subject that have passed through each pixel in the upperphotoelectric conversion layer 41 and a microlens group arranged such that the light fluxes from the subject are guided to the pixel group in the upperphotoelectric conversion layer 41; the pixel group in the upperphotoelectric conversion layer 41 includes Cy pixels, Ye pixels, and Mg pixels having first, second, and third spectral sensitivities, respectively, differing from each other, arranged in a two-dimensional pattern such that behind eachmicrolens 40, there are arranged one Cy pixel, one Ye pixel and two Mg pixels in a two-by-two matrix, with these four pixels receiving four light fluxes A through D, respectively, that pass through fourpupil regions 81 through 84, respectively, of theexit pupil 80. The pixel group in the lowerphotoelectric conversion layer 43 includes R pixels, B pixels, and G pixels having fourth, fifth and sixth spectral sensitivities, respectively, which are complementary to the first, second and third spectral sensitivities, respectively, of the pixel group in the upperphotoelectric conversion layer 41, arranged in a two dimensional pattern, with the positions of the Cy pixel, the Ye pixel and the Mg pixel in the upperphotoelectric conversion layer 41 and the positions of the R pixel, the B pixel and the G pixel in the lowerphotoelectric conversion layer 43 are set such that the R pixel, the B pixel and the G pixel receive light fluxes that have passed through the Cy pixel, the Ye pixel and the Mg pixel, respectively. Thebody control unit 14 generates image signals on the basis of output signals from one of the pixel group in the upperphotoelectric conversion layer 41 and the pixel group in the lowerphotoelectric conversion layer 43 whereas thebody control unit 14 detects the focusing condition on the basis of output signals from the other of the pixel group in the upperphotoelectric conversion layer 41 and the pixel group in the lowerphotoelectric conversion layer 43. With this configuration, the generation of image signals and the focus detection according to the phase detection method on the basis of output signals from theimage sensor 12 can be executed without providing theimage sensor 12 with any pixels that are dedicated for focus detection.
(2) In thedigital camera system 1 in (1) above, it is further configured such that among the Cy pixels, the Ye pixels and the Mg pixels in the pixel group of the upperphotoelectric conversion layer 41, those pixels having substantially the same spectral sensitivities (in other words, pixels having the same color) are adjacently arranged so as to form two-by-two matrices and four other pixels that are adjacent to each of the two-by-two matrices are arranged behind fourdifferent microlenses 40, respectively, such that their positions with respect to the correspondingmicrolenses 40 are different from each other. With this configuration, the incident light fluxes can be optimally photoelectrically converted regardless of the angle of incidence of light to themicrolens 40.
(3) In thedigital camera system 1 in (2) above, a configuration is adopted, in which the pixel group in the upperphotoelectric conversion layer 41 provides output signals relating to Cy, Ye and Mg from the first, second and third pixels, respectively, whereas the pixel group in the lowerphotoelectric conversion layer 43 provides output signals relating to complementary colors of Cy, Ye and Mg from the fourth, fifth and sixth pixels, respectively. With this configuration, red, green and blue color image signals can be obtained from the output signals from theimage sensor 12.
(4) In thedigital camera system 1 in (3) above, a configuration is adopted, in which the pixel groups of the upperphotoelectric conversion layer 41 and the lowerphotoelectric conversion layer 43 each are formed by arranging, in a two-dimensional pattern, a plurality of sets of pixels, each of which sets includes four pixels arrayed in a two-by-two matrix behind onemicrolens 40, the sets include first through fourth sets P1 through P4 in which arrangements of pixels are different from each other. The pixel group in the upperphotoelectric conversion layer 41 is configured such that in the first set P1, a Cy pixel and an Mg pixel are arrayed adjacent to each other in the horizontal direction and an Mg pixel and a Ye pixel are arrayed adjacent to the horizontally arrayed Cy and Mg pixels, respectively, in the vertical direction. In the second set P2, an Mg pixel and a Cy pixel are arrayed adjacent to each other in the horizontal direction and a Ye pixel and an Mg pixel are arrayed adjacent to the horizontally arrayed Mg and Cy pixels, respectively, in the vertical direction. In the third set P3, an Mg pixel and a Ye pixel are arrayed adjacent to each other in the horizontal direction and a Cy pixel and an Mg pixel are arrayed adjacent to the horizontally arrayed Mg and Ye pixels, respectively, in the vertical direction. In the fourth set P4, a Ye pixel and an Mg pixel are arrayed adjacent to each other in the horizontal direction and an Mg pixel and a Cy pixel are arrayed adjacent to the horizontally arrayed Mg and Ye pixels, respectively, in the vertical direction. The first set P1 and the second set P2 are adjacent to each other in the horizontal direction and alternately arrayed in a repeated manner in the horizontal direction. The third set P3 and the fourth set P4 are adjacent to each other in the horizontal direction and alternately arrayed in a repeated manner in the horizontal direction. A first line L1 formed by the first set P1 and the second set P2 and a second line L2 formed by the third set P3 and the fourth set P4 are adjacent to each other in the vertical direction and alternately arrayed in a repeated manner in the vertical direction. With this configuration, focus detection according to the phase detection method can be performed on the basis of an output signal from theimage sensor 12 and any of the first through third image signal processing described above can be executed.
(5) In thedigital camera system 1 in any one of (2) through (4) above, a configuration is adopted, in which thebody control unit 14 adds output signals from four adjacent R pixels in a two-by-two matrix, adds output signals from four adjacent B pixels in a two-by-two matrix, and adds output signals from four adjacent G pixels in a two-by-two matrix, thereby forming an image signal of a Bayer arrangement (that is, executing the second image signal generation processing). With this configuration, optimal image signals can be generated regardless of the incident angle of light into themicrolenses 40. Furthermore, a conventional image processing engine that executes color interpolation of a Bayer arrangement can be used in the color interpolation processing.
(6) In thedigital camera system 1 in any one of (1) through (4) above, a configuration is adopted, in which thebody control unit 14 obtains three color signals at a position of each of the microlenses on the basis of output signals from the R, B and G pixels positioned behind each of the microlenses 40 (that is, executes the first image signal generation processing). With this configuration, color image signals can be obtained without executing color interpolation processing.
(7) In thedigital camera system 1 in any one of (1) through (4) above, a configuration is adopted, in which thebody control unit 14 executes color interpolation processing for generating signals for two other spectral components at each position of R, B and G pixels to obtain three color signals and generates a luminance signal and color difference signals on the basis of the three color signals (that is, executes the third image signal generation processing). With this configuration, image signals having high resolutions can be obtained.
(8) In thedigital camera system 1 in any one of (1) through (7) above, a configuration is adopted, in which thebody control unit 14 detects the focusing condition of theinterchangeable lens 2 on the basis of outputs signals from a pair of pixels out of the pixels in the pixel group of the upperphotoelectric conversion layer 41, the pair of pixels having substantially the same spectral sensitivities and having different positions from each other with respect to themicrolens 40. With this configuration, the focusing condition can be detected according to the phase detection method appropriately based on the output signals from theimage sensor 12.
(9) In thedigital camera system 1 in (4) above, a configuration is adopted, in which thebody control unit 14 detects the focusing condition of theinterchangeable lens 2 in the horizontal direction on the basis of output signals from Mg pixels included in the first set P1 and the second set P2, respectively, among the pixels in the pixel groups in the upperphotoelectric conversion layer 41. With this configuration, the focusing condition can be detected in the horizontal direction of theimage sensor 12 according to the phase detection method appropriately based on the output signals from theimage sensor 12.
(10) In thedigital camera system 1 in (4) above, a configuration is adopted, in which thebody control unit 14 detects the focusing condition of the imaging optical system in the vertical direction on the basis of output signals from the Mg pixels included in the second set P2 and the fourth set P4, respectively, among the pixels in the pixel groups of the upperphotoelectric conversion layer 41. With this configuration, the focusing condition can be detected according to the phase detection method appropriately.
(11) In thedigital camera system 1 in (4) above, a configuration is adopted, in which thebody control unit 14 detects the focusing condition of theinterchangeable lens 2 in a direction oblique to the horizontal direction on the basis of output signals from the Mg pixels included in the second set P2 and the third set P3, respectively, among the pixels in the pixel group of the upperphotoelectric conversion layer 41. With this configuration, the focusing condition can be appropriately detected according to the phase detection method. - In the embodiment described above, the focus detection processing is executed by using output signals from the Mg pixels in the upper
photoelectric conversion layer 41. However, the focus detection processing may also be executed by using output signals from the Cy pixels or the Ye pixels. - The
body control unit 14 according to Variation Example 1 is configured to obtain evaluation values by using output signals from the upperphotoelectric conversion layer 41. The evaluation values are cumulative values of output signals for each of; for instance, Cy, Mg or Ye pixels. When the cumulative value for the Mg pixels is low, it may be possible that output signals from the Mg pixels cannot afford appropriate calculation of defocus amounts. Then, thebody control unit 14 according to Variation Example 1 executes, when the cumulative value for the Mg pixels is equal to or lower than a predetermined threshold value, the above-mentioned focus detection processing by using either one of Cy pixel or Ye pixel that has a greater cumulative value than that of the rest. With this configuration, appropriate focus detection processing can be executed even when a subject having a small amount of Mg components is shot. - In the embodiment described above, out of the first through third image signal generation processing, the processing that is indicated by initial setting is used to generate image signals for recording. However, the present invention is not limited thereto.
- For instance, the
body control unit 14 according to Variation Example 2, when a through image is to be displayed, selects the first image signal generation processing in which image signals can be generated without executing color interpolation processing and generates image signals by using the selected first image signal generation processing. On the other hand, for the images for recording, the third image signal generation processing capable of generating image signals having high resolutions is selected and image signals are generated by using the selected third image signal generation processing. As described above, thebody control unit 14 according to Variation Example 2 is configured to select, upon image signal generation, any of the first, second and third image signal generation processing. With this configuration, image signal generation processing that is suitable for uses of images to be generated can be selected. For instance, the first image signal generation processing which does not require any color interpolation processing is selected in a scene where it is desired to display images on a real-time basis, whereas the third image signal generation processing is selected in a scene where it is desired to record images with high image quality. - In addition, the
body control unit 14 may be configured to generate image signals by the first or second image signal generation processing for video images or by the third image signal generation processing for still images. - In addition, the
body control unit 14 may be configured to generate image signals by using, for instance, both of the first image signal generation processing and the second image signal generation processing. In this case, thebody control unit 14 controls, for instance, both the image generated by the first image signal generation processing and the image generated by the second image signal generation processing to be displayed on a display device (not shown) on the rear side. Thebody control unit 14 records one of the two displayed images, which one is selected by the user through the operatingmember 18 into thememory card 20. - In the embodiment described above, the configuration is adopted in which the defocus amount in the horizontal direction is obtained on the basis of output signal from the
pixel row 90 constituted by the Mg pixel (Mg-b) included in each of the first sets P1 and the Mg pixel (Mg-a) included in each of the second sets P2 among the pixels in the upperphotoelectric conversion layer 41. However, the present invention is not limited thereto. A configuration may be adopted, in which the defocus amount in the horizontal direction is obtained on the basis of output signals from a pixel row constituted by the Mg pixel (Mg-d) included in each of the third sets P3 and the Mg pixel (Mg-c) included in each of the fourth sets P4. Alternatively, the defocus amount in the horizontal direction may be obtained on the basis of both the above pixel row and thepixel row 90. - In addition, in the embodiment described above, the configuration is adopted in which the defocus amount in the vertical direction is obtained on the basis of output signals from the
pixel row 120 constituted by the Mg pixel (Mg-a) included in each of the second sets P2 and the Mg pixel (Mg-c) included in each of the fourth sets P4. However, the present invention is not limited thereto. A configuration may be adopted in which the defocus amount in the vertical direction is obtained on the basis of output signals from a pixel row constituted by the Mg pixel (Mg-b) included in each of the first sets P1 and the Mg pixel (Mg-d) included in each of the third sets P3. Alternatively, the defocus amount in the horizontal direction may be obtained on the basis of both the above pixel row and thepixel row 120. - In the embodiment described above, the configuration is adopted, in which the defocus amount in an oblique direction is obtained on the basis of output signals from the
pixel row 150 constituted by the Mg pixels (Mg-a) and (Mg-d) included in each of the second set P2 and the Mg pixels (Mg-a) and (Mg-d) included in each of the third set P3. However, the present invention is not limited thereto. A configuration may be adopted, in which the defocus amount in the oblique direction is obtained on the basis of the pixel row constituted by the Mg pixels (Mg-b) and (Mg-c) included in each of the first set P1 and the Mg pixels (Mg-b) and (Mg-c) included in each of the fourth set P4. Alternatively, the defocus amount in the oblique direction may be obtained on the basis of both the above pixel row and thepixel row 150. - In the embodiment described above, a configuration is adopted, in which Mg pixels, Cy pixels and Ye pixels are provided in the upper
photoelectric conversion layer 41 and G pixels, R pixels and B pixels are provided in the lowerphotoelectric conversion layer 43. In place of this, a configuration may be adopted, in which G pixels, R pixels and B pixels are provided in the upper photoelectric conversion layer and Mg pixels, Cy pixels and Ye pixels are provided in the lower photoelectric conversion layer - In the embodiment described above, the present invention is applied to the
digital camera system 1 having a configuration such that theinterchangeable lens 2 is mounted to thecamera body 3. However, the present invention is not limited thereto. For instance, the present invention may also be applied to a lens-integrated digital camera. - The above description is by way of example and the present invention is not limited to the embodiment described above. In addition, the embodiment described above may be combined with the configuration(s) of one or more of the variation examples in any desired manner.
- Although various embodiments and variations have been described above, the present invention is not limited thereto. Other aspects that may be conceivable within the scope of technical ideas of the present invention are also embraced by the present invention.
- The disclosure of the following priority application is herein incorporated by reference: Japanese Patent Application No. 2012-081165 (filed on Mar. 30, 2012).
Claims (1)
Priority Applications (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US16/401,436 US20190260951A1 (en) | 2012-03-30 | 2019-05-02 | Image-capturing device and image sensor |
| US17/387,390 US20210358981A1 (en) | 2012-03-30 | 2021-07-28 | Image-capturing device and image sensor |
Applications Claiming Priority (6)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2012081165 | 2012-03-30 | ||
| JP2012-081165 | 2012-03-30 | ||
| PCT/JP2013/059621 WO2013147198A1 (en) | 2012-03-30 | 2013-03-29 | Imaging device and image sensor |
| US201514389626A | 2015-04-08 | 2015-04-08 | |
| US15/794,710 US10389959B2 (en) | 2012-03-30 | 2017-10-26 | Image-capturing device and image sensor |
| US16/401,436 US20190260951A1 (en) | 2012-03-30 | 2019-05-02 | Image-capturing device and image sensor |
Related Parent Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US15/794,710 Continuation US10389959B2 (en) | 2012-03-30 | 2017-10-26 | Image-capturing device and image sensor |
Related Child Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US17/387,390 Continuation US20210358981A1 (en) | 2012-03-30 | 2021-07-28 | Image-capturing device and image sensor |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| US20190260951A1 true US20190260951A1 (en) | 2019-08-22 |
Family
ID=49260439
Family Applications (4)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US14/389,626 Active 2033-11-07 US9826183B2 (en) | 2012-03-30 | 2013-03-29 | Image-capturing device and image sensor |
| US15/794,710 Active US10389959B2 (en) | 2012-03-30 | 2017-10-26 | Image-capturing device and image sensor |
| US16/401,436 Abandoned US20190260951A1 (en) | 2012-03-30 | 2019-05-02 | Image-capturing device and image sensor |
| US17/387,390 Abandoned US20210358981A1 (en) | 2012-03-30 | 2021-07-28 | Image-capturing device and image sensor |
Family Applications Before (2)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US14/389,626 Active 2033-11-07 US9826183B2 (en) | 2012-03-30 | 2013-03-29 | Image-capturing device and image sensor |
| US15/794,710 Active US10389959B2 (en) | 2012-03-30 | 2017-10-26 | Image-capturing device and image sensor |
Family Applications After (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US17/387,390 Abandoned US20210358981A1 (en) | 2012-03-30 | 2021-07-28 | Image-capturing device and image sensor |
Country Status (5)
| Country | Link |
|---|---|
| US (4) | US9826183B2 (en) |
| EP (1) | EP2835965B1 (en) |
| JP (3) | JP6264284B2 (en) |
| CN (2) | CN111711746A (en) |
| WO (1) | WO2013147198A1 (en) |
Families Citing this family (25)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2013147198A1 (en) * | 2012-03-30 | 2013-10-03 | 株式会社ニコン | Imaging device and image sensor |
| US20150062391A1 (en) * | 2012-03-30 | 2015-03-05 | Nikon Corporation | Image sensor, photographing method, and image-capturing device |
| JP5778873B2 (en) * | 2012-12-07 | 2015-09-16 | 富士フイルム株式会社 | Image processing apparatus, image processing method and program, and recording medium |
| TWI620445B (en) | 2013-03-25 | 2018-04-01 | 新力股份有限公司 | Camera element and electronic equipment |
| JP6368993B2 (en) * | 2013-07-24 | 2018-08-08 | 株式会社ニコン | Imaging device |
| JP6397281B2 (en) * | 2013-10-23 | 2018-09-26 | キヤノン株式会社 | Imaging apparatus, control method thereof, and program |
| JP6458343B2 (en) * | 2014-02-27 | 2019-01-30 | 株式会社ニコン | Imaging device |
| JP6408372B2 (en) * | 2014-03-31 | 2018-10-17 | ソニーセミコンダクタソリューションズ株式会社 | SOLID-STATE IMAGING DEVICE, ITS DRIVE CONTROL METHOD, AND ELECTRONIC DEVICE |
| JP2016058559A (en) | 2014-09-10 | 2016-04-21 | ソニー株式会社 | Solid-state imaging device, driving method thereof, and electronic apparatus |
| US9967501B2 (en) | 2014-10-08 | 2018-05-08 | Panasonic Intellectual Property Management Co., Ltd. | Imaging device |
| US10249657B2 (en) | 2015-03-13 | 2019-04-02 | Sony Semiconductor Solutions Corporation | Solid-state image sensing device, drive method, and electronic apparatus |
| JP6541503B2 (en) * | 2015-08-11 | 2019-07-10 | キヤノン株式会社 | Image pickup apparatus, image processing apparatus, image processing method and program |
| WO2017119477A1 (en) * | 2016-01-08 | 2017-07-13 | 株式会社ニコン | Imaging element and imaging device |
| JP6782431B2 (en) * | 2016-01-22 | 2020-11-11 | パナソニックIpマネジメント株式会社 | Imaging device |
| CN107026961B (en) | 2016-01-29 | 2021-02-12 | 松下知识产权经营株式会社 | Image pickup apparatus |
| DE112016007230T5 (en) * | 2016-10-21 | 2019-06-13 | Olympus Corporation | Endoscope probe, endoscope processor and endoscope adapter |
| US10714520B1 (en) * | 2017-08-04 | 2020-07-14 | Facebook Technologies, Llc | Manufacturing an on-chip microlens array |
| US11120082B2 (en) | 2018-04-18 | 2021-09-14 | Oracle International Corporation | Efficient, in-memory, relational representation for heterogeneous graphs |
| JP6766095B2 (en) * | 2018-06-08 | 2020-10-07 | キヤノン株式会社 | Semiconductor substrates for imaging devices, imaging systems, mobiles, and stacks |
| CN115236829B (en) * | 2018-07-20 | 2024-03-19 | 株式会社尼康 | Shooting method |
| CN109905681B (en) * | 2019-02-01 | 2021-07-16 | 华为技术有限公司 | Image sensor, method for acquiring image data therefrom, and imaging device |
| KR102885929B1 (en) * | 2019-10-22 | 2025-11-12 | 삼성전자주식회사 | Image sensor including a plurality of auto focusing pixel groups |
| WO2021174529A1 (en) * | 2020-03-06 | 2021-09-10 | Oppo广东移动通信有限公司 | Image sensor, imaging apparatus, electronic device, image processing system and signal processing method |
| CN111464733B (en) * | 2020-05-22 | 2021-10-01 | Oppo广东移动通信有限公司 | Control method, camera assembly and mobile terminal |
| JP7474129B2 (en) * | 2020-06-22 | 2024-04-24 | キヤノン株式会社 | Image sensor, image pickup device, and focus detection method |
Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20100091161A1 (en) * | 2007-06-16 | 2010-04-15 | Nikon Corporation | Solid-state image sensor and imaging apparatus equipped with solid-state image sensor |
| US20100141771A1 (en) * | 2007-07-09 | 2010-06-10 | Boly Media Communications (Shenzhen) Co., Ltd | Multi-Spectrum Sensing Device And Manufacturing Methods Thereof |
| US20100282945A1 (en) * | 2009-05-11 | 2010-11-11 | Sony Corporation | Two-dimensional solid-state image capture device and polarization-light data processing method therefor |
| US20120193515A1 (en) * | 2011-01-28 | 2012-08-02 | Gennadiy Agranov | Imagers with depth sensing capabilities |
| US9826183B2 (en) * | 2012-03-30 | 2017-11-21 | Nikon Corporation | Image-capturing device and image sensor |
Family Cites Families (46)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPS59108457A (en) * | 1982-12-14 | 1984-06-22 | Olympus Optical Co Ltd | solid-state image sensor |
| EP2317567A1 (en) | 1998-08-19 | 2011-05-04 | The Trustees Of Princeton University | Organic photosensitive optoelectronic device |
| JP2001021792A (en) | 1999-07-09 | 2001-01-26 | Olympus Optical Co Ltd | Focus detection system |
| US7742088B2 (en) | 2002-11-19 | 2010-06-22 | Fujifilm Corporation | Image sensor and digital camera |
| JP4264248B2 (en) | 2002-11-19 | 2009-05-13 | 富士フイルム株式会社 | Color solid-state imaging device |
| JP2004200902A (en) | 2002-12-17 | 2004-07-15 | Nikon Corp | Image processing device, electronic camera, and image processing program |
| US6958862B1 (en) * | 2003-04-21 | 2005-10-25 | Foveon, Inc. | Use of a lenslet array with a vertically stacked pixel array |
| US20040223071A1 (en) * | 2003-05-08 | 2004-11-11 | David Wells | Multiple microlens system for image sensors or display units |
| JP4500706B2 (en) | 2005-02-23 | 2010-07-14 | 富士フイルム株式会社 | Photoelectric conversion film stack type solid-state imaging device |
| KR100665177B1 (en) * | 2005-05-25 | 2007-01-09 | 삼성전기주식회사 | Image sensor for semiconductor photosensitive device and image processing device using same |
| JP2007011070A (en) | 2005-06-30 | 2007-01-18 | Olympus Imaging Corp | Digital single-lens reflex camera |
| JP2007067075A (en) | 2005-08-30 | 2007-03-15 | Nippon Hoso Kyokai <Nhk> | Color image sensor |
| US7419844B2 (en) | 2006-03-17 | 2008-09-02 | Sharp Laboratories Of America, Inc. | Real-time CMOS imager having stacked photodiodes fabricated on SOI wafer |
| JP4935162B2 (en) | 2006-04-11 | 2012-05-23 | 株式会社ニコン | Imaging apparatus, camera, and image processing method |
| US7711261B2 (en) | 2006-04-11 | 2010-05-04 | Nikon Corporation | Imaging device, camera and image processing method |
| JP4952060B2 (en) * | 2006-05-26 | 2012-06-13 | 株式会社ニコン | Imaging device |
| JP2007324405A (en) * | 2006-06-01 | 2007-12-13 | Fujifilm Corp | Solid-state image sensor |
| JP4867552B2 (en) | 2006-09-28 | 2012-02-01 | 株式会社ニコン | Imaging device |
| JP4321579B2 (en) | 2006-11-28 | 2009-08-26 | ソニー株式会社 | Imaging device |
| US7859588B2 (en) | 2007-03-09 | 2010-12-28 | Eastman Kodak Company | Method and apparatus for operating a dual lens camera to augment an image |
| US7676146B2 (en) | 2007-03-09 | 2010-03-09 | Eastman Kodak Company | Camera using multiple lenses and image sensors to provide improved focusing capability |
| JP2008258474A (en) | 2007-04-06 | 2008-10-23 | Sony Corp | Solid-state imaging device and imaging device |
| WO2009020031A1 (en) | 2007-08-06 | 2009-02-12 | Canon Kabushiki Kaisha | Image sensing apparatus |
| JP5171178B2 (en) | 2007-09-13 | 2013-03-27 | 富士フイルム株式会社 | Image sensor and manufacturing method thereof |
| JP2009099867A (en) | 2007-10-18 | 2009-05-07 | Fujifilm Corp | Photoelectric conversion element and imaging element |
| JP5032954B2 (en) | 2007-11-27 | 2012-09-26 | 日本放送協会 | Color imaging device |
| US8106426B2 (en) | 2008-02-04 | 2012-01-31 | Sharp Laboratories Of America, Inc. | Full color CMOS imager filter |
| JP5298638B2 (en) | 2008-02-14 | 2013-09-25 | 株式会社ニコン | Image processing apparatus, imaging apparatus, correction coefficient calculation method, and image processing program |
| JP4902892B2 (en) | 2008-02-22 | 2012-03-21 | パナソニック株式会社 | Imaging device |
| JP5058128B2 (en) * | 2008-10-31 | 2012-10-24 | シャープ株式会社 | Imaging device and portable device |
| TWI422020B (en) * | 2008-12-08 | 2014-01-01 | Sony Corp | Solid-state imaging device |
| JP5359465B2 (en) * | 2009-03-31 | 2013-12-04 | ソニー株式会社 | Solid-state imaging device, signal processing method for solid-state imaging device, and imaging device |
| JP5246424B2 (en) * | 2009-05-11 | 2013-07-24 | ソニー株式会社 | Imaging device |
| JP5671789B2 (en) | 2009-08-10 | 2015-02-18 | ソニー株式会社 | Solid-state imaging device, manufacturing method thereof, and imaging device |
| JP5537905B2 (en) | 2009-11-10 | 2014-07-02 | 富士フイルム株式会社 | Imaging device and imaging apparatus |
| KR101709941B1 (en) | 2009-12-02 | 2017-02-27 | 삼성전자주식회사 | Image sensor, image processing apparatus having the same, and method for manufacturing the image sensor |
| KR101643610B1 (en) | 2010-01-18 | 2016-07-29 | 삼성전자주식회사 | Method and Apparatus for digital imaging process |
| KR101688523B1 (en) | 2010-02-24 | 2016-12-21 | 삼성전자주식회사 | Stack-type image sensor |
| JP5552858B2 (en) | 2010-03-26 | 2014-07-16 | ソニー株式会社 | Solid-state imaging device, driving method of solid-state imaging device, and electronic apparatus |
| US20110317048A1 (en) * | 2010-06-29 | 2011-12-29 | Aptina Imaging Corporation | Image sensor with dual layer photodiode structure |
| JP5521854B2 (en) | 2010-07-26 | 2014-06-18 | コニカミノルタ株式会社 | Imaging device and image input device |
| JP5676988B2 (en) | 2010-09-14 | 2015-02-25 | キヤノン株式会社 | Focus adjustment device |
| JP2012195921A (en) * | 2011-02-28 | 2012-10-11 | Sony Corp | Solid-state imaging element and camera system |
| US20130113967A1 (en) * | 2011-11-04 | 2013-05-09 | Honeywell International Inc. Doing Business As (D.B.A.) Honeywell Scanning & Mobility | Apparatus comprising image sensor array having global shutter shared by a plurality of pixels |
| JP5556823B2 (en) | 2012-01-13 | 2014-07-23 | 株式会社ニコン | Solid-state imaging device and electronic camera |
| US8569700B2 (en) | 2012-03-06 | 2013-10-29 | Omnivision Technologies, Inc. | Image sensor for two-dimensional and three-dimensional image capture |
-
2013
- 2013-03-29 WO PCT/JP2013/059621 patent/WO2013147198A1/en not_active Ceased
- 2013-03-29 JP JP2014508109A patent/JP6264284B2/en active Active
- 2013-03-29 EP EP13767811.6A patent/EP2835965B1/en active Active
- 2013-03-29 CN CN202010418828.4A patent/CN111711746A/en active Pending
- 2013-03-29 CN CN201380017868.6A patent/CN104205808B/en active Active
- 2013-03-29 US US14/389,626 patent/US9826183B2/en active Active
-
2017
- 2017-09-06 JP JP2017171144A patent/JP6536642B2/en active Active
- 2017-10-26 US US15/794,710 patent/US10389959B2/en active Active
-
2019
- 2019-05-02 US US16/401,436 patent/US20190260951A1/en not_active Abandoned
- 2019-06-06 JP JP2019106431A patent/JP7001080B2/en active Active
-
2021
- 2021-07-28 US US17/387,390 patent/US20210358981A1/en not_active Abandoned
Patent Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20100091161A1 (en) * | 2007-06-16 | 2010-04-15 | Nikon Corporation | Solid-state image sensor and imaging apparatus equipped with solid-state image sensor |
| US20100141771A1 (en) * | 2007-07-09 | 2010-06-10 | Boly Media Communications (Shenzhen) Co., Ltd | Multi-Spectrum Sensing Device And Manufacturing Methods Thereof |
| US20100282945A1 (en) * | 2009-05-11 | 2010-11-11 | Sony Corporation | Two-dimensional solid-state image capture device and polarization-light data processing method therefor |
| US20120193515A1 (en) * | 2011-01-28 | 2012-08-02 | Gennadiy Agranov | Imagers with depth sensing capabilities |
| US9826183B2 (en) * | 2012-03-30 | 2017-11-21 | Nikon Corporation | Image-capturing device and image sensor |
Also Published As
| Publication number | Publication date |
|---|---|
| JP2018029342A (en) | 2018-02-22 |
| US9826183B2 (en) | 2017-11-21 |
| EP2835965A4 (en) | 2015-12-09 |
| EP2835965A1 (en) | 2015-02-11 |
| JP7001080B2 (en) | 2022-01-19 |
| US10389959B2 (en) | 2019-08-20 |
| JPWO2013147198A1 (en) | 2015-12-14 |
| CN104205808A (en) | 2014-12-10 |
| CN104205808B (en) | 2020-06-05 |
| WO2013147198A1 (en) | 2013-10-03 |
| JP6536642B2 (en) | 2019-07-03 |
| EP2835965B1 (en) | 2017-05-03 |
| JP2019193280A (en) | 2019-10-31 |
| US20210358981A1 (en) | 2021-11-18 |
| US20150222833A1 (en) | 2015-08-06 |
| US20180048837A1 (en) | 2018-02-15 |
| JP6264284B2 (en) | 2018-01-24 |
| CN111711746A (en) | 2020-09-25 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US10389959B2 (en) | Image-capturing device and image sensor | |
| US12294795B2 (en) | Image sensor and image-capturing device | |
| US8063978B2 (en) | Image pickup device, focus detection device, image pickup apparatus, method for manufacturing image pickup device, method for manufacturing focus detection device, and method for manufacturing image pickup apparatus | |
| JP4421793B2 (en) | Digital camera | |
| US20040090550A1 (en) | Image sensing means for digital camera and digital camera adopting the same | |
| JP2008177903A (en) | Imaging device | |
| JP4858179B2 (en) | Focus detection apparatus and imaging apparatus | |
| JP2018046563A (en) | Image sensor |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
| STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |