WO2006075463A1 - 撮像装置及び撮像結果の処理方法 - Google Patents
撮像装置及び撮像結果の処理方法 Download PDFInfo
- Publication number
- WO2006075463A1 WO2006075463A1 PCT/JP2005/022603 JP2005022603W WO2006075463A1 WO 2006075463 A1 WO2006075463 A1 WO 2006075463A1 JP 2005022603 W JP2005022603 W JP 2005022603W WO 2006075463 A1 WO2006075463 A1 WO 2006075463A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- imaging
- block
- result
- motion
- exposure time
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/68—Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/68—Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
- H04N23/681—Motion detection
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/68—Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
- H04N23/681—Motion detection
- H04N23/6811—Motion detection based on the image signal
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/71—Circuitry for evaluating the brightness variation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/72—Combination of two or more compensation controls
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/73—Circuitry for compensating brightness variation in the scene by influencing the exposure time
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/40—Extracting pixel data from image sensors by controlling scanning circuits, e.g. by modifying the number of pixels sampled or to be sampled
- H04N25/44—Extracting pixel data from image sensors by controlling scanning circuits, e.g. by modifying the number of pixels sampled or to be sampled by partially reading an SSIS array
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/50—Control of the SSIS exposure
- H04N25/51—Control of the gain
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/50—Control of the SSIS exposure
- H04N25/53—Control of the integration time
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/50—Control of the SSIS exposure
- H04N25/53—Control of the integration time
- H04N25/533—Control of the integration time by using differing integration times for different sensor regions
- H04N25/535—Control of the integration time by using differing integration times for different sensor regions by dynamic region selection
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/70—SSIS architectures; Circuits associated therewith
- H04N25/79—Arrangements of circuitry being divided between different or multiple substrates, chips or circuit boards, e.g. stacked image sensors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/14—Picture signal circuitry for video frequency region
- H04N5/144—Movement detection
- H04N5/145—Movement estimation
-
- H—ELECTRICITY
- H10—SEMICONDUCTOR DEVICES; ELECTRIC SOLID-STATE DEVICES NOT OTHERWISE PROVIDED FOR
- H10F—INORGANIC SEMICONDUCTOR DEVICES SENSITIVE TO INFRARED RADIATION, LIGHT, ELECTROMAGNETIC RADIATION OF SHORTER WAVELENGTH OR CORPUSCULAR RADIATION
- H10F39/00—Integrated devices, or assemblies of multiple devices, comprising at least one element covered by group H10F30/00, e.g. radiation detectors comprising photodiode arrays
- H10F39/80—Constructional details of image sensors
- H10F39/802—Geometry or disposition of elements in pixels, e.g. address-lines or gate electrodes
- H10F39/8027—Geometry of the photosensitive area
-
- H—ELECTRICITY
- H10—SEMICONDUCTOR DEVICES; ELECTRIC SOLID-STATE DEVICES NOT OTHERWISE PROVIDED FOR
- H10F—INORGANIC SEMICONDUCTOR DEVICES SENSITIVE TO INFRARED RADIATION, LIGHT, ELECTROMAGNETIC RADIATION OF SHORTER WAVELENGTH OR CORPUSCULAR RADIATION
- H10F39/00—Integrated devices, or assemblies of multiple devices, comprising at least one element covered by group H10F30/00, e.g. radiation detectors comprising photodiode arrays
- H10F39/80—Constructional details of image sensors
- H10F39/805—Coatings
-
- H—ELECTRICITY
- H10—SEMICONDUCTOR DEVICES; ELECTRIC SOLID-STATE DEVICES NOT OTHERWISE PROVIDED FOR
- H10F—INORGANIC SEMICONDUCTOR DEVICES SENSITIVE TO INFRARED RADIATION, LIGHT, ELECTROMAGNETIC RADIATION OF SHORTER WAVELENGTH OR CORPUSCULAR RADIATION
- H10F39/00—Integrated devices, or assemblies of multiple devices, comprising at least one element covered by group H10F30/00, e.g. radiation detectors comprising photodiode arrays
- H10F39/80—Constructional details of image sensors
- H10F39/805—Coatings
- H10F39/8053—Colour filters
-
- H—ELECTRICITY
- H10—SEMICONDUCTOR DEVICES; ELECTRIC SOLID-STATE DEVICES NOT OTHERWISE PROVIDED FOR
- H10F—INORGANIC SEMICONDUCTOR DEVICES SENSITIVE TO INFRARED RADIATION, LIGHT, ELECTROMAGNETIC RADIATION OF SHORTER WAVELENGTH OR CORPUSCULAR RADIATION
- H10F39/00—Integrated devices, or assemblies of multiple devices, comprising at least one element covered by group H10F30/00, e.g. radiation detectors comprising photodiode arrays
- H10F39/80—Constructional details of image sensors
- H10F39/805—Coatings
- H10F39/8057—Optical shielding
-
- H—ELECTRICITY
- H10—SEMICONDUCTOR DEVICES; ELECTRIC SOLID-STATE DEVICES NOT OTHERWISE PROVIDED FOR
- H10F—INORGANIC SEMICONDUCTOR DEVICES SENSITIVE TO INFRARED RADIATION, LIGHT, ELECTROMAGNETIC RADIATION OF SHORTER WAVELENGTH OR CORPUSCULAR RADIATION
- H10F39/00—Integrated devices, or assemblies of multiple devices, comprising at least one element covered by group H10F30/00, e.g. radiation detectors comprising photodiode arrays
- H10F39/80—Constructional details of image sensors
- H10F39/806—Optical elements or arrangements associated with the image sensors
- H10F39/8063—Microlenses
-
- H—ELECTRICITY
- H10—SEMICONDUCTOR DEVICES; ELECTRIC SOLID-STATE DEVICES NOT OTHERWISE PROVIDED FOR
- H10F—INORGANIC SEMICONDUCTOR DEVICES SENSITIVE TO INFRARED RADIATION, LIGHT, ELECTROMAGNETIC RADIATION OF SHORTER WAVELENGTH OR CORPUSCULAR RADIATION
- H10F39/00—Integrated devices, or assemblies of multiple devices, comprising at least one element covered by group H10F30/00, e.g. radiation detectors comprising photodiode arrays
- H10F39/80—Constructional details of image sensors
- H10F39/809—Constructional details of image sensors of hybrid image sensors
-
- H—ELECTRICITY
- H10—SEMICONDUCTOR DEVICES; ELECTRIC SOLID-STATE DEVICES NOT OTHERWISE PROVIDED FOR
- H10F—INORGANIC SEMICONDUCTOR DEVICES SENSITIVE TO INFRARED RADIATION, LIGHT, ELECTROMAGNETIC RADIATION OF SHORTER WAVELENGTH OR CORPUSCULAR RADIATION
- H10F39/00—Integrated devices, or assemblies of multiple devices, comprising at least one element covered by group H10F30/00, e.g. radiation detectors comprising photodiode arrays
- H10F39/80—Constructional details of image sensors
- H10F39/811—Interconnections
Definitions
- the present invention relates to an imaging apparatus and a method of processing imaging results, and can be applied to an imaging apparatus using, for example, a CMOS solid-state imaging device.
- the present invention divides one screen into a plurality of blocks, detects the movement for each block, and controls the exposure time, thereby further improving the image quality compared to the conventional technology with respect to the control of the exposure time by the electronic shutter. To be able to improve. Background art
- an imaging device is configured to control the exposure time by controlling the charge accumulation time of the image sensor with a so-called electronic shutter, so that even when shooting a fast-moving subject, an imaging result without blurring can be obtained.
- a so-called electronic shutter for example, in Japanese Patent Laid-Open No. 6-165 047 and Japanese Patent Laid-Open No. 6-2 6 1 2 5 6, the movement of a subject is detected by a system controller, A method has been proposed in which the exposure time is controlled based on the motion detection result, and gain control of AGC (Auto Gain Control) and control by ALC (Auto Light Control) are executed.
- control of exposure time using a conventional electronic shutter is the same as the background when the exposure time is varied by moving part of the screen by varying the exposure time for all the pixels that make up one screen. Control is also performed so that the exposure time is shortened for the stationary part. As a result, in such a case, there was a problem that the S / N ratio deteriorates for such a stationary part, and the image quality deteriorates accordingly.
- the present invention has been made in consideration of the above points, and intends to propose an imaging apparatus and an imaging result processing method capable of further improving the image quality as compared with the related art with regard to the control of the exposure time by the electronic shutter. Is.
- the present invention is applied to an imaging device, an imaging unit is formed by arranging pixels in a matrix shape, and an imaging result of an optical image formed on the imaging surface is output.
- the motion detection unit detects the amount of motion in the imaging result, and the motion is large based on the detection result by the motion detection unit.
- the control means for controlling the exposure time of the image pickup means for each block so that the exposure time becomes shorter for each block, and the signal level of the image pickup result that changes by the control of the exposure time for each block.
- Signal level correction means for correcting and outputting the bell is provided.
- the imaging device is applied to an imaging device, pixels are arranged in a matrix shape, an imaging surface is formed, and an imaging result of an optical image formed on the imaging surface is output, and the imaging For each block obtained by dividing the surface in the horizontal direction and / or vertical direction, a motion detection unit that detects a motion amount in the imaging result, and a block with a larger motion based on a detection result by the motion detection unit, exposure time If the control means for controlling the exposure time of the imaging means is provided for each block, the exposure time can be controlled in accordance with the movement for each block, thereby As for the background, the exposure time can be lengthened to prevent the deterioration of the SZN ratio, while the moving part can be shortened to prevent blurring.
- the exposure time is controlled for each block in this way by providing, for each block, signal level correction means for correcting and outputting the signal level of the imaging result that changes by controlling the exposure time. Therefore, it is possible to prevent the signal level from changing between the blocks that change.
- the present invention is applied to a method of processing an imaging result, and the imaging surface of the imaging unit is
- the imaging means for each block A control step for controlling the exposure time, and a signal level correction step for correcting and outputting the signal level of the imaging result that changes due to the control of the exposure time for each block.
- the configuration of the present invention it is possible to provide a method of processing an imaging result that can further improve the image quality as compared with the related art with respect to the control of the exposure time by the electronic shutter.
- the image quality can be further improved as compared with the related art with regard to the control of the exposure time by the electronic shutter.
- FIG. 1 is a block diagram showing the overall configuration of an integrated circuit according to an embodiment of the present invention.
- FIG. 2 is a cross-sectional view of a sensor chip of the integrated circuit of FIG.
- FIG. 3 is a side view for explaining the laminated structure of the sensor chip and the logic chip.
- FIG. 4 is a block diagram showing a detailed configuration of the integrated circuit of FIG.
- FIG. 5 is a timing chart for explaining the operation of the pixel block of FIG.
- FIG. 6 is a timing chart for explaining the control of the exposure time of each pixel.
- FIG. 7 is a conceptual diagram for explaining the control of the exposure time of each block.
- FIG. 8 is a block diagram showing an image pickup apparatus according to Embodiment 2 of the present invention.
- FIG. 9 is a block diagram showing an image pickup apparatus according to Embodiment 3 of the present invention.
- BEST MODE FOR CARRYING OUT THE INVENTION Hereinafter, embodiments of the present invention will be described in detail with reference to the drawings as appropriate.
- FIG. 1 is a block diagram showing an integrated circuit applied to the image pickup apparatus according to Embodiment 1 of the present invention.
- the imaging apparatus according to this embodiment forms a desired optical image on the imaging surface of the integrated circuit 1 by automatic aperture control and automatic focus control by a controller using a lens (not shown), and outputs an imaging result of this optical image. To do.
- the integrated circuit 1 is an integrated circuit using an image sensor, and is formed by packaging a stacked body of the sensor chip 2 and the logic chip 3.
- the sensor chip 2 is a semiconductor chip of an imaging element that outputs an imaging result by the XY addressing method.
- a CMOS solid-state imaging element is applied to this imaging element.
- the sensor chip 2 is formed by an imaging unit 4 and a control unit 5 that controls the operation of the imaging unit 4.
- the imaging unit 4 arranges pixels in a matrix form to form an imaging surface, and outputs an imaging result of an optical image formed on the imaging surface.
- the imaging unit 4 is formed by equally dividing the pixels arranged in the matrix shape in the horizontal direction and the vertical direction, and thereby dividing the imaging surface into a plurality of pixel blocks 6.
- Each pixel block 6 is provided with peripheral circuits related to the pixels belonging to each pixel block 6 and is configured to output an imaging result.
- the sensor chip 2 captures images using the plurality of pixel blocks 6. Output results in parallel.
- the sensor chip 2 has an element layer 12 formed of a silicon (S i) layer having a thickness of about 10 to 20 [m].
- the light receiving element 13 is formed.
- the lenses 18 are stacked, thereby forming the pixels 19.
- the peripheral circuit of each pixel block 6, the control unit 5, and the wiring layer 20 for wiring them are formed in the lower layer of the element layer 12, and the substrate that holds the whole on the lower layer side of the wiring layer 20. Support material 21 is provided.
- the wiring layer 20 is formed so as to eliminate various restrictions in the case of providing the wiring layer 20 on the imaging surface side and to greatly improve the degree of freedom of wiring.
- the sensor chip 2 has the wiring layer 20 formed on the side opposite to the imaging surface in this manner, so that a thin semiconductor substrate is processed from the wiring layer 20 side to receive the light receiving element 1 3 and peripheral circuits.
- the wiring layer 20 and the substrate support material 21 are sequentially formed on this semiconductor substrate, and then the semiconductor substrate is turned over and polished by CMP to complete the element layer 12, and the light shielding film 15.
- SiN silicon nitride film
- color filter 17 and microlens 18 sequentially.
- the sensor chip 2 has a logic chip 3 assigned to the substrate support material 21, and is formed on the fine protruding electrode 3 1 formed on the wiring layer 20 side and the logic chip 3.
- the fine projecting electrodes 32 are electrically connected to the logic chip 3 and held.
- the fine protruding electrode 32 is hereinafter referred to as a micro bump. .
- the logic chip 3 is an integrated circuit including a signal processing circuit that processes the imaging result of the sensor chip 2, and in this embodiment, the signal processing circuit determines the exposure time of the imaging unit 4 provided in the sensor chip 2.
- the pulse generation unit 4 1 includes each pixel block 6 of the sensor chip 2.
- the pulse generation block 4 3 is provided corresponding to each of the pulse generation blocks 4 3, and each pulse generation block 4 3 is connected to the corresponding pixel block 6 by the micro bumps 3 1 and 3 2, respectively. Control each time.
- the integrated circuit 1 is formed so as to simultaneously output the imaging results of the plurality of pixel blocks 6 in parallel, and the exposure time of each pixel block 6 can be individually controlled by each pulse generation block 43. Formed as follows.
- FIG. 4 is a block diagram showing the pixel block 6 and the pulse generation block 43.
- the pixels 19 A and 19 B that are continuous in the horizontal direction are connected to the reset control circuit 52 by a common gate line, and the pixels in the vertical direction are also connected.
- the continuous pixels 19 A and 19 B are connected by a common signal line and connected to the horizontal drive control circuit 53.
- Each pixel 19A, 19B photoelectrically converts the incident light to accumulate the accumulated charge, and discharges the accumulated charge by the reset pulses RS Tl and RST2 output from the reset control circuit 52, and the reset control circuit 52 Readout pulses output from the ROUT 1 and ROUT 2 convert the accumulated charge to voltage and output it to the signal line.
- the pixel block 6 changes the exposure time by varying the charge accumulation time under the control of the reset control circuit 52, and outputs the imaging result to the horizontal drive control circuit 53 in line units.
- the reset control circuit 52 resets the operation with the vertical synchronization signal VD output from the control unit 5, and then outputs the reset control circuit 52 from the pulse generation block 43.
- the reset pulse RST 1 is raised at the rising timing of the shatter pulse SHT (Fig. 5 (B) and Fig. 6 (B)) (Fig. 5 (C 1) and Fig. 6 (C 1)).
- the readout pulse ROUT 1 is raised at the timing of the falling edge of the shatter pulse SHT (Fig. 5 (D1) and Fig. 6 (D1)).
- the reset control circuit 52 outputs a reset pulse R ST 1 and a readout pulse ROUT 1 based on the shirt pulse SHT to the pixel 19 A on the first line.
- the reset pulse RST 1 and the readout pulse ROUT 1 are sequentially delayed by a delay time ⁇ t corresponding to the number of lines forming the pixel block 6 to sequentially reset the reset pulses RST related to the pixels 19 B,. 2, ..., RSTn and readout pulse ROUT 2, ..., ROUTn are generated sequentially (Fig. 5 (C 2), (D 2) and Fig. 6 (C 2), (D 2), (C n), (Dn)), these reset pulse RST 2 and readout pulse ROUT 2,... are output to pixels 19 B,.
- the reset control circuit 52 sets the period T 1 and T 2 until the falling after the shot tap pulse SHT rises to be the exposure time of the pixels 19A and 19B, and each pixel 19A, 1 9 A, 1 so that the accumulated charge accumulated in 9 B (Fig. 5 (E 1) and (E 2)) is output to the signal line in line units as the imaging result at the end of the exposure time.
- the reset control circuit 52 outputs the gain control signal G of the amplifier circuit 54 in accordance with the exposure time by the shatter pulse SHT, thereby changing the imaging result signal by controlling the exposure time by the shatter pulse SHT. Correct the level.
- the imaging results of the pixels 19 A, 19 B,... Consecutive in the vertical direction are sequentially output to each signal line by time division (FIG. 6 (E)).
- the horizontal drive control circuit 53 outputs the imaging results output from the pixels 19 A and 19 B to the signal lines in a line unit in a time-division multiplexed manner (FIG. 6 (F)).
- the amplification circuit 54 varies the gain by the gain control signal G output from the reset control circuit 52, and amplifies and outputs the imaging result output from the horizontal drive control circuit 53.
- the integrated circuit 1 corrects and outputs the signal level of the imaging result that changes by controlling the exposure time.
- the analog-digital conversion circuit (ADC) 55 performs analog-digital conversion processing on the output signal of the amplification circuit 54, and outputs image data OU T as a result of the processing to the corresponding pulse generation block 43 of the logic chip 3. Accordingly, in this embodiment, the reset control circuit 52, the horizontal drive control circuit 53, the amplifier circuit 54, and the analog digital conversion circuit 55 can be used to surround the pixels 19A and 19B belonging to each pixel block 6. A circuit is formed.
- the integrated circuit 1 outputs the imaging result OUT of each pixel 19 A, 19 B from each pixel block 6 in the order of raster scanning, and simultaneously outputs the imaging result from each pixel block 6 in parallel.
- OUT is output, and the exposure time by the electronic shutter is individually varied in each pixel block 6 by the control of the corresponding pulse generation block 43, and the signal level at the imaging result OUT by varying the speed of this exposure time It is configured so that each pixel block 6 individually corrects the change of.
- the pulse generation block 43 detects the motion amount of the imaging result output from the corresponding pixel block 6 by the motion amount detection circuit 61, and varies the shutter pulse SHT according to the motion amount detection result. Output.
- the averaging circuit 6 2 The imaging results output from Block 6 are averaged in units of frames. That is, the averaging circuit 62 cumulatively adds the imaging results output from the corresponding pixel block 6 by the adder circuit 63 and the delay circuit 64, and controls the switch circuit 65 provided in the output stage of the adder circuit 63. Thus, the addition result is initialized by outputting the addition result to the frame difference calculation circuit 66 for each frame. As a result, the motion detection circuit 61 calculates the average value of the imaging results of each pixel block 6 by accumulatively adding the imaging results in units of frames.
- the frame difference calculation circuit 66 delays the output value of the averaging circuit 62 by a delay circuit (1 V delay) 67 for one frame period.
- the frame difference calculation circuit 66 subtracts the output value of the delay circuit 67 from the output value of the averaging circuit 62 by the subtraction circuit 68, thereby calculating the inter-frame difference value ⁇ S.
- the motion amount detection circuit 61 approximately calculates the motion amount based on the inter-frame difference value AS.
- the shutter pulse generation circuit 69 generates a shatter pulse SHT according to the motion amount detection result by the motion amount detection circuit 6 1 obtained in this way.
- the shutter pulse generation circuit 69 increases the inter-frame difference value ⁇ S output from the motion amount detection circuit 61 so that the period during which the shotta pulse SHT rises is shortened accordingly. Generate pulse SHT. If the movement increases as a result of this, a shirt pulse SHT is generated so that the exposure time is reduced accordingly.
- the shotta pulse generation circuit 69 delays the rise timing of the shottapulse SHT, and this rise The timing of the fall of the shirt tapulse SHT is advanced by the amount of delay of the timing of.
- the shatter pulse generation circuit 69 is set so as not to change at the center timing of the exposure time in each pixel block 6.
- the reset control circuit 52 generates the reset pulse RST 1 and the read pulse ROUT 1 by the shatter pulse SHT, and the reset pulse RST 1 and the read pulse.
- the pulse ROUT 1 is sequentially delayed to generate reset pulses R ST 2,... And read pulses ROUT 2,..., And these reset pulses are generated according to the first half and the second half of the vertical synchronization period. It is configured to output RST 1, RST 2,... And read pulse ROUT 1, ROUT 2,.
- the integrated circuit 1 accumulates the imaging results OUT output from the pixel blocks 6 in parallel in this way in the image memory provided in the sensor chip 2, and multiplexes them in the order of raster scanning. Output by one system.
- each of the pixels 19 A, 19 B, the reset control circuit 52, and the horizontal drive control circuit 53 constitute an imaging unit that outputs the imaging result of the optical image formed on the imaging surface
- the motion amount detection circuit 61 constitutes motion detection means for detecting the motion amount in the imaging result for each block obtained by dividing the imaging surface in the horizontal direction and / or the vertical direction.
- the shirt pulse generation circuit 69 constitutes a control means for controlling the exposure time of the imaging means for each block so that the exposure time is shortened for the block with the larger motion based on the detection result by the motion detection means.
- the reset control circuit 52 and the amplification circuit 54 constitute signal level correction means that corrects and outputs the signal level of the imaging result that changes by controlling the exposure time for each block.
- this imaging apparatus In the above configuration, in this imaging apparatus (FIG. 1), an optical image is formed on the imaging surface of the integrated circuit 1 via the lens, and the imaging result of this optical image is acquired by the integrated circuit 1. Is output.
- the pixels 19A and 19B by the light receiving elements 13 are arranged in a matrix to form an imaging surface.
- the imaging result OUT is obtained by the pixel block 6 equally divided in the vertical and vertical directions.
- each pixel block 6 read pulses ROUT 1, ROUT 2,... Are sequentially output from the reset control circuit 52 on each line, and each pixel 19 is output by this read pulse ROUT 1, ROUT 2,.
- the imaging results out 1, out 2,... By A and 19B are input to the horizontal drive control circuit 53 in line units, where they are time-division multiplexed. Further, after the signal level of the imaging result is corrected by the subsequent amplifier circuit 54, it is converted into a digital signal by the analog-digital conversion circuit 55 and output. As a result, an imaging result OUT is acquired for each pixel block 6 obtained by dividing the imaging surface, and the imaging result OUT is output simultaneously in parallel.
- the imaging results OUT of each pixel block 6 are input to the corresponding pulse generation block 43, where the amount of motion ⁇ S of each pixel block 6 is detected, and the pixel block with the larger motion is detected by this amount of motion AS.
- the charge storage time of the corresponding pixel block 6 is controlled so as to shorten the exposure time.
- the exposure time can be controlled according to the movement for each block that divides one screen, and for stationary backgrounds, the exposure time is lengthened to prevent deterioration of the SZN ratio.
- the exposure time can be shortened to prevent blurring.
- by detecting the movement for each block it is possible to detect the movement of even a small object, thereby preventing the blur of such a small object.
- time control the image quality is further improved compared to the conventional method.
- the exposure time is controlled in this way, and the gain of the amplification circuit 54 is increased and the signal level of the imaging result OUT is corrected by the amount corresponding to the shortened exposure time.
- the exposure time is controlled for each pixel block, and this imaging apparatus acquires the imaging result OUT for each pixel block 6 and outputs it simultaneously in parallel.
- the pulse generation block 43 that detects the movement amount ⁇ S of each pixel block 6 controls the exposure time.
- the imaging device can control the exposure time for each block by simultaneous parallel processing in a plurality of blocks independently of each block. Be improved.
- the sensor chip 2 and the logic chip 3 are connected and stacked by the microphone opening bumps 3 1 and 3 2, and the imaging result OUT is acquired by the sensor chip 2, and the logic chip 3 performs exposure.
- the sensor chip is configured to control the exposure time for each block by simultaneous parallel processing in a plurality of blocks as described above. By applying it to the laminated structure of 2 and logic chip 3, the overall configuration can be simplified.
- peripheral circuits such as a reset control circuit 52 and a horizontal drive control circuit 53 are arranged on the opposite side of the imaging surface, and the logic chip 3 is laminated on the opposite side.
- a reset control circuit 52 and a horizontal drive control circuit 53 are arranged on the opposite side of the imaging surface, and the logic chip 3 is laminated on the opposite side.
- the motion detection is performed to control the exposure time
- the averaging circuit 62 accumulates the imaging result OUT in units of frames and performs imaging in each frame.
- the frame difference calculation circuit 66 calculates the interframe difference of the average value to obtain the motion amount ⁇ S.
- the shutter pulse SHT is generated by the shatter pulse generation circuit 69 based on the amount of movement AS, and the reset pulse RST 1, RST 2,... Is generated by the reset control circuit 52 by the shot pulse SHT. reading
- the exposure time is varied by changing the timing of the output pulses ROUT 1, ROUT 2,.
- the amount of movement ⁇ S is calculated by a simple process for obtaining a frame difference, and the exposure control process is executed, and accordingly, the exposure control process is executed for each block by a simple configuration. It becomes possible.
- one screen is divided into a plurality of blocks, the motion is detected for each block, and the exposure time is controlled.
- the image quality can be further improved.
- the imaging results of each block are output simultaneously in parallel, and the imaging results of each block are processed separately, and the amount of motion is detected for each block.
- the exposure time can be controlled by independent and parallel processing for each block, and the image quality can be further improved compared to the conventional method with regard to the exposure time control by the electronic shutter. Can do.
- the pixels that are the imaging means, the peripheral circuits of the pixels, and the configuration of the motion detection means are formed on different semiconductor chips and connected by micro bumps, so that the entire configuration is integrated to simplify the configuration. Therefore, it is possible to prevent the aperture ratio of the pixel from being lowered and to ensure a sufficient degree of design freedom.
- the imaging apparatus 71 is provided with an encoding means 72 and a recording means 73 in addition to the configuration of the integrated circuit 1 described above for the embodiment 1.
- the imaging result OUT is encoded by the encoding means 7 2 and the recording means 7 3 and recorded on the recording medium.
- the imaging apparatus 71 according to the present embodiment forms the imaging system 74 by the integrated circuit 1 described above for the first embodiment, and after multiplexing the imaging result OUT output from the imaging system 74, The luminance signal and the color difference signal are converted and input to the encoding unit 72, and the data encoded by the encoding unit 72 is recorded. 7 Record on the recording medium according to 3.
- the encoding means 72 is an encoding means related to data compression using a motion vector.
- the imaging result OUT is subjected to an encoding process by the method of H.264.
- the number of pixels constituting the pixel block 6 is set so as to correspond to the macroblock used for detection of the motion vector by the encoding means 72.
- the pixel block may be set so as to correspond to either the luminance signal or the color difference signal, and in H. 2 6 4 motion detection is performed using a plurality of types of macro blocks of different sizes. Accordingly, the pixel block may be set so as to correspond to any macro block of the plurality of types of macro blocks.
- the imaging system 74 sets the pixel block 6 in this way, and outputs the motion amount ⁇ S detected by each pixel block 6 to the sign means 72.
- the encoding means 72 determines the presence or absence of motion by determining the amount of motion ⁇ S based on a predetermined threshold value. If it is determined that there is motion based on this determination result, the motion vector is detected by normal processing and the imaging result OUT is encoded. On the other hand, if it is determined that there is no motion based on this determination result, the motion vector detection is stopped, the motion vector is set to 0, and the imaging result OU T is encoded.
- the motion detection is performed for each block and the exposure time is controlled, and the motion detection result used for the control of the exposure time is used for the encoding process. Reduce the burden.
- the imaging result is blocked according to the size corresponding to the motion vector detection unit in the encoding process, the motion is detected for each block, and the exposure time is controlled.
- the image pickup apparatus 81 includes a motion detection circuit 61 (see FIG. 4) of the image pickup system 74 described above for the second embodiment.
- the amount of motion is detected by detecting the motion vector. That is, the motion vector V is detected, the motion vector is converted into an absolute value, the length of the motion vector V is detected, and the magnitude of the motion is detected based on the length of the motion vector V.
- various detection methods such as a block matching method and a gradient method can be applied.
- the encoding means 82 stops the motion vector detection process when detecting the motion vector V for the block having a size corresponding to the pixel block 6, and detects the motion detected by the motion detection circuit 61. Encodes using vector V.
- the imaging device 81 according to this embodiment is configured in the same manner as the imaging device 71 described above with respect to the second embodiment except for the configuration related to the motion vector V.
- the exposure time is controlled by varying the timing of both the reset pulse and the readout pulse.
- the present invention is not limited to this, and exposure is performed by varying the timing of only the reset pulse. You may make it control time.
- the present invention is not limited to this, and the imaging results for one screen are output together. Then, this imaging result may be classified and processed for each pixel block, and the exposure time may be controlled for each pixel block. In this way, with respect to the output of the imaging results, even when using a conventional imaging device that outputs the imaging results in the order of raster scanning, the exposure time by the electronic shutter is the same as in the above-described embodiment. With this control, the image quality can be further improved compared to the conventional method.
- the case where the sensor chip is configured by the CMOS solid-state image sensor has been described.
- the present invention is not limited to this, and the case where the sensor chip is configured by various solid-state image sensors based on the XY address system.
- it can be widely applied to the case where a sensor chip is constituted by a CCD solid-state imaging device.
- the case where the pixel block is formed by dividing the imaging surface in the horizontal direction and the vertical direction has been described.
- the present invention is not limited to this, and only one of the horizontal direction and the vertical direction is described.
- the imaging surface may be divided into pixel blocks.
- the present invention is not limited to this, and the front surface side of the sensor chip 2
- the present invention can be widely applied to the case where the peripheral circuit of each pixel is provided, and the logic chip is formed separately.
- the present invention can be applied to, for example, an imaging apparatus using a CMOS solid state imaging device.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Transforming Light Signals Into Electric Signals (AREA)
- Studio Devices (AREA)
- Solid State Image Pick-Up Elements (AREA)
- Facsimile Image Signal Circuits (AREA)
Abstract
Description
Claims
Priority Applications (13)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US11/794,996 US8345108B2 (en) | 2005-01-13 | 2005-12-02 | Image pickup apparatus and processing method for result of image pickup |
| CA002593719A CA2593719A1 (en) | 2005-01-13 | 2005-12-02 | Image pickup device and image pickup result processing method |
| KR1020137026191A KR101362604B1 (ko) | 2005-01-13 | 2005-12-02 | 촬상 장치 및 촬상 결과의 처리 방법 |
| AU2005324643A AU2005324643B2 (en) | 2005-01-13 | 2005-12-02 | Image pickup device and image pickup result processing method |
| KR1020127031162A KR20120137443A (ko) | 2005-01-13 | 2005-12-02 | 촬상 장치 및 촬상 결과의 처리 방법 |
| BRPI0519834-8A BRPI0519834A2 (pt) | 2005-01-13 | 2005-12-02 | aparelho de captação de imagem, e, método de processamento para um resultado de captação de imagem |
| EP05814419.7A EP1838088B1 (en) | 2005-01-13 | 2005-12-02 | Image pickup device and image pickup result processing method |
| MX2007008426A MX2007008426A (es) | 2005-01-13 | 2005-12-02 | Dispositivo de recuperacion de imagenes y metodo de procesamiento de resultado de recuperacion de imagenes. |
| US13/690,999 US8687069B2 (en) | 2005-01-13 | 2012-11-30 | Image pickup apparatus and processing method for result of image pickup |
| US14/159,616 US9185304B2 (en) | 2005-01-13 | 2014-01-21 | Image pickup apparatus and processing method for result of image pickup |
| US14/750,088 US20150296113A1 (en) | 2005-01-13 | 2015-06-25 | Image pickup apparatus and processing method for result of image pickup |
| US15/047,467 US9729764B2 (en) | 2005-01-13 | 2016-02-18 | Image pickup apparatus and processing method for result of image pickup |
| US15/590,909 US10057515B2 (en) | 2005-01-13 | 2017-05-09 | Image pickup apparatus and processing method for result of image pickup |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2005006178A JP4277216B2 (ja) | 2005-01-13 | 2005-01-13 | 撮像装置及び撮像結果の処理方法 |
| JP2005-006178 | 2005-03-03 |
Related Child Applications (2)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US11/794,996 A-371-Of-International US8345108B2 (en) | 2005-01-13 | 2005-12-02 | Image pickup apparatus and processing method for result of image pickup |
| US13/690,999 Division US8687069B2 (en) | 2005-01-13 | 2012-11-30 | Image pickup apparatus and processing method for result of image pickup |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2006075463A1 true WO2006075463A1 (ja) | 2006-07-20 |
Family
ID=36677495
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/JP2005/022603 Ceased WO2006075463A1 (ja) | 2005-01-13 | 2005-12-02 | 撮像装置及び撮像結果の処理方法 |
Country Status (11)
| Country | Link |
|---|---|
| US (6) | US8345108B2 (ja) |
| EP (1) | EP1838088B1 (ja) |
| JP (1) | JP4277216B2 (ja) |
| KR (3) | KR101362604B1 (ja) |
| CN (1) | CN101099381A (ja) |
| AU (1) | AU2005324643B2 (ja) |
| BR (1) | BRPI0519834A2 (ja) |
| CA (1) | CA2593719A1 (ja) |
| MX (1) | MX2007008426A (ja) |
| RU (1) | RU2343644C1 (ja) |
| WO (1) | WO2006075463A1 (ja) |
Families Citing this family (55)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP5023805B2 (ja) | 2007-05-16 | 2012-09-12 | ソニー株式会社 | 画像処理装置および画像処理方法、並びにプログラム |
| US8063942B2 (en) * | 2007-10-19 | 2011-11-22 | Qualcomm Incorporated | Motion assisted image sensor configuration |
| EP2063630A1 (en) * | 2007-11-26 | 2009-05-27 | Thomson Licensing | Video capture device with variable shutter integration time |
| JP5247397B2 (ja) * | 2008-12-05 | 2013-07-24 | キヤノン株式会社 | 撮像装置及び撮像方法 |
| ATE543215T1 (de) * | 2009-03-24 | 2012-02-15 | Sony Corp | Festkörper-abbildungsvorrichtung, ansteuerverfahren für festkörper- abbildungsvorrichtung und elektronische vorrichtung |
| CA2757571A1 (en) * | 2009-04-13 | 2010-10-21 | Showscan Digital Llc | Method and apparatus for photographing and projecting moving images |
| JP5218309B2 (ja) * | 2009-07-14 | 2013-06-26 | ソニー株式会社 | 固体撮像素子およびその制御方法、並びにカメラシステム |
| JP5521721B2 (ja) * | 2009-08-28 | 2014-06-18 | ソニー株式会社 | 撮像素子およびカメラシステム |
| JP5737921B2 (ja) * | 2010-12-13 | 2015-06-17 | キヤノン株式会社 | 固体撮像装置、撮像システム及び固体撮像装置の駆動方法 |
| RU2446471C1 (ru) * | 2010-12-23 | 2012-03-27 | Государственное образовательное учреждение высшего профессионального образования "Тамбовский государственный технический университет" ГОУ ВПО ТГТУ | Способ обнаружения движущихся объектов и определения их параметров |
| JP5622648B2 (ja) | 2011-04-14 | 2014-11-12 | 日立オートモティブシステムズ株式会社 | 画像処理装置 |
| US9013615B2 (en) * | 2011-09-21 | 2015-04-21 | Semiconductor Components Industries, Llc | Image sensor with flexible interconnect capabilities |
| RU2478241C1 (ru) * | 2011-10-03 | 2013-03-27 | Открытое акционерное общество "Научно-исследовательский институт телевидения" | Устройство формирования видеосигнала от кмоп-матрицы |
| IN2014DN08388A (ja) | 2012-03-30 | 2015-05-08 | Nikon Corp | |
| CN110177227B (zh) | 2012-03-30 | 2022-04-05 | 株式会社尼康 | 拍摄单元、拍摄装置及拍摄控制程序 |
| RU2649967C2 (ru) | 2012-05-02 | 2018-04-06 | Никон Корпорейшн | Устройство формирования изображений |
| KR101887988B1 (ko) * | 2012-07-03 | 2018-08-14 | 삼성전자 주식회사 | 이미지 센서 칩, 이의 동작 방법, 및 이를 포함하는 시스템 |
| CN110086980B (zh) * | 2013-05-31 | 2022-06-24 | 株式会社尼康 | 电子设备 |
| WO2015001646A1 (ja) * | 2013-07-04 | 2015-01-08 | 株式会社ニコン | 電子機器、電子機器の制御方法、及び制御プログラム |
| JP6231869B2 (ja) * | 2013-12-12 | 2017-11-15 | キヤノン株式会社 | 撮像装置、通信機器およびそれらの制御方法、プログラム |
| CN103744322B (zh) * | 2013-12-13 | 2016-08-17 | 青岛歌尔声学科技有限公司 | 一种控制信号生成电路 |
| JP6245474B2 (ja) | 2014-04-21 | 2017-12-13 | ソニー株式会社 | 固体撮像素子、固体撮像素子の製造方法、並びに、電子機器 |
| JP2016019076A (ja) | 2014-07-07 | 2016-02-01 | ソニー株式会社 | 撮像装置、撮像方法、およびプログラム、並びに再生装置 |
| US10284771B2 (en) | 2014-12-03 | 2019-05-07 | Nikon Corporation | Image-capturing apparatus, electronic device, and program |
| JP2016192605A (ja) * | 2015-03-30 | 2016-11-10 | 株式会社ニコン | 電子機器、記録媒体およびプログラム |
| JP6604384B2 (ja) | 2015-09-30 | 2019-11-13 | 株式会社ニコン | 撮像装置 |
| JP6516014B2 (ja) * | 2015-09-30 | 2019-05-22 | 株式会社ニコン | 撮像装置および画像処理装置 |
| US10810716B2 (en) | 2015-09-30 | 2020-10-20 | Nikon Corporation | Image capturing device, image processing device and display device for setting different exposure conditions |
| WO2017170723A1 (ja) | 2016-03-31 | 2017-10-05 | 株式会社ニコン | 撮像装置、画像処理装置、および電子機器 |
| JPWO2017170724A1 (ja) * | 2016-03-31 | 2019-02-14 | 株式会社ニコン | 撮像装置、レンズ調節装置、および電子機器 |
| JP2018022935A (ja) * | 2016-08-01 | 2018-02-08 | ソニー株式会社 | 撮像装置、および、撮像装置の制御方法 |
| KR102647268B1 (ko) * | 2016-09-16 | 2024-03-12 | 소니 세미컨덕터 솔루션즈 가부시키가이샤 | 촬상 장치 및 전자 기기 |
| JP6715463B2 (ja) * | 2016-09-30 | 2020-07-01 | パナソニックIpマネジメント株式会社 | 画像生成装置、画像生成方法、プログラムおよび記録媒体 |
| CN106412454B (zh) * | 2016-10-18 | 2019-06-04 | 南京大学 | 暗场景下基于ccd传感器实时获取清晰图像的装置和方法 |
| US11044411B2 (en) * | 2017-02-10 | 2021-06-22 | Canon Kabushiki Kaisha | Imaging element and imaging apparatus |
| JP6580111B2 (ja) * | 2017-02-10 | 2019-09-25 | キヤノン株式会社 | 撮像素子および撮像装置 |
| JP2020096211A (ja) * | 2017-03-31 | 2020-06-18 | 株式会社ニコン | 電子機器 |
| JP2020098943A (ja) * | 2017-03-31 | 2020-06-25 | 株式会社ニコン | 電子機器 |
| JP6755405B2 (ja) * | 2017-08-18 | 2020-09-16 | 富士フイルム株式会社 | 撮像装置、撮像装置の制御方法、及び撮像装置の制御プログラム |
| EP3462731B1 (en) * | 2017-09-29 | 2021-11-10 | Canon Kabushiki Kaisha | Imaging device, imaging system, and moving body |
| WO2019082831A1 (ja) | 2017-10-27 | 2019-05-02 | キヤノン株式会社 | 撮像装置、撮像装置の制御方法、および、プログラム |
| JP7286294B2 (ja) | 2017-10-27 | 2023-06-05 | キヤノン株式会社 | 撮像装置、撮像装置の制御方法、および、プログラム |
| TW202406328A (zh) * | 2018-03-15 | 2024-02-01 | 日商尼康股份有限公司 | 感測器以及具備其之攝像裝置 |
| JP7299680B2 (ja) * | 2018-08-23 | 2023-06-28 | キヤノン株式会社 | 撮像装置及び撮像システム |
| JP7281897B2 (ja) * | 2018-12-12 | 2023-05-26 | キヤノン株式会社 | 撮像装置及びその制御方法並びにプログラム |
| JP7516031B2 (ja) | 2019-11-19 | 2024-07-16 | キヤノン株式会社 | 撮像装置及び撮像システム |
| JP2020074596A (ja) | 2020-01-23 | 2020-05-14 | 株式会社ニコン | 撮像装置 |
| JP7458819B2 (ja) * | 2020-02-21 | 2024-04-01 | キヤノン株式会社 | 撮像装置、制御方法、プログラム及び記憶媒体 |
| JP7468632B2 (ja) * | 2020-04-15 | 2024-04-16 | 株式会社ニコン | 撮像素子、撮像装置及び装置 |
| JP7562306B2 (ja) * | 2020-06-23 | 2024-10-07 | キヤノン株式会社 | 光電変換装置、光電変換システム、および移動体 |
| JP7803283B2 (ja) | 2020-11-20 | 2026-01-21 | 株式会社ニコン | 撮像装置および撮像方法 |
| JP2023031696A (ja) * | 2021-08-25 | 2023-03-09 | 株式会社ニコン | 撮像素子および撮像装置 |
| US20240381005A1 (en) * | 2021-08-25 | 2024-11-14 | Nikon Corporation | Imaging element and imaging device |
| JP7786077B2 (ja) * | 2021-08-25 | 2025-12-16 | 株式会社ニコン | 撮像素子および撮像装置 |
| KR102826195B1 (ko) * | 2023-09-08 | 2025-06-26 | 아주대학교산학협력단 | 움직임 및 움직임 경로 감지 장치 및 방법과 이를 구비하는 비전 센서 |
Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPH0738803A (ja) * | 1993-07-20 | 1995-02-07 | Mitsubishi Electric Corp | ビデオカメラ |
| JPH07284006A (ja) * | 1994-04-12 | 1995-10-27 | Canon Inc | 撮像装置 |
| JPH1188763A (ja) * | 1997-06-20 | 1999-03-30 | Samsung Electron Co Ltd | デジタルスチルカメラのシャッター速度制御装置及び方法 |
| JP2000013690A (ja) * | 1998-06-19 | 2000-01-14 | Toshiba Corp | 画像撮影装置及びその方法 |
| JP2000278595A (ja) * | 1999-03-26 | 2000-10-06 | Minolta Co Ltd | デジタルカメラ及び撮像方法 |
| US20030020827A1 (en) | 2001-07-25 | 2003-01-30 | Bean Heather Noel | Image capturing device capable of single pixel exposure duration control |
| US20030174772A1 (en) | 2001-09-12 | 2003-09-18 | Transchip, Inc. | Systems and methods for utilizing activity detection information in relation to image processing |
Family Cites Families (30)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| SU1003383A1 (ru) * | 1981-10-28 | 1983-03-07 | Предприятие П/Я В-8450 | Фотоприемник движущихс изображений (его варианты) |
| JP3642793B2 (ja) | 1992-11-25 | 2005-04-27 | ソニー株式会社 | 固体撮像装置及び撮像装置の制御方法 |
| JPH06261256A (ja) | 1993-03-04 | 1994-09-16 | Fuji Photo Optical Co Ltd | 固体撮像素子を用いた撮像装置 |
| DE69739272D1 (de) * | 1996-11-05 | 2009-04-02 | Bae Systems Information | Vorrichtung zur elektro-optischen fernerkundung mit bewegungskompensation |
| JPH10336511A (ja) * | 1997-06-05 | 1998-12-18 | Sanyo Electric Co Ltd | カメラ装置 |
| GB2332800B (en) * | 1997-12-18 | 2000-09-27 | Simage Oy | Device for imaging radiation |
| US6665010B1 (en) * | 1998-07-21 | 2003-12-16 | Intel Corporation | Controlling integration times of pixel sensors |
| US6303923B1 (en) * | 1998-12-18 | 2001-10-16 | California Institute Of Technology | Fabricating a hybrid imaging device having non-destructive sense nodes |
| WO2000055810A1 (en) * | 1999-03-16 | 2000-09-21 | Hamamatsu Photonics K. K. | High-speed vision sensor |
| US6778210B1 (en) * | 1999-07-15 | 2004-08-17 | Olympus Optical Co., Ltd. | Image pickup apparatus with blur compensation |
| JP3430994B2 (ja) * | 1999-09-28 | 2003-07-28 | ミノルタ株式会社 | カメラ |
| US6995800B2 (en) * | 2000-01-27 | 2006-02-07 | Canon Kabushiki Kaisha | Image pickup apparatus utilizing a plurality of converging lenses |
| US6765619B1 (en) * | 2000-04-04 | 2004-07-20 | Pixim, Inc. | Method and apparatus for optimizing exposure time in image acquisitions |
| JP2001298175A (ja) * | 2000-04-12 | 2001-10-26 | Toshiba Corp | 撮像システム |
| US6809769B1 (en) * | 2000-06-22 | 2004-10-26 | Pixim, Inc. | Designs of digital pixel sensors |
| US7139028B2 (en) * | 2000-10-17 | 2006-11-21 | Canon Kabushiki Kaisha | Image pickup apparatus |
| JP2002231918A (ja) * | 2001-02-06 | 2002-08-16 | Olympus Optical Co Ltd | 固体撮像装置及びその製造方法 |
| JP3794942B2 (ja) * | 2001-07-09 | 2006-07-12 | 松下電器産業株式会社 | マルチチップモジュール及びその接続テスト方法 |
| JP3759435B2 (ja) * | 2001-07-11 | 2006-03-22 | ソニー株式会社 | X−yアドレス型固体撮像素子 |
| RU2235443C2 (ru) * | 2001-08-16 | 2004-08-27 | Федеральное государственное унитарное предприятие "Научно-исследовательский институт промышленного телевидения "Растр" | Телевизионная камера на матрице приборов с зарядовой связью |
| JP3833088B2 (ja) * | 2001-09-05 | 2006-10-11 | 独立行政法人科学技術振興機構 | 画像検出処理装置 |
| US7158180B2 (en) * | 2001-12-31 | 2007-01-02 | Texas Instruments Incorporated | System and method for varying exposure time for different parts of a field of view while acquiring an image |
| CN1234234C (zh) * | 2002-09-30 | 2005-12-28 | 松下电器产业株式会社 | 固体摄像器件及使用该固体摄像器件的设备 |
| JP3804617B2 (ja) * | 2003-02-14 | 2006-08-02 | コニカミノルタフォトイメージング株式会社 | 画像処理装置及び方法 |
| JP3709873B2 (ja) * | 2003-02-19 | 2005-10-26 | ソニー株式会社 | 固体撮像装置及び撮像カメラ |
| US6989589B2 (en) * | 2003-07-21 | 2006-01-24 | Motorola, Inc. | Programmable sensor array |
| JP2005056998A (ja) * | 2003-08-01 | 2005-03-03 | Fuji Photo Film Co Ltd | 固体撮像装置およびその製造方法 |
| US7091469B2 (en) * | 2004-05-20 | 2006-08-15 | St Assembly Test Services Ltd. | Packaging for optoelectronic devices |
| US7583305B2 (en) * | 2004-07-07 | 2009-09-01 | Eastman Kodak Company | Extended dynamic range imaging system |
| US7705884B2 (en) * | 2004-07-21 | 2010-04-27 | Zoran Corporation | Processing of video data to compensate for unintended camera motion between acquired image frames |
-
2005
- 2005-01-13 JP JP2005006178A patent/JP4277216B2/ja not_active Expired - Fee Related
- 2005-12-02 KR KR1020137026191A patent/KR101362604B1/ko not_active Expired - Fee Related
- 2005-12-02 EP EP05814419.7A patent/EP1838088B1/en not_active Expired - Lifetime
- 2005-12-02 KR KR1020127031162A patent/KR20120137443A/ko not_active Ceased
- 2005-12-02 US US11/794,996 patent/US8345108B2/en not_active Expired - Fee Related
- 2005-12-02 RU RU2007126682/09A patent/RU2343644C1/ru not_active IP Right Cessation
- 2005-12-02 MX MX2007008426A patent/MX2007008426A/es active IP Right Grant
- 2005-12-02 CA CA002593719A patent/CA2593719A1/en not_active Abandoned
- 2005-12-02 BR BRPI0519834-8A patent/BRPI0519834A2/pt not_active IP Right Cessation
- 2005-12-02 AU AU2005324643A patent/AU2005324643B2/en not_active Ceased
- 2005-12-02 KR KR1020077015722A patent/KR20070102678A/ko not_active Ceased
- 2005-12-02 WO PCT/JP2005/022603 patent/WO2006075463A1/ja not_active Ceased
- 2005-12-02 CN CNA2005800465648A patent/CN101099381A/zh active Pending
-
2012
- 2012-11-30 US US13/690,999 patent/US8687069B2/en not_active Expired - Fee Related
-
2014
- 2014-01-21 US US14/159,616 patent/US9185304B2/en not_active Expired - Fee Related
-
2015
- 2015-06-25 US US14/750,088 patent/US20150296113A1/en not_active Abandoned
-
2016
- 2016-02-18 US US15/047,467 patent/US9729764B2/en not_active Expired - Fee Related
-
2017
- 2017-05-09 US US15/590,909 patent/US10057515B2/en not_active Expired - Lifetime
Patent Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPH0738803A (ja) * | 1993-07-20 | 1995-02-07 | Mitsubishi Electric Corp | ビデオカメラ |
| JPH07284006A (ja) * | 1994-04-12 | 1995-10-27 | Canon Inc | 撮像装置 |
| JPH1188763A (ja) * | 1997-06-20 | 1999-03-30 | Samsung Electron Co Ltd | デジタルスチルカメラのシャッター速度制御装置及び方法 |
| JP2000013690A (ja) * | 1998-06-19 | 2000-01-14 | Toshiba Corp | 画像撮影装置及びその方法 |
| JP2000278595A (ja) * | 1999-03-26 | 2000-10-06 | Minolta Co Ltd | デジタルカメラ及び撮像方法 |
| US20030020827A1 (en) | 2001-07-25 | 2003-01-30 | Bean Heather Noel | Image capturing device capable of single pixel exposure duration control |
| US20030174772A1 (en) | 2001-09-12 | 2003-09-18 | Transchip, Inc. | Systems and methods for utilizing activity detection information in relation to image processing |
Non-Patent Citations (1)
| Title |
|---|
| See also references of EP1838088A4 |
Also Published As
| Publication number | Publication date |
|---|---|
| US20080284871A1 (en) | 2008-11-20 |
| US9729764B2 (en) | 2017-08-08 |
| MX2007008426A (es) | 2007-09-06 |
| AU2005324643B2 (en) | 2010-10-07 |
| KR20120137443A (ko) | 2012-12-20 |
| KR101362604B1 (ko) | 2014-02-12 |
| EP1838088A4 (en) | 2009-12-09 |
| CN101099381A (zh) | 2008-01-02 |
| JP4277216B2 (ja) | 2009-06-10 |
| AU2005324643A1 (en) | 2006-07-20 |
| US10057515B2 (en) | 2018-08-21 |
| US20140132785A1 (en) | 2014-05-15 |
| EP1838088A1 (en) | 2007-09-26 |
| US8687069B2 (en) | 2014-04-01 |
| US20160165100A1 (en) | 2016-06-09 |
| US20150296113A1 (en) | 2015-10-15 |
| EP1838088B1 (en) | 2020-07-15 |
| US20170244913A1 (en) | 2017-08-24 |
| KR20130114760A (ko) | 2013-10-17 |
| CA2593719A1 (en) | 2006-07-20 |
| US8345108B2 (en) | 2013-01-01 |
| US20130088608A1 (en) | 2013-04-11 |
| RU2343644C1 (ru) | 2009-01-10 |
| KR20070102678A (ko) | 2007-10-19 |
| JP2006197192A (ja) | 2006-07-27 |
| BRPI0519834A2 (pt) | 2009-03-17 |
| US9185304B2 (en) | 2015-11-10 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| JP4277216B2 (ja) | 撮像装置及び撮像結果の処理方法 | |
| JP4290066B2 (ja) | 固体撮像装置および撮像システム | |
| CN108028898B (zh) | 图像传感器和摄像设备 | |
| JP5852324B2 (ja) | 撮像装置及びその制御方法、プログラム | |
| US10397502B2 (en) | Method and apparatus for imaging an object | |
| WO2018198766A1 (ja) | 固体撮像装置および電子機器 | |
| JP2010147785A (ja) | 固体撮像素子及び撮像装置並びにその画像補正方法 | |
| JP4266726B2 (ja) | 撮像装置 | |
| US8026965B2 (en) | Image pickup apparatus and method for controlling the same | |
| JP2009284328A (ja) | 撮像装置 | |
| US11252354B2 (en) | Image sensor, control method thereof, and image capturing apparatus | |
| JP7661099B2 (ja) | 信号処理装置、光電変換装置、撮像システム及び信号処理装置の制御方法 | |
| JP2012235363A (ja) | 撮像装置 | |
| JP2007143067A (ja) | 撮像装置及び撮像システム | |
| JP2010166159A (ja) | カメラ、カメラのぶれ補正方法 | |
| JP5664010B2 (ja) | 撮像装置 | |
| JP2011101170A (ja) | Ccd型固体撮像素子のスミア補正方法及び撮像装置 | |
| JP2007336333A (ja) | 撮像装置 | |
| JP2013258652A (ja) | 固体撮像素子及びこれを用いた撮像装置 | |
| JP2015231157A (ja) | 撮像装置及びその駆動方法 |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
| WWE | Wipo information: entry into national phase |
Ref document number: 2005324643 Country of ref document: AU |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 2593719 Country of ref document: CA |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 2005814419 Country of ref document: EP Ref document number: 5184/DELNP/2007 Country of ref document: IN |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 1020077015722 Country of ref document: KR |
|
| WWE | Wipo information: entry into national phase |
Ref document number: MX/a/2007/008426 Country of ref document: MX |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 2007126682 Country of ref document: RU |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 200580046564.8 Country of ref document: CN |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| ENP | Entry into the national phase |
Ref document number: 2005324643 Country of ref document: AU Date of ref document: 20051202 Kind code of ref document: A |
|
| WWP | Wipo information: published in national office |
Ref document number: 2005324643 Country of ref document: AU |
|
| WWP | Wipo information: published in national office |
Ref document number: 2005814419 Country of ref document: EP |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 11794996 Country of ref document: US |
|
| ENP | Entry into the national phase |
Ref document number: PI0519834 Country of ref document: BR Kind code of ref document: A2 |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 1020127031162 Country of ref document: KR |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 1020137026191 Country of ref document: KR |