WO2021117633A1 - 撮像装置、情報処理装置、撮像方法、および情報処理方法 - Google Patents
撮像装置、情報処理装置、撮像方法、および情報処理方法 Download PDFInfo
- Publication number
- WO2021117633A1 WO2021117633A1 PCT/JP2020/045246 JP2020045246W WO2021117633A1 WO 2021117633 A1 WO2021117633 A1 WO 2021117633A1 JP 2020045246 W JP2020045246 W JP 2020045246W WO 2021117633 A1 WO2021117633 A1 WO 2021117633A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- polarization
- subject
- light
- unit
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
- G06T7/586—Depth or shape recovery from multiple images from multiple light sources, e.g. photometric stereo
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01B—MEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
- G01B11/00—Measuring arrangements characterised by the use of optical techniques
- G01B11/24—Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B11/00—Filters or other obturators specially adapted for photographic purposes
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B15/00—Special procedures for taking photographs; Apparatus therefor
- G03B15/02—Illuminating scene
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B15/00—Special procedures for taking photographs; Apparatus therefor
- G03B15/02—Illuminating scene
- G03B15/03—Combinations of cameras with lighting apparatus; Flash units
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B19/00—Cameras
- G03B19/02—Still-picture cameras
- G03B19/04—Roll-film cameras
- G03B19/06—Roll-film cameras adapted to be loaded with more than one film, e.g. with exposure of one or the other at will
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B35/00—Stereoscopic photography
- G03B35/08—Stereoscopic photography by simultaneous recording
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/56—Cameras or camera modules comprising electronic image sensors; Control thereof provided with illuminating means
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/10—Circuitry of solid-state image sensors [SSIS]; Control thereof for transforming different wavelengths into image signals
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/28—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 for polarising
- G02B27/288—Filters employing polarising elements, e.g. Lyot or Solc filters
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10141—Special mode during image acquisition
- G06T2207/10152—Varying illumination
Definitions
- the present disclosure relates to an imaging device, an information processing device, an imaging method, and an information processing method.
- illuminance difference stereo method As one of the methods for measuring the three-dimensional shape of a subject.
- the illuminance difference stereo method light is sequentially and individually applied to a subject from a plurality of directions, and the distance to the subject and the three-dimensional shape of the subject are measured from the difference in shadow (see, for example, Patent Document 1).
- the present disclosure proposes an imaging device, an information processing device, an imaging method, and an information processing method capable of accurately measuring the three-dimensional shape of an animal body.
- an imaging device includes an image pickup unit, a separation unit, a calculation unit, and an estimation unit.
- the imaging unit includes a plurality of illuminations having different polarization directions of the light irradiating the subject and a polarization sensor, and captures an image of the subject to which the light is simultaneously irradiated from the plurality of illuminations.
- the separation unit separates the pixel signals corresponding to each of the polarization directions from the image captured by the imaging unit to generate an image for each of the polarization directions.
- the calculation unit calculates the normal on the surface of the subject by the illuminance difference stereo method from the image for each polarization direction.
- the estimation unit estimates the shape of the subject based on the normal calculated by the calculation unit.
- this method is difficult to apply to an animal subject because the position shift occurs due to the movement of the subject during the switching of lighting and accurate measurement cannot be performed, and only the subject of a stationary object is accurate.
- the three-dimensional shape cannot be measured.
- wavelength division multiplexing measurement method in which the illumination is multiplexed by using the difference in the color of the light irradiating the subject, and the three-dimensional shape of the subject is measured by one imaging without switching the illumination.
- the wavelength multiplexing method when a subject is simultaneously irradiated with light having a different wavelength (color) from multiple lights to image the subject, each color component is extracted from the captured image, and each light is illuminated by itself. The three-dimensional shape of the subject is measured by obtaining the shadow of the subject.
- the wavelength division multiplexing method it is possible to measure the three-dimensional shape of the subject by one imaging without switching the illumination. Therefore, even when the subject is an animal body, the three-dimensional shape of the subject is obtained. Can be measured.
- the color of the light emitted to the subject is changed by applying a different narrow-band bandpass filter to each illumination, so that the amount of light is reduced by passing through the narrow-band bandpass filter.
- the S / N ratio may deteriorate and the measurement accuracy may decrease.
- the imaging device, the information processing device, the imaging method, and the information processing method according to the present disclosure accurately measure the three-dimensional shape of the animal body without increasing the cost.
- FIG. 1 is an explanatory diagram showing an outline of an imaging method and an information processing method according to the present disclosure.
- an animal is imaged by multiplexing illumination using polarized light and simultaneously irradiating a subject with light having different polarization directions from a plurality of directions in one imaging. Allows measurement of the three-dimensional shape of the body.
- one camera 10 and a plurality of (here, four) light sources L 1 , L 2 , L 3 , and L 4 are prepared.
- the light sources L 1 , L 2 , L 3 , and L 4 are arranged in different directions (hereinafter, referred to as light source directions) S 1 , S 2 , S 3 , and S 4 with respect to the subject 100.
- the number of light sources is not limited to four.
- the light source L 1, L 2, L 3 , L 4 , the polarization filter is a polarization direction different from the irradiation portion of light F 1, F 2, F 3 , F 4 is provided, subject to each polarization direction different light Irradiate to 100. Further, the light source L 1, L 2, L 3 , L 4 , the polarization direction and the light source direction of the irradiated light (the position with respect to the subject) is associated in advance.
- the camera 10 includes a polarization sensor 11.
- each polarization direction component is separated from the image data acquired by the polarization sensor 11 by signal separation processing. Then, based on the correspondence between the polarization direction and the light source direction of the light associated in advance, the image I that would be obtained when the light sources L 1 , L 2 , L 3 , and L 4 in each direction are irradiated independently. Calculate 1 , I 2 , I 3 , and I 4.
- the normal image I 5 is calculated by performing the normal calculation process for each image I 1 , I 2 , I 3 , and I 4 , and the subject is subjected to the distance estimation process using the normal image I 5.
- FIG. 2 is a block diagram showing a configuration example of the imaging device according to the present disclosure.
- the imaging device 1 according to the present disclosure includes an imaging unit 2 and an information processing device 3.
- the imaging unit 2 includes light sources L 1 , L 2 , L 3 , L 4 provided with polarizing filters F 1 , F 2 , F 3 , and F 4 shown in FIG. 1, and a camera 10.
- Imaging unit 2 a polarization filter F 1, F 2, F 3 , F 4 each light source L 1 through a, L 2, L 3, L 4 subject 100 in a state where polarization directions different light is irradiated simultaneously from The image is captured by the camera 10 and the image data of the captured image is output to the information processing device 3.
- the information processing device 3 includes, for example, a microcomputer having a CPU (Central Processing Unit), a ROM (Read Only Memory), a RAM (Random Access Memory), and various circuits.
- a CPU Central Processing Unit
- ROM Read Only Memory
- RAM Random Access Memory
- the information processing device 3 includes a calibration unit 4, a signal separation unit 5, and a normal calculation unit 6 that function by executing an information processing program stored in a ROM by a CPU using RAM as a work area. , A distance estimation unit 7 is provided.
- Part or all of the calibration unit 4, signal separation unit 5, normal calculation unit 6, and distance estimation unit 7 included in the information processing device 3 are ASIC (Application Specific Integrated Circuit) or FPGA (Field Programmable Gate Array). ) Etc. may be configured. Further, the signal separation unit 5, the normal calculation unit 6, and the distance estimation unit 7 do not necessarily have to be provided in the information processing device 3, and may be provided, for example, on the cloud.
- the calibration unit 4, the signal separation unit 5, the normal calculation unit 6, and the distance estimation unit 7 included in the information processing device 3 realize or execute the information processing operations described below, respectively.
- the internal configuration of the information processing device 3 is not limited to the configuration shown in FIG. 2, and may be another configuration as long as it is a configuration for performing information processing described later.
- the calibration unit 4 includes a storage unit, and provides information on linearization of the output brightness of the camera 10, information on shading data generated when illuminated by each light source, information in which the light source direction and the polarization direction are associated with each other, and the like.
- FIG. 3 is an explanatory diagram showing the relationship between the exposure amount of the camera and the output brightness according to the present disclosure. As shown in FIG. 3, in the camera 10, the larger the exposure amount i, the higher the output brightness i'in the captured image. However, as shown by the solid line in FIG. 3, the change in the output brightness i'of the camera 10 with the change in the exposure amount i is not linear.
- the calibration unit 4 measures the brightness of the images to be sequentially captured while changing the exposure amount i of the camera 10, and determines the relationship between the exposure amount i of the camera 10 and the output brightness i'shown by the solid line in FIG. Calculate the characteristic function shown and store the inverse conversion function of the characteristic function.
- the inverse transformation function is used when the preprocessing unit 51 (see FIG. 13) included in the signal separation unit 5, which will be described later, linearizes the image signal. As a result, the characteristic of the output brightness i'of the camera 10 is corrected to be linear as shown by the dotted line in FIG.
- the calibration unit 4 acquires and stores shading data indicating the state of shadows generated when light is irradiated by the light sources L 1 , L 2 , L 3 , and L 4.
- FIG. 4 and 5 are explanatory views of a shading data acquisition method according to the present disclosure.
- a gray plate 101 is arranged at a place where the subject 100 (see FIG. 1) is placed. Then, the gray plate 101 is sequentially irradiated with light by the light sources L 1 , L 2 , L 3 , and L 4 , and the image of the gray plate 101 is captured by the camera 10.
- the positions of the light sources L 1 , L 2 , L 3 , and L 4 with respect to the gray plate 101 are different from each other. Therefore, as shown in FIG. 4, for example, in the lit light sources L 1, is the image I 11 are imaging shadow becomes darker as direction from the lower left of the image to the upper right, to the lit light source L 2, the image Image I 12 in which the shadow becomes darker from the upper left to the lower right of the image is captured.
- the calibration unit 4 acquires and stores the image data of these images I 11 , I 12 , I 13 , and I 14 as shading data.
- the calibration unit 4 detects each light source direction.
- 6 to 8 are explanatory views of a light source direction detection method according to the present disclosure. As shown in FIG. 6, when detecting the direction of the light source, for example, a sphere 102 having a mirror-finished surface is placed at a place where the subject 100 (see FIG. 1) is placed. Then, the light sources L 1 , L 2 , L 3 , and L 4 sequentially irradiate the sphere 102 with light, and the camera 10 captures an image of the sphere 102.
- the positions of the light sources L 1 , L 2 , L 3 , and L 4 with respect to the sphere 102 are different from each other. Therefore, as shown in FIG. 7, for example, while the light source L 1 is lit, the image I 21 in which the position near the lower left of the sphere 102 is reflected as a high-brightness highlight portion is imaged, and the light source L 2 is lit. Image I 22 is captured in which the position near the upper left of the sphere 102 is reflected as a highlight portion.
- the image I 23 in which the position near the upper right of the sphere 102 is reflected as a highlight portion is imaged, and while the light source L 4 is lit, the position near the lower right of the sphere 102 is high.
- the image I 24 reflected as a light unit is imaged.
- the calibration unit 4 can calculate the light source directions of the light sources L 1 , L 2 , L 3 , and L 4 by the following equation (1).
- the calibration unit 4 associates the light source directions of the light sources L 1 , L 2 , L 3 , and L 4 with the polarization directions of the emitted light.
- the gray plate 101 When associating the light source direction with the polarization direction, the gray plate 101 (see FIG. 4) is imaged by the camera 10 in the same manner as when the shading data is acquired.
- the calibration unit 4 performs linearization and shading correction of image data imaged determines the polarization angle phi i performs estimation of the polarization model which will be described later, and the polarization direction of the polarized illuminating i the polarization angle phi i, previously It is associated with the detected light source direction and stored as a light source direction / polarization direction correspondence table.
- FIG. 9 is an explanatory diagram showing an example of a light source direction / polarization direction correspondence table according to the present disclosure.
- the light source L 1 of the light source number (i) is 1, the polarization direction (phi i) as phi 1 is associated, S 1 is associated as a light source direction (S i).
- the light source L 2 having the light source number (i) is associated with ⁇ 2 as the polarization direction ( ⁇ i ) and S 2 as the light source direction (S i ).
- the light source number (i) is the light source L 3 of 3, the polarization direction (phi i) as phi 3 is associated, is S 3 are associated as a light source direction (S i).
- the light source L 4 having the light source number (i) is associated with ⁇ 4 as the polarization direction ( ⁇ i ) and S 4 as the light source direction (S i ).
- the imaging unit 2 includes a plurality of polarized lights 21, 22, 23, 24, a camera 10, and a shooting control unit 12.
- Each polarized light 21, 22, 23, 24 uses a white LED light source L 1 , L 2 , L 3 , L 4, and polarizing filters F 1 , F 2 , F 3 , F 4 having different polarization directions, respectively.
- the polarizing filters F 1 , F 2 , F 3 , and F 4 are filters that selectively transmit light having polarization directions of 0 °, 45 °, 90 °, and 135 °, respectively.
- the camera 10 includes a polarization sensor 11. As shown in FIG. 11, the polarization sensor 11 selectively causes a pixel array 13 in which a plurality of image pickup elements are arranged in a matrix and light in different polarization directions associated with each image pickup element to enter the image pickup element. A polarizing filter 14 and a microlens 15 provided for each image sensor are provided.
- the polarizing filter 14 has regions 14a, 14b, 14c in which wire grids in four directions (0 °, 45 °, 90 °, 135 °) are arranged at positions facing each image sensor. , 14d are arranged, for example, in the same arrangement as the Bayer arrangement.
- the photographing control unit 12 turns on the plurality of polarized lights 21, 22, 23, 24 at the same time, and after that, the camera 10 images the subject 100 (see FIG. 1), and then the polarized lights 21, Turn off 22, 23, 24.
- the photographing control unit 12 when performing moving image imaging, causes the camera 10 to repeatedly perform imaging with the plurality of polarized lights 21, 22, 23, 24 turned on at the same time, and the user stops the imaging. Continue imaging until instructed.
- the photographing control unit 12 turns off the polarized lights 21, 22, 23, 24 after the moving image imaging is completed.
- the shooting control unit 12 acquires image data of the captured image from the camera 10 and outputs the image data to the signal separation unit 5 in the subsequent stage.
- FIG. 13 is a block diagram showing a configuration example of the signal separation unit according to the present disclosure.
- the signal separation unit 5 according to the present disclosure includes a preprocessing unit 51, a polarization demosaic unit 52, a polarization model estimation unit 53, and a polarization luminance calculation unit 54.
- the preprocessing unit 51 linearizes the output brightness i'of the camera 10 in the image data input from the camera 10 and performs shading correction.
- the preprocessing unit 51 linearizes the output brightness i'using the following equation (2).
- Pre-processing unit 51 as shown in equation (2), the output luminance j x of each pixel in the y, the output luminance j'x was linearized by applying the inverse transformation function characteristic function of the camera 10, y Is calculated.
- the preprocessing unit 51 performs shading correction using the following equation (3). As shown in the equation (3), the preprocessing unit 51 converts the linearized output luminance j'x , y into the luminance l of the corresponding pixel in the images I 11 , I 12 , I 13 , and I 14 shown in FIG. The shading-corrected output luminance j ′′ x, y is calculated by dividing by 1 l 2 ... l M.
- the polarization demosaic unit 52 obtains data in these four directions for each pixel from the data in each direction of 0 °, 45 °, 90 °, and 135 ° assigned to each pixel (j ′′ ′′ x, y (j ′′ ′′ x, y ( 0), j ′′ ′′ x, y (45), j ′′ ′′ x, y (90), j ′′ ′′ x, y (135)).
- FIG. 14 is an explanatory diagram of the polarized light demosaic process according to the present disclosure.
- the polarization demosaic unit 52 uses the data a, b, c, d of the image pickup device provided with the 90 ° wire grid (see FIG. 12) to form A, B, C.
- the data of the image I 31 subjected to the polarization demosaic processing is calculated.
- the polarization demosaic unit 52 calculates the data for each pixel by the same method for the data in the polarization directions of 0 °, 45 °, and 135 °. Returning to FIG. 13, the polarization demosaic unit 52 uses the calculated data (j ′′ ′′ x, y (0), j ′′ ′′ x, y (45), j ′′ ′′ x, y (90), j ′′. '' X, y (135)) is output to the polarization model estimation unit 53.
- the polarization model estimation unit 53 estimates a polarization model that shows the correspondence between the deflection angle and the brightness.
- FIG. 15 is an explanatory diagram showing an example of the polarization model according to the present disclosure.
- the polarization model estimation unit 53 uses the polarization sensor data (j ′′ ′′ x, y (0), j ′′ ′′ x, y (45), j ′′ ′′ x, y (90) obtained for each pixel. , J ′′ ′′ x, y (135)) to estimate the polarization model shown in FIG.
- the signal separation unit 5 can estimate the brightness I ( ⁇ ) at an arbitrary deflection angle ( ⁇ ).
- the polarization model shown in FIG. 15 is expressed by the following equation (5).
- the polarization model estimation unit 53 obtains the unknown parameters I max , I min , and ⁇ in the above equation (5) from the captured data I ( ⁇ 1 ), ..., I ( ⁇ M).
- the polarization model estimation unit 53 can obtain the unknown parameters I max , I min, and ⁇ by the following equation (9).
- the polarization luminance calculation unit 54 corresponds to each light source L 1 , L 2 , L 3 , L 4 by using the unknown parameters I max , I min , ⁇ , and equation (5) obtained by the polarization model estimation unit 53.
- the polarization brightness calculation unit 54 uses the angle of the polarization direction in the light source direction / polarization direction correspondence table 55.
- FIG. 16 is an explanatory diagram of a normal calculation method according to the present disclosure.
- the normal calculation unit 6 has, for example, when there are three light sources L 1 , L 2 , and L 3, the light source vector S corresponding to the light source direction in the light source direction / polarization direction correspondence table 55.
- the normal vector n is calculated for each pixel by performing the operations of the following equations (10) and (12) using i and the brightness i of each pixel input from the polarization brightness calculation unit 54.
- the normal calculation unit 6 calculates the normal vector n for each pixel by performing the operations of the following equations (11) and (12). calculate.
- M in the equation (11) is 4.
- the distance estimation unit 7 calculates the distance Z from a certain reference point to the corresponding point on the subject for each pixel by using the normal information obtained for each pixel.
- the distance estimation unit 7 calculates the distance Z by using, for example, the Francot-Chellappa algorithm represented by the following equation (13) using a Fourier basis.
- P and q in the above formula (13) are x component and y component of the normal vector n calculated by the normal calculation unit 6.
- F is a Fourier transform
- ⁇ x is a spatial frequency (x)
- ⁇ x is a spatial frequency (y).
- the distance estimation unit 7 sets a certain reference point in advance, integrates the gradient field from the reference point, and estimates the shape (distance Z) of the subject. At this time, the distance estimation unit 7 calculates the distance Z so that the derivatives of the gradient field and the shape match.
- FIGS. 17 to 19. 17 to 19 are flowcharts showing processes executed by the imaging apparatus according to the present disclosure.
- FIG. 17 shows an example of the calibration process executed by the calibration unit 4.
- FIG. 18 shows an example of the three-dimensional shape measurement process performed by the image pickup apparatus.
- FIG. 19 shows an example of signal separation processing in the three-dimensional shape measurement processing.
- the calibration unit 4 When performing the calibration process, as shown in FIG. 17, the calibration unit 4 first performs an inverse conversion function of the camera characteristics (characteristic function indicating the relationship between the exposure amount of the camera 10 and the output brightness shown in FIG. 3). Calculate and store (step S101). Subsequently, the calibration unit 4 acquires and stores shading data (see FIG. 5) (step S102).
- the calibration unit 4 calculates the light source direction of each light source (light source L 1 , L 2 , L 3 , L 4 ) (step S103), and each light source (light source L 1 , L 2 , L 3 , L 4). ) Is calculated (step S104). Then, the calibration unit 4 associates each light source direction with the polarization direction and stores it as a light source direction / polarization direction correspondence table 55 (step S105), and ends the calibration process.
- the image pickup unit 2 When performing the three-dimensional shape measurement process, as shown in FIG. 18, the image pickup unit 2 first performs the image pickup process (step S201). In the imaging process, the imaging unit 2 simultaneously irradiates the three-dimensionally shaped measurement object with light having different polarization directions from a plurality of directions, and the polarization sensor 11 images the measurement object. Subsequently, the signal separation unit 5 performs a signal separation process for separating an image signal corresponding to each polarization direction from the image captured by the image pickup unit 2 (step S202).
- the signal separation unit 5 first performs preprocessing (step S301).
- the signal separation unit 5 linearizes the output brightness of the camera 10 in the captured image using the inverse conversion function of the camera characteristics, and performs shading correction of the captured image using the shading data.
- the signal separation unit 5 performs polarization demosaic processing (step S302).
- the signal separation unit 5 generates image data for each polarization direction by demosacing the captured image after shading correction (see FIG. 14) and complementing the captured image.
- the signal separation unit 5 performs the polarization model estimation process (step S303).
- the signal separation unit 5 calculates unknown parameters (I max , I min , ⁇ ) in the polarization model (Equation (5)) from the image data (brightness for each pixel) for each polarization direction. Estimate the polarization model by
- the signal separation unit 5 performs the polarization luminance calculation process (step S304).
- the signal separation unit 5 determines the brightness of each pixel in the image for each light source direction based on the polarization direction corresponding to the light source direction included in the light source direction / polarization direction correspondence table 55 and the polarization model. It is calculated and output to the normal line calculation unit 6.
- the normal calculation unit 6 performs the normal calculation process (step S203).
- the normal calculation unit 6 sets the surface of the object to be measured pixel by pixel based on the brightness of each pixel in the image for each light source direction calculated by the signal separation process and the known light source direction. Calculate the normal vector.
- the distance estimation unit 7 performs the distance estimation process (step S204).
- the distance estimation unit 7 calculates the distance from a predetermined reference point to a point on the measurement object for each pixel by using the normal vector for each pixel calculated by the normal calculation process. Measures the three-dimensional shape of the object to be measured.
- FIG. 20 is an explanatory diagram showing a modified example of the camera according to the present disclosure.
- the camera according to the modified example includes the polarization sensor 10A shown in FIG.
- the polarizing sensor 10A includes beam splitters 15a, 15b, 15c, 15d, image sensors 10a, 10b, 10c, 10d, and polarizing filters 11a, 11b, 11c, 11d.
- the beam splitters 15a, 15b, 15c, 15d divide the incident light into a plurality of light rays.
- the image sensors 10a, 10b, 10c, and 10d receive each light beam.
- the polarizing filters 11a, 11b, 11c, 11d are provided between the image sensors 10a, 10b, 10c, 10d and the beam splitters 15a, 15b, 15c, 15d, and the polarization directions are set for each of the image sensors 10a, 10b, 10c, 10d. Is different.
- the polarizing filter 11a selectively transmits light having a polarization angle of 0 °, for example.
- the polarizing filter 11b selectively transmits light having a polarization angle of, for example, 45 °.
- the polarizing filter 11c selectively transmits light having a polarization angle of 90 °, for example.
- the polarizing filter 11d selectively transmits, for example, light having a polarization angle of 135 °.
- the image sensor 10a can capture an image of a subject to which only light having a polarization angle of 0 ° is applied.
- the image sensor 10b can capture an image of a subject to which only light having a polarization angle of 45 ° is applied.
- the image sensor 10c can capture an image of a subject to which only light having a polarization angle of 90 ° is applied.
- the image sensor 10d can capture an image of a subject to which only light having a polarization angle of 135 ° is applied.
- the imaging device 1 By providing the polarization sensor 10A shown in FIG. 20 in place of the polarization sensor 11, the imaging device 1 eliminates the need for the polarization demosaic processing performed by the polarization demosaic unit 52, so that the processing load can be reduced.
- the image pickup apparatus includes an image pickup unit 2, a signal separation unit 5, a normal calculation unit 6, and a distance estimation unit 7.
- the image pickup unit 2 includes a plurality of polarized lights 21, 22, 23, 24 and a polarizing sensor 11 having different polarization directions of the light irradiating the subject 100, and the light is simultaneously emitted from the plurality of polarized lights 21, 22, 23, 24. An image of the subject 100 to be irradiated is captured.
- the signal separation unit 5 separates the pixel signals corresponding to each polarization direction from the image captured by the image pickup unit 2 to generate an image for each polarization direction.
- the normal calculation unit 6 calculates the normal on the surface of the subject from the image for each polarization direction by the illuminance difference stereo method.
- the distance estimation unit 7 estimates the shape of the subject based on the normal calculated by the normal calculation unit. As a result, the image pickup apparatus 1 can accurately measure the three-dimensional shape of the animal body.
- the image pickup apparatus 1 relates to the polarized light 21,22,23,24, the polarization direction of the light emitted by the polarized light 21,22,23,24, and the subject 100 of the polarized light 21,22,23,24.
- the signal separation unit 5 is a polarization model showing a correspondence relationship between an arbitrary polarization direction and the brightness of each pixel in an image of a subject irradiated with light in the polarization direction, based on the brightness of each pixel in an image for each polarization direction.
- the image pickup apparatus 1 can calculate the accurate brightness of each pixel in the image for each of the polarized lights 21, 22, 23, 24.
- the normal calculation unit 6 calculates the normal on the surface of the subject 100 based on the brightness of each pixel in the image for each of the polarized lights 21, 22, 23, and 24 and the corresponding information. As a result, the image pickup apparatus 1 can calculate a more accurate normal.
- the plurality of polarized illuminations 21, 22, 23, 24 are provided with polarizing filters F 1 , F 2 , F 3 , and F 4 having different polarization directions on the light emitting surfaces of the light sources L 1 , L 2 , L 3 , and L 4. Be done.
- the image pickup apparatus 1 increases the number of light source directions without increasing the cost, for example, by simply increasing the number of light sources that emit white light and the number of deflection directions of light transmitted by the polarizing filter provided in the light sources. be able to.
- the polarizing sensor 11 has a pixel array 13 and a polarizing filter 14.
- a pixel array 13 a plurality of image pickup elements are arranged in a matrix.
- the polarizing filter 14 selectively causes light in different polarization directions associated with the image sensor to enter the image sensor.
- the image pickup apparatus 1 can capture a plurality of images for each light having different polarization directions by one pixel array 13.
- the polarizing sensor 11 includes beam splitters 15a, 15b, 15c, 15d, image sensors 10a, 10b, 10c, 10d, and polarizing filters 11a, 11b, 11c, 11d.
- the beam splitters 15a, 15b, 15c, 15d split the incident light into a plurality of light rays.
- the image sensors 10a, 10b, 10c, and 10d receive each light beam.
- the polarizing filters 11a, 11b, 11c, 11d are provided between the image sensors 10a, 10b, 10c, 10d and the beam splitters 15a, 15b, 15c, 15d, and the polarization directions are set for each of the image sensors 10a, 10b, 10c, 10d. Is different. As a result, the image pickup apparatus 1 does not require the polarization demosaicing process, so that the processing load can be reduced.
- the computer captures an image of a subject that is simultaneously irradiated with light from a plurality of illuminations having different polarization directions of the light emitted to the subject 100 by the polarization sensor 11, and each polarization is obtained from the image captured by the polarization sensor 11.
- the pixel signals corresponding to the directions are separated to generate an image for each polarization direction, the normal on the surface of the subject is calculated from the image for each polarization direction by the illuminance difference stereo method, and the shape of the subject is estimated based on the normal. Including doing.
- the imaging method can accurately measure the three-dimensional shape of the animal body.
- the information processing device 3 has a storage unit, a signal separation unit 5, a normal calculation unit 6, and a distance estimation unit 7.
- the storage unit includes a plurality of polarized lights 21, 22, 23, 24 having different polarization directions of the light irradiating the subject 100, the polarization directions of the light emitted by the polarized lights 21, 22, 23, 24, and the polarized lights.
- the light source direction / polarization direction correspondence table 55 which is the correspondence information in which the directions of 21, 22, 23, and 24 with respect to the subject are associated with each other, is stored.
- the signal separation unit 5 separates the pixel signals corresponding to each polarization direction from the image captured by the polarization sensor 11 by the subject 100, which is simultaneously irradiated with light from the plurality of polarized illuminations 21, 22, 23, 24, and provides corresponding information. An image for each illumination is generated based on.
- the normal calculation unit 6 calculates the normal on the surface of the subject 100 by the illuminance difference stereo method from the images for each of the polarized lights 21, 22, 23, and 24.
- the distance estimation unit 7 estimates the shape of the subject 100 based on the normal calculated by the normal calculation unit 6. As a result, the information processing device 3 can accurately measure the three-dimensional shape of the animal body.
- the information processing method includes a plurality of lights 21, 22, 23, 24 having different polarization directions of the light emitted by the computer to the subject 100, the polarization directions of the light emitted by the lights 21, 22, 23, 24, and the said light.
- the light source direction / polarization direction correspondence table 55 which is the correspondence information in which the directions of the polarized illuminations 21, 22, 23, 24 with respect to the subject 100 are associated with each other, is stored, and light is simultaneously emitted from the plurality of polarized illuminations 21, 22, 23, 24.
- the irradiated subject 100 separates the pixel signals corresponding to each polarization direction from the image captured by the polarization sensor 11, generates images for each of the polarized illuminations 21, 22, 23, and 24 based on the corresponding information, and the polarized illumination. This includes calculating the normal line on the surface of the subject 100 by the illuminance difference stereo method from the images for every 21, 22, 23, and 24, and estimating the shape of the subject 100 based on the normal line. As a result, the information processing method can accurately measure the three-dimensional shape of the animal body.
- the present technology can also have the following configurations.
- An imaging unit that includes a plurality of illuminations having different polarization directions of light irradiating the subject and a polarization sensor, and captures an image of the subject that is simultaneously irradiated with light from the plurality of illuminations.
- a separation unit that separates pixel signals corresponding to each polarization direction from an image captured by the imaging unit to generate an image for each polarization direction.
- a calculation unit that calculates a normal on the surface of the subject by the illuminance difference stereo method from the image for each polarization direction.
- An imaging device including an estimation unit that estimates the shape of the subject based on a normal calculated by the calculation unit.
- a storage unit further includes a storage unit that stores corresponding information in which the lighting, the polarization direction of the light emitted by the lighting, and the direction of the lighting with respect to the subject are associated with each other.
- the separation part Based on the brightness of each pixel in the image for each polarization direction, a polarization model showing the correspondence between the brightness of each pixel in the image of the subject irradiated with the light in the arbitrary polarization direction is estimated, and the polarization model is estimated.
- the imaging device according to (1) above, which calculates the brightness of each pixel in the image for each illumination based on the polarization model and the corresponding information.
- the calculation unit The imaging device according to (2), wherein a normal on the surface of the subject is calculated based on the brightness of each pixel in the image for each illumination and the corresponding information.
- the plurality of lights The imaging device according to any one of (1) to (3) above, wherein a polarizing filter having a different polarization direction is provided on a light emitting surface of a light source.
- the polarization sensor A pixel array in which multiple image sensors are arranged in a matrix, and The image pickup apparatus according to any one of (1) to (4) above, which has a polarizing filter that selectively incidents light in different polarization directions associated with the image pickup device onto the image pickup device.
- the polarization sensor A beam splitter that splits the incident light into multiple rays, An image sensor that receives each light beam and The imaging apparatus according to any one of (1) to (4) above, which is provided between the image sensor and the beam splitter and has a polarizing filter having a different polarization direction for each image sensor.
- the computer An image of the subject, which is simultaneously irradiated with light from a plurality of illuminations having different polarization directions of the light emitted to the subject, is captured by a polarization sensor. Pixel signals corresponding to each polarization direction are separated from the image captured by the polarization sensor to generate an image for each polarization direction.
- the normal on the surface of the subject is calculated from the image for each polarization direction by the illuminance difference stereo method.
- An imaging method including estimating the shape of the subject based on the normal.
- a storage unit that stores a plurality of illuminations having different polarization directions of light irradiating the subject, corresponding information in which the polarization directions of the light emitted by the illuminations and the directions of the illuminations with respect to the subject are associated with each other. Separation in which the subject, which is simultaneously irradiated with light from the plurality of illuminations, separates the pixel signals corresponding to the respective polarization directions from the image captured by the polarization sensor and generates an image for each illumination based on the corresponding information.
- the Department and A calculation unit that calculates a normal on the surface of the subject by the illuminance difference stereo method from the image for each illumination.
- An information processing device having an estimation unit that estimates the shape of the subject based on a normal calculated by the calculation unit.
- the computer A plurality of lights having different polarization directions of the light irradiating the subject, the polarization directions of the light radiated by the lights, and the corresponding information in which the directions of the lights with respect to the subject are associated with each other are stored.
- the subject which is simultaneously irradiated with light from the plurality of illuminations, separates the pixel signals corresponding to the respective polarization directions from the image captured by the polarization sensor, and generates an image for each illumination based on the corresponding information.
- the normal on the surface of the subject is calculated from the image for each illumination by the illuminance difference stereo method.
- An information processing method including estimating the shape of the subject based on the normal.
Landscapes
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Length Measuring Devices By Optical Means (AREA)
- Cameras In General (AREA)
- Stereoscopic And Panoramic Photography (AREA)
- Blocking Light For Cameras (AREA)
- Studio Devices (AREA)
- Image Analysis (AREA)
Abstract
Description
基本的な照度差ステレオ法では、被写体に対する光の照射方向が異なる複数の照明を順次切替えながら被写体へ光を当て、各照明によって照らされた被写体の画像を撮像し、各画像における被写体の陰影の違いに基づいて被写体の3次元形状を測定する。
まず、図1を参照して本開示に係る撮像方法および情報処理方法の概要について説明する。図1は、本開示に係る撮像方法および情報処理方法の概要を示す説明図である。
次に、図2を参照して本開示に係る撮像装置の構成について説明する。図2は、本開示に係る撮像装置の構成例を示すブロック図である。図2に示すように、本開示に係る撮像装置1は、撮像部2と、情報処理装置3とを備える。
キャリブレーション部4は、記憶部を備えており、カメラ10の出力輝度のリニア化に関する情報、各光源で照明した場合に生じるシェーディングデータの情報、光源方向と偏光方向とを対応付けした情報等を記憶する。
図3は、本開示に係るカメラの露光量と出力輝度との関係を示す説明図である。図3に示すように、カメラ10は、露光量iが多くなるほど、撮像画像における出力輝度i´が高くなる。ただし、図3に実線で示すように、露光量iの変化に伴うカメラ10の出力輝度i´の変化は線形(リニア)ではない。
また、キャリブレーション部4は、各光源L1,L2,L3,L4によって光を照射した場合に生じる陰影の状態を示すシェーディングデータを取得して記憶する。
また、キャリブレーション部4は、各光源方向の検出を行う。図6~図8は、本開示に係る光源方向の検出方法の説明図である。図6に示すように、光源方向の検出を行う場合には、被写体100(図1参照)が置かれる場所に、例えば、表面が鏡面仕上げされた形状が既知の球体102を配置する。そして、各光源L1,L2,L3,L4によって球体102へ順次光を照射し、カメラ10によって球体102の画像を撮像する。
また、キャリブレーション部4は、各光源L1,L2,L3,L4の光源方向と照射する光の偏光方向との対応付けを行う。
ここで、信号分離部5の説明を行う前に、図10~図12を参照して撮像部2の構成について説明する。図10~図12は、本開示に係る撮像部の説明図である。図10に示すように、撮像部2は、複数の偏光照明21,22,23,24と、カメラ10と、撮影制御部12とを含む。
次に、図13を参照して信号分離部について説明する。図13は、本開示に係る信号分離部の構成例を示すブロック図である。図13に示すように、本開示に係る信号分離部5は、前処理部51と、偏光デモザイク部52と、偏光モデル推定部53と、偏光輝度算出部54とを備える。
次に、図16を参照して法線算出部6による法線算出方法について説明する。図16は、本開示に係る法線算出方法の説明図である。
次に、距離推定部7について説明する。距離推定部7は、各画素ごとに求められた法線情報を用いて、各画素ごとに、ある基準点から対応する被写体上の点までの距離Zを算出する。距離推定部7は、例えば、フーリエ基底を用いた下記式(13)で表されるFrankot-Chellappaアルゴリズムを用いて距離Zを算出する。
次に、図17~図19を参照して本開示に係る撮像装置が実行する処理の一例について説明する。図17~図19は、本開示に係る撮像装置が実行する処理を示すフローチャートである。
次に、図20を参照してカメラ10の変形例について説明する。図20は、本開示に係るカメラの変形例を示す説明図である。変形例に係るカメラは、図20に示す偏光センサ10Aを備える。
撮像装置は、撮像部2と、信号分離部5と、法線算出部6と、距離推定部7とを有する。撮像部2は、被写体100へ照射する光の偏光方向が異なる複数の偏光照明21,22,23,24と偏光センサ11とを備え、複数の偏光照明21,22,23,24から同時に光が照射される被写体100の画像を撮像する。信号分離部5は、撮像部2によって撮像される画像から各偏光方向に対応する画素信号を分離して偏光方向毎の画像を生成する。法線算出部6は、偏光方向毎の画像から照度差ステレオ法によって被写体の表面における法線を算出する。距離推定部7は、法線算出部によって算出される法線に基づいて被写体の形状を推定する。これにより、撮像装置1は、動物体の3次元形状を正確に測定することができる。
(1)
被写体へ照射する光の偏光方向が異なる複数の照明と偏光センサとを備え、前記複数の照明から同時に光が照射される前記被写体の画像を撮像する撮像部と、
前記撮像部によって撮像される画像から各前記偏光方向に対応する画素信号を分離して前記偏光方向毎の画像を生成する分離部と、
前記偏光方向毎の画像から照度差ステレオ法によって前記被写体の表面における法線を算出する算出部と、
前記算出部によって算出される法線に基づいて前記被写体の形状を推定する推定部と
を有する撮像装置。
(2)
各前記照明、当該照明によって照射される光の偏光方向、および当該照明の前記被写体に対する方向が対応付けられた対応情報を記憶する記憶部
をさらに備え、
前記分離部は、
前記偏光方向毎の画像における各画素の輝度に基づいて、任意の偏光方向と当該偏光方向の光が照射された被写体の画像における各画素の輝度との対応関係を示す偏光モデルを推定し、前記偏光モデルと前記対応情報とに基づいて前記照明毎の画像における各画素の輝度を算出する
前記(1)に記載の撮像装置。
(3)
前記算出部は、
前記照明毎の画像における各画素の輝度と前記対応情報とに基づいて、前記被写体の表面における法線を算出する
前記(2)に記載の撮像装置。
(4)
前記複数の照明は、
それぞれ偏光方向が異なる偏光フィルタが光源の光出射面に設けられる
前記(1)から(3)のいずれかに記載の撮像装置。
(5)
前記偏光センサは、
複数の撮像素子が行列状に配列される画素アレイと、
前記撮像素子に対応付けられた異なる偏光方向の光を選択的に前記撮像素子へ入射させる偏光フィルタと
を有する前記(1)から(4)のいずれかに記載の撮像装置。
(6)
前記偏光センサは、
入射光を複数の光線に分割するビームスプリッタと、
各光線を受光するイメージセンサと、
前記イメージセンサと前記ビームスプリッタとの間に設けられ、前記イメージセンサ毎に偏光方向が異なる偏光フィルタと
を有する前記(1)から(4)のいずれかに記載の撮像装置。
(7)
コンピュータが、
被写体へ照射する光の偏光方向が異なる複数の照明から同時に光が照射される前記被写体の画像を偏光センサによって撮像し、
前記偏光センサによって撮像される画像から各前記偏光方向に対応する画素信号を分離して前記偏光方向毎の画像を生成し、
前記偏光方向毎の画像から照度差ステレオ法によって前記被写体の表面における法線を算出し、
前記法線に基づいて前記被写体の形状を推定する
ことを含む撮像方法。
(8)
被写体へ照射する光の偏光方向が異なる複数の照明、当該照明によって照射される光の偏光方向、および当該照明の前記被写体に対する方向が対応付けられた対応情報を記憶する記憶部と、
前記複数の照明から同時に光が照射される前記被写体が偏光センサによって撮像された画像から各前記偏光方向に対応する画素信号を分離して前記対応情報に基づいて前記照明毎の画像を生成する分離部と、
前記照明毎の画像から照度差ステレオ法によって前記被写体の表面における法線を算出する算出部と、
前記算出部によって算出される法線に基づいて前記被写体の形状を推定する推定部と
を有する情報処理装置。
(9)
コンピュータが、
被写体へ照射する光の偏光方向が異なる複数の照明、当該照明によって照射される光の偏光方向、および当該照明の前記被写体に対する方向が対応付けられた対応情報を記憶し、
前記複数の照明から同時に光が照射される前記被写体が偏光センサによって撮像された画像から各前記偏光方向に対応する画素信号を分離して前記対応情報に基づいて前記照明毎の画像を生成し、
前記照明毎の画像から照度差ステレオ法によって前記被写体の表面における法線を算出し、
前記法線に基づいて前記被写体の形状を推定する
ことを含む情報処理方法。
2 撮像部
3 情報処理装置
4 キャリブレーション部
5 信号分離部
6 法線算出部
7 距離推定部
10 カメラ
11 偏光センサ
12 撮影制御部
51 前処理部
52 偏光デモザイク部
53 偏光モデル推定部
54 偏光輝度算出部
55 光源方向/偏光方向対応表
Claims (9)
- 被写体へ照射する光の偏光方向が異なる複数の照明と偏光センサとを備え、前記複数の照明から同時に光が照射される前記被写体の画像を撮像する撮像部と、
前記撮像部によって撮像される画像から各前記偏光方向に対応する画素信号を分離して前記偏光方向毎の画像を生成する分離部と、
前記偏光方向毎の画像から照度差ステレオ法によって前記被写体の表面における法線を算出する算出部と、
前記算出部によって算出される法線に基づいて前記被写体の形状を推定する推定部と
を有する撮像装置。 - 各前記照明、当該照明によって照射される光の偏光方向、および当該照明の前記被写体に対する方向が対応付けられた対応情報を記憶する記憶部
をさらに備え、
前記分離部は、
前記偏光方向毎の画像における各画素の輝度に基づいて、任意の偏光方向と当該偏光方向の光が照射された被写体の画像における各画素の輝度との対応関係を示す偏光モデルを推定し、前記偏光モデルと前記対応情報とに基づいて前記照明毎の画像における各画素の輝度を算出する
請求項1に記載の撮像装置。 - 前記算出部は、
前記照明毎の画像における各画素の輝度と前記対応情報とに基づいて、前記被写体の表面における法線を算出する
請求項2に記載の撮像装置。 - 前記複数の照明は、
それぞれ偏光方向が異なる偏光フィルタが光源の光出射面に設けられる
請求項1に記載の撮像装置。 - 前記偏光センサは、
複数の撮像素子が行列状に配列される画素アレイと、
前記撮像素子に対応付けられた異なる偏光方向の光を選択的に前記撮像素子へ入射させる偏光フィルタと
を有する請求項1に記載の撮像装置。 - 前記偏光センサは、
入射光を複数の光線に分割するビームスプリッタと、
各光線を受光するイメージセンサと、
前記イメージセンサと前記ビームスプリッタとの間に設けられ、前記イメージセンサ毎に偏光方向が異なる偏光フィルタと
を有する請求項1に記載の撮像装置。 - コンピュータが、
被写体へ照射する光の偏光方向が異なる複数の照明から同時に光が照射される前記被写体の画像を偏光センサによって撮像し、
前記偏光センサによって撮像される画像から各前記偏光方向に対応する画素信号を分離して前記偏光方向毎の画像を生成し、
前記偏光方向毎の画像から照度差ステレオ法によって前記被写体の表面における法線を算出し、
前記法線に基づいて前記被写体の形状を推定する
ことを含む撮像方法。 - 被写体へ照射する光の偏光方向が異なる複数の照明、当該照明によって照射される光の偏光方向、および当該照明の前記被写体に対する方向が対応付けられた対応情報を記憶する記憶部と、
前記複数の照明から同時に光が照射される前記被写体が偏光センサによって撮像された画像から各前記偏光方向に対応する画素信号を分離して前記対応情報に基づいて前記照明毎の画像を生成する分離部と、
前記照明毎の画像から照度差ステレオ法によって前記被写体の表面における法線を算出する算出部と、
前記算出部によって算出される法線に基づいて前記被写体の形状を推定する推定部と
を有する情報処理装置。 - コンピュータが、
被写体へ照射する光の偏光方向が異なる複数の照明、当該照明によって照射される光の偏光方向、および当該照明の前記被写体に対する方向が対応付けられた対応情報を記憶し、
前記複数の照明から同時に光が照射される前記被写体が偏光センサによって撮像された画像から各前記偏光方向に対応する画素信号を分離して前記対応情報に基づいて前記照明毎の画像を生成し、
前記照明毎の画像から照度差ステレオ法によって前記被写体の表面における法線を算出し、
前記法線に基づいて前記被写体の形状を推定する
ことを含む情報処理方法。
Priority Applications (3)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US17/756,776 US12293534B2 (en) | 2019-12-13 | 2020-12-04 | Imaging device, information processing device, imaging method, and information processing method |
| CN202080079371.7A CN114731368B (zh) | 2019-12-13 | 2020-12-04 | 成像装置、信息处理装置、成像方法和信息处理方法 |
| JP2021563921A JP7447916B2 (ja) | 2019-12-13 | 2020-12-04 | 撮像装置、情報処理装置、撮像方法、および情報処理方法 |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2019-225887 | 2019-12-13 | ||
| JP2019225887 | 2019-12-13 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2021117633A1 true WO2021117633A1 (ja) | 2021-06-17 |
Family
ID=76330334
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/JP2020/045246 Ceased WO2021117633A1 (ja) | 2019-12-13 | 2020-12-04 | 撮像装置、情報処理装置、撮像方法、および情報処理方法 |
Country Status (4)
| Country | Link |
|---|---|
| US (1) | US12293534B2 (ja) |
| JP (1) | JP7447916B2 (ja) |
| CN (1) | CN114731368B (ja) |
| WO (1) | WO2021117633A1 (ja) |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP2021117232A (ja) * | 2020-01-27 | 2021-08-10 | コグネックス・コーポレイション | 複数タイプの光を用いるビジョン検査のためのシステム及び方法 |
| JP2022100524A (ja) * | 2020-12-24 | 2022-07-06 | オムロン株式会社 | 検査システムおよび検査方法 |
Families Citing this family (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2024256008A1 (en) * | 2023-06-15 | 2024-12-19 | Hexagon Technology Center Gmbh | Measuring system for measuring a surface of an object or skin of a person |
Citations (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20140118502A1 (en) * | 2011-07-13 | 2014-05-01 | Dongguk University Gyeongju Campus Industry-Academy Cooperation Foundation | System and method for extracting a 3d shape of a hot metal surface |
| US20140168382A1 (en) * | 2011-06-22 | 2014-06-19 | Dongguk University Gyeongju Campus Industry-Academy Cooperation Foundation | Method and system for reliable 3d shape extraction of metal surface |
| JP2018073122A (ja) * | 2016-10-28 | 2018-05-10 | キヤノン株式会社 | 画像処理装置および画像処理方法 |
Family Cites Families (26)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN101044507B (zh) | 2005-09-01 | 2010-08-18 | 松下电器产业株式会社 | 图像处理方法以及图像处理装置 |
| JP2007206797A (ja) | 2006-01-31 | 2007-08-16 | Omron Corp | 画像処理方法および画像処理装置 |
| CN101422035B (zh) * | 2006-05-29 | 2012-02-29 | 松下电器产业株式会社 | 光源推定装置、光源推定系统与光源推定方法以及图像高分辨率化装置与图像高分辨率化方法 |
| JP4435867B2 (ja) * | 2008-06-02 | 2010-03-24 | パナソニック株式会社 | 法線情報を生成する画像処理装置、方法、コンピュータプログラム、および、視点変換画像生成装置 |
| CN101960859B (zh) * | 2008-07-08 | 2013-04-10 | 松下电器产业株式会社 | 图像处理方法、图像处理装置、图像合成方法和图像合成装置 |
| WO2010073547A1 (ja) | 2008-12-25 | 2010-07-01 | パナソニック株式会社 | 画像処理装置及び擬似立体画像生成装置 |
| EP2223650A1 (en) * | 2009-02-25 | 2010-09-01 | The Provost, Fellows and Scholars of the College of the Holy and Undivided Trinity of Queen Elizabeth near Dublin | Method and apparatus for imaging tissue topography |
| CN103037752B (zh) * | 2010-09-24 | 2015-05-13 | 松下电器产业株式会社 | 图像处理装置 |
| JP2012143363A (ja) | 2011-01-11 | 2012-08-02 | Panasonic Corp | 画像処理装置 |
| JP5503573B2 (ja) * | 2011-02-21 | 2014-05-28 | 日本放送協会 | 撮像装置および画像処理情報生成プログラム |
| WO2013080658A1 (ja) * | 2011-11-29 | 2013-06-06 | オリンパスメディカルシステムズ株式会社 | 偏光観察装置 |
| CN103826530B (zh) * | 2012-07-25 | 2015-12-30 | 松下电器产业株式会社 | 摄像系统 |
| US20170065178A1 (en) * | 2014-03-12 | 2017-03-09 | Sony Corporation | Measurement device and measurement method |
| JP6432968B2 (ja) * | 2014-06-26 | 2018-12-05 | 国立大学法人岐阜大学 | 物体形状推定装置及びプログラム |
| JP6673327B2 (ja) * | 2015-02-27 | 2020-03-25 | ソニー株式会社 | 画像処理装置と画像処理方法および撮像素子 |
| CN107533370B (zh) * | 2015-04-30 | 2021-05-11 | 索尼公司 | 图像处理设备、图像处理方法和程序 |
| WO2017013913A1 (ja) * | 2015-07-17 | 2017-01-26 | ソニー株式会社 | 視線検出装置、アイウェア端末、視線検出方法及びプログラム |
| JP6556013B2 (ja) * | 2015-10-08 | 2019-08-07 | キヤノン株式会社 | 処理装置、処理システム、撮像装置、処理方法、プログラム、および記録媒体 |
| JP6679289B2 (ja) * | 2015-11-30 | 2020-04-15 | キヤノン株式会社 | 処理装置、処理システム、撮像装置、処理方法、処理プログラムおよび記録媒体 |
| WO2018017897A1 (en) * | 2016-07-20 | 2018-01-25 | Mura Inc. | Systems and methods for 3d surface measurements |
| CN108267896A (zh) * | 2016-12-30 | 2018-07-10 | 深圳超多维光电子有限公司 | 偏振成像装置和方法 |
| US10108261B1 (en) * | 2017-07-05 | 2018-10-23 | Oculus Vr, Llc | Eye tracking based on light polarization |
| JP7103357B2 (ja) * | 2017-07-26 | 2022-07-20 | ソニーグループ株式会社 | 情報処理装置、情報処理方法、及びプログラム |
| JP6892603B2 (ja) * | 2017-12-07 | 2021-06-23 | 富士通株式会社 | 距離計測装置、距離計測方法および距離計測プログラム |
| CN109285213A (zh) * | 2018-07-18 | 2019-01-29 | 西安电子科技大学 | 全方位偏振三维重建方法 |
| JP7236854B2 (ja) * | 2018-12-11 | 2023-03-10 | キヤノン株式会社 | 形状情報取得装置および形状情報取得方法、プログラム並びに記憶媒体 |
-
2020
- 2020-12-04 WO PCT/JP2020/045246 patent/WO2021117633A1/ja not_active Ceased
- 2020-12-04 US US17/756,776 patent/US12293534B2/en active Active
- 2020-12-04 JP JP2021563921A patent/JP7447916B2/ja active Active
- 2020-12-04 CN CN202080079371.7A patent/CN114731368B/zh active Active
Patent Citations (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20140168382A1 (en) * | 2011-06-22 | 2014-06-19 | Dongguk University Gyeongju Campus Industry-Academy Cooperation Foundation | Method and system for reliable 3d shape extraction of metal surface |
| US20140118502A1 (en) * | 2011-07-13 | 2014-05-01 | Dongguk University Gyeongju Campus Industry-Academy Cooperation Foundation | System and method for extracting a 3d shape of a hot metal surface |
| JP2018073122A (ja) * | 2016-10-28 | 2018-05-10 | キヤノン株式会社 | 画像処理装置および画像処理方法 |
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP2021117232A (ja) * | 2020-01-27 | 2021-08-10 | コグネックス・コーポレイション | 複数タイプの光を用いるビジョン検査のためのシステム及び方法 |
| JP7413293B2 (ja) | 2020-01-27 | 2024-01-15 | コグネックス・コーポレイション | 複数タイプの光を用いるビジョン検査のためのシステム及び方法 |
| JP2022100524A (ja) * | 2020-12-24 | 2022-07-06 | オムロン株式会社 | 検査システムおよび検査方法 |
| JP7686965B2 (ja) | 2020-12-24 | 2025-06-03 | オムロン株式会社 | 検査システムおよび検査方法 |
Also Published As
| Publication number | Publication date |
|---|---|
| US12293534B2 (en) | 2025-05-06 |
| CN114731368A (zh) | 2022-07-08 |
| CN114731368B (zh) | 2024-07-16 |
| US20230342963A1 (en) | 2023-10-26 |
| JP7447916B2 (ja) | 2024-03-12 |
| JPWO2021117633A1 (ja) | 2021-06-17 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Feng et al. | General solution for high dynamic range three-dimensional shape measurement using the fringe projection technique | |
| CN104634276B (zh) | 三维测量系统、拍摄设备和方法、深度计算方法和设备 | |
| JP7447916B2 (ja) | 撮像装置、情報処理装置、撮像方法、および情報処理方法 | |
| JP6478713B2 (ja) | 計測装置および計測方法 | |
| JP7056131B2 (ja) | 画像処理システム、画像処理プログラム、および画像処理方法 | |
| JP5948496B2 (ja) | 3次元形状測定装置の高さ測定方法 | |
| JP6556013B2 (ja) | 処理装置、処理システム、撮像装置、処理方法、プログラム、および記録媒体 | |
| JP2016038315A (ja) | 計測装置、基板検査装置、及びその制御方法 | |
| CN104634277A (zh) | 拍摄设备和方法、三维测量系统、深度计算方法和设备 | |
| JP2009025189A (ja) | 計測器 | |
| JP2022177166A (ja) | 検査方法、プログラム、及び、検査システム | |
| WO2019066724A1 (en) | LIGHT PROJECTION SYSTEMS | |
| KR101766468B1 (ko) | 트리플 주파수 패턴을 이용한 3차원 형상 측정 방법 | |
| CN114689604B (zh) | 具有光滑表面的待测物于光学检测上的影像处理方法及其检测系统 | |
| JP2018116032A (ja) | 被計測物の形状を計測する計測装置 | |
| JPWO2017168477A1 (ja) | 撮像装置および画像処理方法 | |
| KR20130064564A (ko) | 역구배를 가지는 대상물체의 입체형상 측정 장치 | |
| US20200250847A1 (en) | Surface reconstruction of an illuminated object by means of photometric stereo analysis | |
| JP6432968B2 (ja) | 物体形状推定装置及びプログラム | |
| JP2010078339A (ja) | 3次元形状計測装置および方法並びにプログラム | |
| US20250047986A1 (en) | Method and arrangements for obtaining and associating multispectral 2d image data with 3d image data from light triangulation | |
| KR101750883B1 (ko) | 비전 검사 시스템의 3차원 형상 측정 방법 | |
| KR101223800B1 (ko) | Rgb 컬러가 적용된 모아레를 이용한 형상 측정장치 | |
| JP2011252835A (ja) | 三次元形状計測装置 | |
| JP6939501B2 (ja) | 画像処理システム、画像処理プログラム、および画像処理方法 |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20898419 Country of ref document: EP Kind code of ref document: A1 |
|
| ENP | Entry into the national phase |
Ref document number: 2021563921 Country of ref document: JP Kind code of ref document: A |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| 122 | Ep: pct application non-entry in european phase |
Ref document number: 20898419 Country of ref document: EP Kind code of ref document: A1 |
|
| WWG | Wipo information: grant in national office |
Ref document number: 17756776 Country of ref document: US |