US20120154540A1 - Three dimensional measurement apparatus and three dimensional measurement method - Google Patents
Three dimensional measurement apparatus and three dimensional measurement method Download PDFInfo
- Publication number
- US20120154540A1 US20120154540A1 US13/314,577 US201113314577A US2012154540A1 US 20120154540 A1 US20120154540 A1 US 20120154540A1 US 201113314577 A US201113314577 A US 201113314577A US 2012154540 A1 US2012154540 A1 US 2012154540A1
- Authority
- US
- United States
- Prior art keywords
- pattern light
- target object
- measurement target
- bright
- period
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01B—MEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
- G01B11/00—Measuring arrangements characterised by the use of optical techniques
- G01B11/24—Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures
- G01B11/25—Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures by projecting a pattern, e.g. one or more lines, moiré fringes on the object
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/521—Depth or shape recovery from laser ranging, e.g. using interferometry; from the projection of structured light
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/363—Image reproducers using image projection screens
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10141—Special mode during image acquisition
- G06T2207/10152—Varying illumination
Definitions
- the present invention relates to a three dimensional measurement apparatus and a method using a space encoding method for generating spatial modulation by projecting pattern lights including bright parts and dark parts arranged at an arbitrary width.
- a projection apparatus and an imaging apparatus are arranged in a known relationship.
- the projection apparatus projects pattern lights to a subject, and the imaging apparatus images the subject.
- the principle of triangulation is used based on the relationship between the projection apparatus and the imaging apparatus to calculate the distance to the subject.
- a first bright and dark pattern light including bright parts and dark parts alternately arranged at an arbitrary width and a second bright and dark pattern light in which the phase of the first bright and dark pattern light is shifted are projected. Tone distributions of photographed images are acquired, and positions of intersections between the first and second pattern lights on the imaging element are calculated.
- the tone distributions of the first and second bright and dark pattern lights on the imaging element are sampled by imaging pixels, and straight line approximation is applied to the sampling points of the pattern lights to calculate the intersections.
- the intersections include sampling errors due to the straight line approximation if the intersections are calculated by the conventional method. Therefore, there are errors in the intersection distances if the intersection distances are calculated from the intersections.
- an embodiment of the present invention provides a three dimensional measurement apparatus and a method capable of accurate three dimensional measurement even if sampling per period of pattern lights is performed by a small number of pixels of an imaging element.
- a three dimensional measurement apparatus comprising: a projection unit that projects, to a measurement target object, a first pattern light including alternately arranged bright parts and dark parts and a second pattern light in which a phase of the first pattern light is shifted; an imaging unit that images the measurement target object, on which the first or second pattern light is projected, as image data; and a calculation unit that calculates a three dimensional position by calculating intersection positions of tone values from tone values of the image data of the first pattern light and the image data of the second pattern light obtained by the imaging unit, wherein when a period of repetitions of the bright parts and the dark parts of the first or second pattern light is one period, a range of imaging on the measurement target object surface by one pixel included in the imaging unit is an image distance, and the length of one period of the projected first pattern light or second pattern light on the measurement target object surface is M times the image distance, the projection unit and the imaging unit are arranged to satisfy
- N is an integer value equal to or greater than 2.
- a three dimensional measurement method by a three dimensional measurement apparatus comprising: a projection unit that projects, to a measurement target object, a first pattern light including alternately arranged bright parts and dark parts and a second pattern light in which a phase of the first pattern light is shifted; and an imaging unit that images the measurement target object on which the first or second pattern light is projected, the method comprising: a first acquisition step by a first acquisition unit acquiring first image data by causing the projection unit to project the first pattern light to cause the imaging unit to image the measurement target object; a second acquisition step by a second acquisition unit acquiring second image data by causing the projection unit to project the second pattern light to cause the imaging unit to image the measurement target object; and a calculation step by a calculation unit calculating a three dimensional position by calculating intersection positions of tone values of the first pattern light in the first image data and tone values of the second pattern light in the second image data, wherein in the first and second acquisition steps, when a period of
- N is an integer value equal to or greater than 2.
- FIGS. 1A and 1B are diagrams showing a basic configuration of a three dimensional measurement apparatus according to an embodiment.
- FIG. 2 is a diagram illustrating a relationship between bright and dark pattern lights on a measurement target object surface and pixels of a CCD.
- FIG. 3 is a diagram illustrating tone values acquired from image data obtained by setting the number of pixels of the CCD in a period of the bright and dark pattern lights to an even number.
- FIG. 4 is a diagram showing an example when a condition of the embodiment is not satisfied.
- FIG. 5 is a diagram showing error amounts of intersection distances when a period of the bright and dark pattern lights on the measurement target object surface is sampled by pixels of a CCD camera.
- FIG. 1A is a diagram showing an example of configuration of a three dimensional measurement apparatus according to the embodiment.
- reference numeral 101 denotes a measurement target object
- 102 denotes a projector as an example of a projection apparatus
- 103 denotes a CCD camera as an example of an imaging apparatus.
- Reference numeral 105 denotes a computer (hereinafter, PC 105 ) that inputs an image (pattern to be projected) to the projector 102
- 106 denotes a computer (hereinafter, PC 106 ) that acquires tone values of image data imaged by the CCD camera 103 to calculate the position of the measurement target object.
- the PC 105 and the PC 106 are connected to be able to communicate each other. It is obvious that the PC 105 and the PC 106 can be realized by one computer.
- FIG. 1B shows an example of pattern lights used in the present embodiment.
- a first pattern light including alternately arranged bright parts and dark parts and a second pattern light in which the phase of the bright parts and the dark parts of the first pattern light is shifted are used in the present embodiment.
- a first bright and dark pattern light 110 a is an example of the first pattern light
- a second bright and dark pattern light 110 b in which the phase of the first bright and dark pattern light 110 a is shifted is an example of the second pattern light.
- the first and second pattern lights are not limited to the illustrated examples, and any pattern lights that satisfy the requirements described below can be used.
- the PC 105 and the PC 106 calculate the position of the measurement target object 101 .
- the PC 105 projects the first bright and dark pattern light 110 a to the projector 102
- the PC 106 causes the CCD camera 103 to image the measurement target object 101 , on which the first bright and dark pattern light 110 a is projected, to acquire first image data.
- the PC 105 projects the second bright and dark pattern light 110 b to the projector 102
- the PC 106 causes the CCD camera 103 to image the measurement target object 101 , on which the second bright and dark pattern light 110 b is projected, to acquire second image data.
- the CCD camera 103 images a luminance distribution on the measurement target object 101 , on which the first bright and dark pattern light 110 a and the second bright and dark pattern light 110 b are projected through the projector 102 , and digital image data is imported to the PC 106 .
- values of the image data of the first and second bright and dark pattern lights will be called tone values.
- the three dimensional measurement apparatus calculates the intersection positions of the tone values as described below based on the tone values of the image data (first image data) of the first bright and dark pattern light and the tone values of the image data (second image data) of the second bright and dark pattern light obtained by the CCD camera 103 .
- a three dimensional position is calculated based on the calculated intersection positions.
- tone values of the first and second bright and dark pattern lights 110 a and 110 b are connected along a CCD pixel column in the first and second image data, lines connecting the tones intersect as described later in FIG. 3 . It is preferable that the direction along the CCD pixel column is a direction orthogonal to the bright and dark alignment in the bright and dark pattern light.
- FIG. 2 is an expanded view of part on the measurement target object 101 shown in FIG. 1A and illustrates the relationship between the bright and dark pattern light projected to the measurement target object and the CCD pixels.
- FIG. 3 shows a state after imaging under the first bright and dark pattern light 110 a and the second bright and dark pattern light 110 b , acquiring the tone values from the image data, and detecting boundaries of the tones from the first and second bright and dark pattern lights.
- Curves V and W denote intensity distributions on the imaging element of the first and second bright and dark pattern lights projected to the measurement target object 101
- D white rectangles
- Points A black circles
- points B black triangles
- Points C denote a point sequence formed by intersections between a line obtained by connecting the point sequence A by a straight line and a line obtained by connecting the point sequence B by a straight line.
- the true value of the intersection distance is half the four pixels, or 2.0. Even if the points D and C are different, the intersection distance obtained from the point C is the true value, or 2.0, as a result of sampling a period of the bright and dark pattern lights by even pixels of the imaging element as in the present embodiment.
- the error amount of each intersection distance can be reduced by forming a period of the bright and dark pattern lights on the surface of the measurement target object 101 to be M times (M is even) the image distance on the surface of the measurement target object 101 .
- M is an even value ⁇ 0.2.
- the intersection distances can be close to the true value by arranging the measurement target object 101 , the projector 102 , and the CCD camera 103 to satisfy
- intersection distances allow accurate detection of the intersection positions, and highly accurate three dimensional measurement can be further realized.
- FIG. 4 shows an example in which a period of the bright and dark pattern lights on the measurement target object surface is not an even multiple of the image distance on the measurement target object surface.
- M 4.5
- FIG. 5 shows the result of the calculation of error amounts of the intersection distances when a period of the bright and dark pattern lights projected to the measurement target object is sampled by M pixels of the imaging element.
- the abscissa denotes the number of sampling (M) per period of the bright and dark pattern lights
- the ordinate denotes error amounts (unit: %) generated, by sampling, from the intersection distances calculated from the three dimensional measurement relative to the width of a period of the bright and dark pattern lights.
- Values close to the true value can be acquired for the intersection distances if the number of sampling M satisfies Expression (1), and the three dimensional measurement can be accurately detected. If the number of sampling M is out of Expression (1), the acquired intersection distances are out of the true value, and it is difficult to accurately detect the three dimensional measurement.
- intersection distances between adjacent intersections are calculated from the three dimensional measurement when the number of sampling M satisfies
- intersection distances calculated at each section are closer to the true value, and an accurate position of the measurement object can be calculated.
- the number of sampling M is greater than 10, a period of the projected image of the projector 102 is imaged by many CCD pixels. Therefore, the greater the value of M, the smaller the error amount of each intersection distance even if Expression (1) is not satisfied. However, if the number of sampling M is 10 or less, the errors of the intersection distances are large unless the condition of Expression (1) or (2) is satisfied, and the values are out of the true value. Therefore, the position detection accuracy can be improved, and the three dimensional measurement can be accurately performed by constituting the projection apparatus and the imaging apparatus so that M of Expression (1) or (2) satisfies 10 or less.
- the projection apparatus and the imaging apparatus are constituted so that the width of a period of the bright and dark pattern lights projected to the measurement target object is set to image the measurement target object by about an even number of pixels (the number of sampling) of the imaging element of the imaging unit.
- the tone values are acquired from the digital image data obtained by the imaging apparatus imaging the bright and dark pattern lights
- the tone values are sampled by the pixels of the imaging element.
- the tone values acquired from the images of the first bright and dark pattern light and the second bright and dark pattern light are used to detect the boundaries of the bright and dark pattern lights, sampling errors caused by the straight line approximation of the sampling are generated at the detected boundary positions. As a result, there are errors in the intersection distances if the intersection distances are calculated from the intersections.
- intersection distances can be close to the true value, and accurate three dimensional measurement is possible in the present embodiment by setting the width of a period of the bright and dark pattern lights projected to the measurement target object close to an even multiple of the width for imaging the measurement target object by one pixel of imaging element.
- the distances between the intersections can be accurately calculated, and accurate three dimensional measurement is possible by setting the number of sampling close to an even number.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Physics & Mathematics (AREA)
- Optics & Photonics (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Length Measuring Devices By Optical Means (AREA)
- Image Processing (AREA)
- Image Analysis (AREA)
Abstract
Description
- 1. Field of the Invention
- The present invention relates to a three dimensional measurement apparatus and a method using a space encoding method for generating spatial modulation by projecting pattern lights including bright parts and dark parts arranged at an arbitrary width.
- 2. Description of the Related Art
- In a well known measurement method in three dimensional measurement, a projection apparatus and an imaging apparatus are arranged in a known relationship. The projection apparatus projects pattern lights to a subject, and the imaging apparatus images the subject. The principle of triangulation is used based on the relationship between the projection apparatus and the imaging apparatus to calculate the distance to the subject. For example, there is a method of three dimensional measurement as in Japanese Patent Laid-Open No. 2007-192608. In the method, a first bright and dark pattern light including bright parts and dark parts alternately arranged at an arbitrary width and a second bright and dark pattern light in which the phase of the first bright and dark pattern light is shifted are projected. Tone distributions of photographed images are acquired, and positions of intersections between the first and second pattern lights on the imaging element are calculated.
- It is fundamental in ranging to calculate the positions on the imaging element of the intersections calculated from the first and second pattern lights. It is more important to correctly calculate intersection distances of the first and second bright and dark pattern lights than calculate absolute values of the intersection positions. This is because although the absolute values of the intersections are necessary to calculate the absolute position of the specimen, the shape of the specimen is mainly measured in the three dimensional measurement, and in that case, it is only necessary to accurately calculate the relative positional relationship between the intersections. The measurement of the absolute position of the specimen can be attained by preparing a known index of the absolute position and measuring, i.e. calibrating, the relative relationship between the index and the specimen.
- In a general intersection detection method, the tone distributions of the first and second bright and dark pattern lights on the imaging element are sampled by imaging pixels, and straight line approximation is applied to the sampling points of the pattern lights to calculate the intersections. When a period of the patterns projected to the measurement target object is sampled by relatively few imaging element pixels, the intersections include sampling errors due to the straight line approximation if the intersections are calculated by the conventional method. Therefore, there are errors in the intersection distances if the intersection distances are calculated from the intersections.
- If a period of the pattern lights is sampled by many pixels of the imaging element, the errors in the intersections and the errors in the intersection distances can be reduced. However, a high-resolution imaging element is necessary, and this leads to an increase of the apparatus size and cost.
- In view of the problem, an embodiment of the present invention provides a three dimensional measurement apparatus and a method capable of accurate three dimensional measurement even if sampling per period of pattern lights is performed by a small number of pixels of an imaging element.
- According to one aspect of the present invention, there is provided a three dimensional measurement apparatus comprising: a projection unit that projects, to a measurement target object, a first pattern light including alternately arranged bright parts and dark parts and a second pattern light in which a phase of the first pattern light is shifted; an imaging unit that images the measurement target object, on which the first or second pattern light is projected, as image data; and a calculation unit that calculates a three dimensional position by calculating intersection positions of tone values from tone values of the image data of the first pattern light and the image data of the second pattern light obtained by the imaging unit, wherein when a period of repetitions of the bright parts and the dark parts of the first or second pattern light is one period, a range of imaging on the measurement target object surface by one pixel included in the imaging unit is an image distance, and the length of one period of the projected first pattern light or second pattern light on the measurement target object surface is M times the image distance, the projection unit and the imaging unit are arranged to satisfy
-
2×N−0.2≦M≦2×N+0.2, - where N is an integer value equal to or greater than 2.
- Also, according to another aspect of the present invention, there is provided a three dimensional measurement method by a three dimensional measurement apparatus, the apparatus comprising: a projection unit that projects, to a measurement target object, a first pattern light including alternately arranged bright parts and dark parts and a second pattern light in which a phase of the first pattern light is shifted; and an imaging unit that images the measurement target object on which the first or second pattern light is projected, the method comprising: a first acquisition step by a first acquisition unit acquiring first image data by causing the projection unit to project the first pattern light to cause the imaging unit to image the measurement target object; a second acquisition step by a second acquisition unit acquiring second image data by causing the projection unit to project the second pattern light to cause the imaging unit to image the measurement target object; and a calculation step by a calculation unit calculating a three dimensional position by calculating intersection positions of tone values of the first pattern light in the first image data and tone values of the second pattern light in the second image data, wherein in the first and second acquisition steps, when a period of repetitions of the bright parts and the dark parts of the first or second pattern light is one period, a range of imaging on the measurement target object by one pixel included in the imaging unit is an image distance, and the length of one period of the projected first pattern light or second pattern light on the measurement target object surface is M times the image distance, the projection unit and the imaging unit are arranged to satisfy
-
2×N−0.2≦M≦2×N+0.2, - where N is an integer value equal to or greater than 2.
- Further features of the present invention will become apparent from the following description of exemplary embodiments with reference to the attached drawings.
-
FIGS. 1A and 1B are diagrams showing a basic configuration of a three dimensional measurement apparatus according to an embodiment. -
FIG. 2 is a diagram illustrating a relationship between bright and dark pattern lights on a measurement target object surface and pixels of a CCD. -
FIG. 3 is a diagram illustrating tone values acquired from image data obtained by setting the number of pixels of the CCD in a period of the bright and dark pattern lights to an even number. -
FIG. 4 is a diagram showing an example when a condition of the embodiment is not satisfied. -
FIG. 5 is a diagram showing error amounts of intersection distances when a period of the bright and dark pattern lights on the measurement target object surface is sampled by pixels of a CCD camera. - An example of a preferred embodiment of the present invention will now be described with reference to the attached drawings.
-
FIG. 1A is a diagram showing an example of configuration of a three dimensional measurement apparatus according to the embodiment. InFIG. 1A ,reference numeral 101 denotes a measurement target object, 102 denotes a projector as an example of a projection apparatus, and 103 denotes a CCD camera as an example of an imaging apparatus.Reference numeral 105 denotes a computer (hereinafter, PC 105) that inputs an image (pattern to be projected) to the 102, and 106 denotes a computer (hereinafter, PC 106) that acquires tone values of image data imaged by theprojector CCD camera 103 to calculate the position of the measurement target object. The PC 105 and the PC 106 are connected to be able to communicate each other. It is obvious that the PC 105 and the PC 106 can be realized by one computer. -
FIG. 1B shows an example of pattern lights used in the present embodiment. A first pattern light including alternately arranged bright parts and dark parts and a second pattern light in which the phase of the bright parts and the dark parts of the first pattern light is shifted are used in the present embodiment. A first bright anddark pattern light 110 a is an example of the first pattern light, and a second bright anddark pattern light 110 b in which the phase of the first bright anddark pattern light 110 a is shifted is an example of the second pattern light. The first and second pattern lights are not limited to the illustrated examples, and any pattern lights that satisfy the requirements described below can be used. - The PC 105 and the PC 106 calculate the position of the
measurement target object 101. In a first acquisition process, the PC 105 projects the first bright anddark pattern light 110 a to theprojector 102, and the PC 106 causes theCCD camera 103 to image themeasurement target object 101, on which the first bright anddark pattern light 110 a is projected, to acquire first image data. In the following second acquisition process, the PC 105 projects the second bright anddark pattern light 110 b to theprojector 102, and the PC 106 causes theCCD camera 103 to image themeasurement target object 101, on which the second bright anddark pattern light 110 b is projected, to acquire second image data. - In this way, the
CCD camera 103 images a luminance distribution on themeasurement target object 101, on which the first bright anddark pattern light 110 a and the second bright anddark pattern light 110 b are projected through theprojector 102, and digital image data is imported to the PC 106. Hereinafter, values of the image data of the first and second bright and dark pattern lights will be called tone values. The three dimensional measurement apparatus calculates the intersection positions of the tone values as described below based on the tone values of the image data (first image data) of the first bright and dark pattern light and the tone values of the image data (second image data) of the second bright and dark pattern light obtained by theCCD camera 103. A three dimensional position is calculated based on the calculated intersection positions. - If the tone values of the first and second bright and
110 a and 110 b are connected along a CCD pixel column in the first and second image data, lines connecting the tones intersect as described later indark pattern lights FIG. 3 . It is preferable that the direction along the CCD pixel column is a direction orthogonal to the bright and dark alignment in the bright and dark pattern light. -
FIG. 2 is an expanded view of part on themeasurement target object 101 shown inFIG. 1A and illustrates the relationship between the bright and dark pattern light projected to the measurement target object and the CCD pixels. InFIG. 2 , solid lines show ranges imaged by the pixels of theCCD camera 103 on the measurement target object. If a range imaged by one pixel of theCCD camera 103 on themeasurement target object 101 is an image distance S, and a period (distance) of the bright and dark pattern light projected on themeasurement target object 101 is P, P=M×S (M>0) is satisfied. In the present embodiment, theprojector 102 and theCCD camera 103 are constituted so that M=4 (even), i.e. P=4×S, is satisfied. -
FIG. 3 shows a state after imaging under the first bright and dark pattern light 110 a and the second bright and dark pattern light 110 b, acquiring the tone values from the image data, and detecting boundaries of the tones from the first and second bright and dark pattern lights. Curves V and W denote intensity distributions on the imaging element of the first and second bright and dark pattern lights projected to themeasurement target object 101, and D (white rectangles) denotes a sequence of intersections between the intensity distributions V and W. Points A (black circles) and points B (black triangles) denote point sequences of tone distributions obtained by sampling the intensity distributions V and W by imaging pixels, respectively. - Points C (black stars) denote a point sequence formed by intersections between a line obtained by connecting the point sequence A by a straight line and a line obtained by connecting the point sequence B by a straight line. In the present embodiment, since a period of the bright and dark pattern lights is sampled by even pixels, i.e. four pixels, of the imaging element, the true value of the intersection distance is half the four pixels, or 2.0. Even if the points D and C are different, the intersection distance obtained from the point C is the true value, or 2.0, as a result of sampling a period of the bright and dark pattern lights by even pixels of the imaging element as in the present embodiment.
- The error amount of each intersection distance can be reduced by forming a period of the bright and dark pattern lights on the surface of the
measurement target object 101 to be M times (M is even) the image distance on the surface of themeasurement target object 101. Although it is difficult to accurately set the value of M to be an integer in the implementation, accurate detection of intersection positions is possible if M is an even value ±0.2. More specifically, when the number of sampling is M, the intersection distances can be close to the true value by arranging themeasurement target object 101, theprojector 102, and theCCD camera 103 to satisfy -
2×N−0.2≦M≦2×N+0.2 (N: integer value equal to or greater than 2) Expression (1) - The accurate calculation of the intersection distances allows accurate detection of the intersection positions, and highly accurate three dimensional measurement can be further realized.
-
FIG. 4 shows an example in which a period of the bright and dark pattern lights on the measurement target object surface is not an even multiple of the image distance on the measurement target object surface. In the example ofFIG. 4 , M=4.5, and a period of the bright and dark pattern lights projected from theprojector 102 is 4.5 times the image distance S (P=4.5×S) of theCCD camera 103. In this way, if the intersection distances are calculated from the tone values sampled by imaging when the condition of Expression (1) is not satisfied, the intersection distances are significantly out of the true value, and it is difficult to accurately detect the intersection positions. -
FIG. 5 shows the result of the calculation of error amounts of the intersection distances when a period of the bright and dark pattern lights projected to the measurement target object is sampled by M pixels of the imaging element. The abscissa denotes the number of sampling (M) per period of the bright and dark pattern lights, and the ordinate denotes error amounts (unit: %) generated, by sampling, from the intersection distances calculated from the three dimensional measurement relative to the width of a period of the bright and dark pattern lights. Values close to the true value can be acquired for the intersection distances if the number of sampling M satisfies Expression (1), and the three dimensional measurement can be accurately detected. If the number of sampling M is out of Expression (1), the acquired intersection distances are out of the true value, and it is difficult to accurately detect the three dimensional measurement. - Particularly, if the intersection distances between adjacent intersections are calculated from the three dimensional measurement when the number of sampling M satisfies
-
2×N−0.1≦M≦2×N+0.1 (N: integer value equal to or greater than 2), Expression (2) - the intersection distances calculated at each section are closer to the true value, and an accurate position of the measurement object can be calculated.
- If the number of sampling M is greater than 10, a period of the projected image of the
projector 102 is imaged by many CCD pixels. Therefore, the greater the value of M, the smaller the error amount of each intersection distance even if Expression (1) is not satisfied. However, if the number of sampling M is 10 or less, the errors of the intersection distances are large unless the condition of Expression (1) or (2) is satisfied, and the values are out of the true value. Therefore, the position detection accuracy can be improved, and the three dimensional measurement can be accurately performed by constituting the projection apparatus and the imaging apparatus so that M of Expression (1) or (2) satisfies 10 or less. - As described, according to the present embodiment, the projection apparatus and the imaging apparatus are constituted so that the width of a period of the bright and dark pattern lights projected to the measurement target object is set to image the measurement target object by about an even number of pixels (the number of sampling) of the imaging element of the imaging unit.
- When the tone values are acquired from the digital image data obtained by the imaging apparatus imaging the bright and dark pattern lights, the tone values are sampled by the pixels of the imaging element. When the tone values acquired from the images of the first bright and dark pattern light and the second bright and dark pattern light are used to detect the boundaries of the bright and dark pattern lights, sampling errors caused by the straight line approximation of the sampling are generated at the detected boundary positions. As a result, there are errors in the intersection distances if the intersection distances are calculated from the intersections. Therefore, the intersection distances can be close to the true value, and accurate three dimensional measurement is possible in the present embodiment by setting the width of a period of the bright and dark pattern lights projected to the measurement target object close to an even multiple of the width for imaging the measurement target object by one pixel of imaging element.
- According to the present invention, the distances between the intersections can be accurately calculated, and accurate three dimensional measurement is possible by setting the number of sampling close to an even number.
- While the present invention has been described with reference to exemplary embodiments, it is to be understood that the invention is not limited to the disclosed exemplary embodiments. The scope of the following claims is to be accorded the broadest interpretation so as to encompass all such modifications.
- This application claims the benefit of Japanese Patent Application No. 2010-283730, filed Dec. 20, 2010, which is hereby incorporated by reference herein in its entirety.
Claims (4)
2×N−0.1≦M≦2×N+0.1.
2×N−0.2≦M≦2×N+0.2,
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2010-283730 | 2010-12-20 | ||
| JP2010283730A JP5611022B2 (en) | 2010-12-20 | 2010-12-20 | Three-dimensional measuring apparatus and three-dimensional measuring method |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20120154540A1 true US20120154540A1 (en) | 2012-06-21 |
| US8708497B2 US8708497B2 (en) | 2014-04-29 |
Family
ID=46233863
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US13/314,577 Expired - Fee Related US8708497B2 (en) | 2010-12-20 | 2011-12-08 | Three dimensional measurement apparatus and three dimensional measurement method |
Country Status (2)
| Country | Link |
|---|---|
| US (1) | US8708497B2 (en) |
| JP (1) | JP5611022B2 (en) |
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US8810801B2 (en) | 2011-05-10 | 2014-08-19 | Canon Kabushiki Kaisha | Three-dimensional measurement apparatus, method for controlling a three-dimensional measurement apparatus, and storage medium |
| US20150204662A1 (en) * | 2014-01-17 | 2015-07-23 | Canon Kabushiki Kaisha | Three-dimensional-shape measurement apparatus, three-dimensional-shape measurement method, and non-transitory computer-readable storage medium |
| US9332244B2 (en) | 2011-12-06 | 2016-05-03 | Canon Kabushiki Kaisha | Information processing apparatus, control method for the same and storage medium |
| KR101795690B1 (en) | 2016-09-29 | 2017-11-08 | 이화여자대학교 산학협력단 | Method for high speed computation of projection image |
Families Citing this family (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP6161276B2 (en) | 2012-12-12 | 2017-07-12 | キヤノン株式会社 | Measuring apparatus, measuring method, and program |
| US9881235B1 (en) | 2014-11-21 | 2018-01-30 | Mahmoud Narimanzadeh | System, apparatus, and method for determining physical dimensions in digital images |
| KR102464368B1 (en) * | 2017-11-07 | 2022-11-07 | 삼성전자주식회사 | Meta projector and electronic apparatus including the same |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20070025612A1 (en) * | 2004-03-31 | 2007-02-01 | Brother Kogyo Kabushiki Kaisha | Image input-and-output apparatus |
| US20070064246A1 (en) * | 2003-09-22 | 2007-03-22 | Bernhard Braunecker | Method and system for determining the spatial position of a hand-held measuring appliance |
| US20080112054A1 (en) * | 2006-11-10 | 2008-05-15 | Fuji Xerox Co., Ltd. | Longitudinal interference fringe pattern projection lens, optical system, and three-dimensional image acquisition apparatus |
| US7388679B2 (en) * | 2005-09-21 | 2008-06-17 | Omron Corporation | Pattern light irradiation device, three dimensional shape measuring device, and method pattern light irradiation |
Family Cites Families (8)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPH0264403A (en) * | 1988-08-31 | 1990-03-05 | Fanuc Ltd | Forming method for slit light image |
| JPH0656616B2 (en) * | 1988-10-29 | 1994-07-27 | オージー情報システム株式会社 | Binarization method of captured image information |
| JP3370418B2 (en) * | 1994-03-18 | 2003-01-27 | 株式会社資生堂 | 3D shape measurement system |
| JP3411829B2 (en) * | 1997-07-02 | 2003-06-03 | 旭硝子株式会社 | Method and apparatus for evaluating surface shape |
| JP3871963B2 (en) * | 2002-05-16 | 2007-01-24 | 住友化学株式会社 | Surface inspection apparatus and surface inspection method |
| JP4874657B2 (en) | 2006-01-18 | 2012-02-15 | ローランドディー.ジー.株式会社 | Three-dimensional shape measuring method and apparatus |
| JP4931728B2 (en) * | 2007-08-08 | 2012-05-16 | シーケーディ株式会社 | 3D measuring device and board inspection machine |
| US9007602B2 (en) * | 2010-10-12 | 2015-04-14 | Canon Kabushiki Kaisha | Three-dimensional measurement apparatus, three-dimensional measurement method, and computer-readable medium storing control program |
-
2010
- 2010-12-20 JP JP2010283730A patent/JP5611022B2/en not_active Expired - Fee Related
-
2011
- 2011-12-08 US US13/314,577 patent/US8708497B2/en not_active Expired - Fee Related
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20070064246A1 (en) * | 2003-09-22 | 2007-03-22 | Bernhard Braunecker | Method and system for determining the spatial position of a hand-held measuring appliance |
| US20070025612A1 (en) * | 2004-03-31 | 2007-02-01 | Brother Kogyo Kabushiki Kaisha | Image input-and-output apparatus |
| US7388679B2 (en) * | 2005-09-21 | 2008-06-17 | Omron Corporation | Pattern light irradiation device, three dimensional shape measuring device, and method pattern light irradiation |
| US20080112054A1 (en) * | 2006-11-10 | 2008-05-15 | Fuji Xerox Co., Ltd. | Longitudinal interference fringe pattern projection lens, optical system, and three-dimensional image acquisition apparatus |
Cited By (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US8810801B2 (en) | 2011-05-10 | 2014-08-19 | Canon Kabushiki Kaisha | Three-dimensional measurement apparatus, method for controlling a three-dimensional measurement apparatus, and storage medium |
| US9332244B2 (en) | 2011-12-06 | 2016-05-03 | Canon Kabushiki Kaisha | Information processing apparatus, control method for the same and storage medium |
| US9998728B2 (en) | 2011-12-06 | 2018-06-12 | Canon Kabushiki Kaisha | Information processing apparatus, control method for the same and storage medium |
| US20150204662A1 (en) * | 2014-01-17 | 2015-07-23 | Canon Kabushiki Kaisha | Three-dimensional-shape measurement apparatus, three-dimensional-shape measurement method, and non-transitory computer-readable storage medium |
| US9557167B2 (en) * | 2014-01-17 | 2017-01-31 | Canon Kabushiki Kaisha | Three-dimensional-shape measurement apparatus, three-dimensional-shape measurement method, and non-transitory computer-readable storage medium |
| GB2522551B (en) * | 2014-01-17 | 2018-06-27 | Canon Kk | Three-dimensional-shape measurement apparatus, three-dimensional-shape measurement method, and non-transitory computer-readable storage medium |
| KR101795690B1 (en) | 2016-09-29 | 2017-11-08 | 이화여자대학교 산학협력단 | Method for high speed computation of projection image |
Also Published As
| Publication number | Publication date |
|---|---|
| JP5611022B2 (en) | 2014-10-22 |
| US8708497B2 (en) | 2014-04-29 |
| JP2012132727A (en) | 2012-07-12 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US8708497B2 (en) | Three dimensional measurement apparatus and three dimensional measurement method | |
| EP2588836B1 (en) | Three-dimensional measurement apparatus, three-dimensional measurement method, and storage medium | |
| US8345954B2 (en) | Method and apparatus for the three-dimensional digitization of objects | |
| CN101308012B (en) | Double monocular white light three-dimensional measuring systems calibration method | |
| JP6170281B2 (en) | Three-dimensional measuring device, control method for three-dimensional measuring device, and program | |
| US20100201809A1 (en) | Calibration method, calibration device, and calibration system including the device | |
| JP2011017700A (en) | Method of determining three-dimensional coordinate of object | |
| EP3405746B1 (en) | Imaging unit and system for obtaining a three-dimensional image | |
| US12072177B2 (en) | Three-dimensional shape measuring apparatus, three-dimensional shape measuring method, program, and storage medium | |
| CN100381780C (en) | Tilt angle detection device and tilt angle detection method | |
| JP2018087732A (en) | Wire rope measurement device and method | |
| JP4797109B2 (en) | Three-dimensional shape measuring apparatus and three-dimensional shape measuring method | |
| JP7011093B1 (en) | Displacement measuring device and displacement measuring method | |
| JP4516949B2 (en) | Three-dimensional shape measuring apparatus and three-dimensional shape measuring method | |
| KR101750883B1 (en) | Method for 3D Shape Measuring OF Vision Inspection System | |
| EP3070432B1 (en) | Measurement apparatus | |
| JP2008170282A (en) | Shape measuring device | |
| TWI837061B (en) | System and method for 3d profile measurements using color fringe projection techniques | |
| JP2014194436A (en) | Three-dimensional measurement device, three-dimensional measurement method and program | |
| JP5226250B2 (en) | 3D shape measurement system and 3D shape measurement method | |
| US20020089673A1 (en) | 3-D measurements of improved accuracy | |
| JP2006330772A (en) | Distortion correction method for captured images | |
| JP6532158B2 (en) | Surface shape strain measuring apparatus and method of measuring surface shape strain | |
| Tsujioka et al. | High-density sampling technique using special projections of light variation patterns in three-dimensional shape identification | |
| JP2011153922A (en) | Shape measurement apparatus and shape measurement method |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: CANON KABUSHIKI KAISHA, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:TSUYUKI, MOTOMI;REEL/FRAME:027930/0842 Effective date: 20111206 Owner name: CANON KABUSHIKI KAISHA, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:TSUYUKI, MOTOMI;REEL/FRAME:027930/0919 Effective date: 20111206 |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551) Year of fee payment: 4 |
|
| FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
| FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20220429 |