US20240155210A1 - Data generation apparatus and control method - Google Patents
Data generation apparatus and control method Download PDFInfo
- Publication number
- US20240155210A1 US20240155210A1 US18/491,843 US202318491843A US2024155210A1 US 20240155210 A1 US20240155210 A1 US 20240155210A1 US 202318491843 A US202318491843 A US 202318491843A US 2024155210 A1 US2024155210 A1 US 2024155210A1
- Authority
- US
- United States
- Prior art keywords
- light image
- image
- unit
- invisible light
- subject
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/10—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths
- H04N23/11—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths for generating image signals from visible and infrared light wavelengths
-
- G06T5/002—
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/70—Denoising; Smoothing
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/60—Extraction of image or video features relating to illumination properties, e.g. using a reflectance or lighting model
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/07—Target detection
Definitions
- the present invention relates to techniques for generating supervisory data to be used in learning of a learning model for subject detection processing of an invisible light image.
- Japanese Patent Laid-Open No. 2019-118043 discloses a method of performing subject detection processing of a visible light image and an invisible light image using machine learning.
- a large amount of supervisory data is necessary for learning processing of a learning model for subject detection processing of visible light images; conventionally, learning models for visible light images are the mainstream of learning models for subject detection processing; it takes a lot of effort and man-hours to generate a large amount of supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image.
- the present invention has been made in consideration of the aforementioned problems and realizes techniques for allowing efficient and accurate generation of supervisory data to be used in learning of a learning model for detecting a subject in an invisible light image.
- the present invention provides a data generation apparatus comprising: a first image acquiring unit that acquires a visible light image; a second image acquiring unit that acquires an invisible light image; a subject detection unit that detects a subject in the visible light image; and a supervisory data generation unit that generates supervisory data to be used in learning of a learning model for subject detection of an invisible light image based on the invisible light image and a subject detection result acquired by the subject detection unit.
- the present invention provides a method of controlling a data generation apparatus, the method comprising: acquiring a visible light image and an invisible light image; detecting a subject in the visible light image; and generating supervisory data to be used in learning of a learning model for subject detection of an invisible light image based on the invisible light image and the subject detection result.
- the present invention provides a non-transitory computer-readable storage medium storing a program for causing a computer to function as a data generation apparatus comprising: a first image acquiring unit that acquires a visible light image; a second image acquiring unit that acquires an invisible light image; a subject detection unit that detects a subject in the visible light image; and a supervisory data generation unit that generates supervisory data to be used in learning of a learning model for subject detection of an invisible light image based on the invisible light image and a subject detection result acquired by the subject detection unit.
- supervisory data to be used in learning of a learning model for subject detection of an invisible light image can be efficiently and accurately generated.
- FIG. 1 is a block diagram illustrating an apparatus configuration according to a first embodiment.
- FIG. 2 is a diagram illustrating a relationship among a visible light image, an invisible light image, and a detection result of a subject detected in the visible light image according to the first embodiment.
- FIG. 3 is a diagram illustrating image processing for generating supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image according to the first embodiment.
- FIG. 4 is a diagram illustrating an example of display of a visible light image, an invisible light image, and a subject detection result according to the first embodiment.
- FIG. 5 is a diagram illustrating an example of display of a generation result of supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image according to the first embodiment.
- FIG. 6 is a flowchart exemplifying processing for generating supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image according to the first embodiment.
- FIG. 7 is a diagram of a system configuration according to a second embodiment.
- FIG. 8 is a flowchart exemplifying processing for generating supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image according to the second embodiment.
- An image capture apparatus 100 of the first embodiment captures a visible light image and an invisible light image and generates supervisory data for a learning model for subject detection processing of an invisible light image based on the invisible light image and a subject detection result (type, position, and size of the subject) of the visible light image.
- the supervisory data of the present embodiment is data to be used in machine learning of a learning model for subject detection of an invisible light image, and learning processing is executed with supervisory data as input data (an invisible light image) and output data (a detection result of a subject detected from a visible light image).
- FIG. 1 is a block diagram illustrating a configuration of the image capture apparatus 100 according to the first embodiment.
- the image capture apparatus 100 includes a first optical system 101 a , a second optical system 101 b , a first image capturing unit 102 a , a second image capturing unit 102 b , a first image processing unit 103 a , a second image processing unit 103 b , a subject detection unit 104 , a supervisory data generation determination unit 105 , a supervisory data generation unit 106 , a supervisory data storage unit 107 , a learning unit 108 , a display unit 109 , a memory 110 , and a control unit 111 .
- the first optical system 101 a includes one or more lenses and forms an image of subject image light in a visible light range on the first image capturing unit 102 a .
- the second optical system 101 b includes one or more lenses and forms an image of subject image light in an invisible light range on the second image capturing unit 102 b .
- An image in a visible light range is, for example, an image in a wavelength range of 400 to 800 nm.
- An image in an invisible light range is, for example, a near-infrared image in a near-infrared wavelength range of 800 to 2500 nm but may be a far-infrared image in a far-infrared wavelength range of 4 ⁇ m to 1000 ⁇ m or an ultraviolet image in an ultraviolet wavelength range of 380 nm or less.
- the first image capturing unit 102 a includes an image sensor, such as a CMOS sensor, for converting subject image light of a visible light range formed into an image by the first optical system 101 a into an electric signal.
- the image sensor includes, for example, color filters in an RGB Bayer array.
- the first image capturing unit 102 a includes an AD converter for converting an analog electric signal into a digital signal.
- the second image capturing unit 102 b includes an image sensor, such as a CMOS, for converting subject image light of an invisible light range formed into an image by the second optical system 101 b into an electric signal.
- the image sensor includes, for example, a color filter that passes light of a near-infrared wavelength range.
- the second image capturing unit 102 b includes an AD converter for converting an analog electric signal into a digital signal.
- the first image processing unit 103 a includes a processor (GPU) for executing predetermined image processing on a digital signal obtained by the first image capturing unit 102 a and generates visible light image data.
- the predetermined image processing includes, for example, distortion correction processing, noise removal processing, exposure correction processing, white balance processing, and edge enhancement processing.
- the second image processing unit 103 b includes a processor (GPU) for executing predetermined image processing on a digital signal obtained by the second image capturing unit 102 b and generates invisible light image data.
- the predetermined image processing includes, for example, distortion correction processing, noise removal processing, exposure correction processing, white balance processing, and edge enhancement processing.
- the subject detection unit 104 detects one or more subjects in a visible light image processed by the first image processing unit 103 a . Then, the subject detection unit 104 detects class information indicating the type of the detected subject, position information indicating a position (center coordinates) of the subject, size information indicating a size of the subject, and the like.
- Subject detection can be realized by image analysis processing or image recognition processing in which a learning model for which learning processing has been performed by machine learning is used.
- a learning model is, for example, a neural network, and the class information (such as whether a person is included or a car is included), the position information, and the size information of a subject in an image is detected using a learning model trained with supervisory data.
- the invisible light image supervisory data of the present embodiment is generated based on an invisible light image and the class information, the position information, and the size information of a subject obtained from a visible light image.
- FIG. 2 is a diagram illustrating a relationship among a visible light image, an invisible light image, and a detection result of a subject detected in the visible light image according to the present embodiment.
- a visible light image 201 is an image that has been processed by the first image processing unit 103 a
- the subject detection unit 104 detects a subject 202 for which a subject region 203 has been surrounded by a rectangular frame.
- a subject detection result 204 includes the class information, the size information (X, Y), and the position information (center coordinates) of the subject.
- An invisible light image 205 is an image that has been processed by the second image processing unit 103 b and is an image that has been captured at the same angle of view as that of the visible light image 201 .
- a subject 206 is a subject that corresponds to the subject 202 of the visible light image 201 .
- a subject region 207 and a subject detection result 208 are the same as the subject region 203 and the subject detection result 204 of the visible light image.
- the supervisory data generation determination unit 105 determines whether invisible light image supervisory data can be generated based on a subject detection result obtained by the subject detection unit 104 and an invisible light image obtained by the second image processing unit 103 b .
- a determination method for example, a histogram of luminance values of an invisible light image in the same region as a subject region of a visible light image obtained by the subject detection unit 104 is generated, and when the luminance values are within a predetermined range, it is determined that supervisory data can be generated and when the luminance values are outside of the predetermined range, it is determined that supervisory data cannot be generated.
- a histogram generated based on the entire image may be used.
- the supervisory data generation determination unit 105 includes a subject type designation unit (not illustrated) and can determine whether supervisory data for a pre-designated subject can be generated by a subject type for which to generate supervisory data being set by a user operation or a subject type for which to generate supervisory data being selected by a user operation in superimposed display of an invisible light image and a subject detection result, which will be described later in FIG. 5 .
- the supervisory data generation unit 106 When it is determined by the supervisory data generation determination unit 105 that supervisory data can be generated, the supervisory data generation unit 106 generates invisible light image supervisory data based on an invisible light image and a subject detection result of the subject detection unit 104 .
- the supervisory data generation unit 106 includes an image processing unit (not illustrated) and can generate a plurality of pieces of supervisory data from a single invisible light image by executing specific image processing on the invisible light image.
- the specific image processing includes, for example, at least one of processing for creating bokeh, processing of creating blurring, and processing for correcting luminance.
- FIG. 3 illustrates processing 301 for correcting luminance, processing 302 for creating blurring, and processing 303 for creating bokeh as examples of specific image processing to be performed on an invisible light image.
- processing 301 for correcting luminance processing 302 for creating blurring
- processing 303 for creating bokeh processing 301 for correcting luminance
- processing 302 for creating blurring processing 302 for creating blurring
- processing 303 for creating bokeh processing 301 for correcting luminance
- processing 302 for creating blurring processing 303 for creating bokeh
- the supervisory data storage unit 107 stores invisible light image supervisory data generated by the supervisory data generation unit 106 .
- the learning unit 108 performs learning processing by machine learning on a learning model for subject detection processing of an invisible light image using invisible light image supervisory data stored in the supervisory data storage unit 107 .
- the learning processing of the learning unit 108 is executed with the supervisory data as input data and output data of the learning model for subject detection processing of an invisible light image.
- a graphics processing unit GPU can perform efficient computation by processing more data in parallel processing; therefore, it is useful to perform processing using the GPU when performing learning processing a plurality of times using a learning model, such as in machine learning.
- computation may be performed by the control unit 111 , which will be described later, and the GPU cooperating with each other or computation may be performed by the control unit 111 or the GPU alone.
- a configuration is taken such that the image capture apparatus 100 includes the learning unit 108 ; however, the learning unit 108 may be configured to be separate from the image capture apparatus 100 .
- the display unit 109 displays a subject detection result 401 obtained by the subject detection unit 104 in a superimposed manner on one or both of the visible light image obtained by the first image processing unit 103 a and the invisible light image obtained by the second image processing unit 103 b .
- the displayed subject detection result 401 is a rectangular frame surrounding the subject region.
- the display unit 109 displays the invisible light image supervisory data obtained by the supervisory data generation unit 106 for preview.
- the class information, the size information, and the position information of the subject are displayed as invisible light image supervisory data 501 in a superimposed manner on the visible light image or the invisible light image for preview.
- the user can change the class information of the subject, which serves as the supervisory data, and change the determination result as to whether supervisory data can be generated by an operation on the display screen.
- the memory 110 includes a non-volatile memory (ROM), a volatile memory (RAM), and the like and stores a control program for controlling the overall operation of the image capture apparatus 100 and various parameters.
- ROM non-volatile memory
- RAM volatile memory
- the control unit 111 includes a processor (Central Processing Unit: CPU) for controlling the entire operation of the image capture apparatus 100 by executing the program stored in the memory 110 .
- CPU Central Processing Unit
- FIG. 6 is a flowchart illustrating processing for generating invisible light image supervisory data the present embodiment.
- FIG. 6 The processing of FIG. 6 is realized by the control unit 111 of the image capture apparatus 100 illustrated in FIG. 1 controlling respective components by executing the program stored in the memory 110 .
- step S 601 the control unit 111 generates a visible light image and an invisible light image of the same angle of view by capturing the visible light image and the invisible light image simultaneously or consecutively using the first image capturing unit 102 a and the second image capturing unit 102 b . Further, the control unit 111 executes predetermined image processing on the visible light image and the invisible light image using the first image processing unit 103 a and the second image processing unit 103 b.
- step S 602 the control unit 111 performs subject detection processing on the visible light image obtained in step S 601 using the subject detection unit 104 . Then, the control unit 111 determines whether a subject has been detected and advances the processing to step S 603 when the control unit 111 determines that a subject has been detected and terminates the processing when the control unit 111 determines that a subject has not been detected.
- step S 603 the control unit 111 determines whether invisible light image supervisory data can be generated by the supervisory data generation determination unit 105 based on the subject detection result of step S 602 and the invisible light image obtained in step S 601 .
- the control unit 111 advances the processing to step S 604 when the control unit 111 determines that the supervisory data can be generated and terminates the processing when the control unit 111 determines that the supervisory data cannot be generated.
- step S 604 the control unit 111 generates invisible light image supervisory data using the supervisory data generation unit 106 based on the subject detection result of step S 602 and the invisible light image obtained in step S 601 .
- supervisory data to be used for machine learning of a learning model for subject detection of an invisible light image can be efficiently and accurately generated based on an invisible light image and a detection result of a subject in a visible light image. This makes it possible to generate a large amount of supervisory data for machine learning of a learning model for subject detection processing of an invisible light image.
- FIG. 7 is a block diagram illustrating a configuration of a system 700 according to the second embodiment.
- the system 700 of the second embodiment includes the image capture apparatus 710 and the supervisory data generation apparatus 720 .
- the image capture apparatus 710 includes a first optical system 711 a , a second optical system 711 b , a first image capturing unit 712 a , a second image capturing unit 712 b , a first image processing unit 713 a , a second image processing unit 713 b , a memory 714 , and a control unit 715 .
- the image capture apparatus 710 is, for example, a monitor camera or a fixed-point camera capable of simultaneously or consecutively capturing a visible light image and an invisible light image of the same angle of view.
- the image capture apparatus 710 is capable of capturing a visible light image and an invisible light image at a specific timing, continuously or at regular intervals.
- the first optical system 711 a , the second optical system 711 b , the first image capturing unit 712 a , the second image capturing unit 712 b , the first image processing unit 713 a , and the second image processing unit 713 b are similar to the first optical system 101 a , the second optical system 101 b , the first image capturing unit 102 a the second image capturing unit 102 b , the first image processing unit 103 a , and the second image processing unit 103 b of the first embodiment.
- the memory 714 includes a non-volatile memory (ROM), a volatile memory (RAM), and the like and stores a control program for controlling the overall operation of the image capture apparatus 710 and various parameters.
- ROM non-volatile memory
- RAM volatile memory
- the control unit 715 includes a processor (CPU) for controlling the entire operation of the image capture apparatus 710 by executing the program stored in the memory 714 .
- a processor CPU
- the supervisory data generation apparatus 720 includes a first image acquiring unit 721 a , a second image acquiring unit 721 b , a subject detection unit 722 , a supervisory data generation determination unit 723 , a supervisory data generation unit 724 , a supervisory data storage unit 725 , a learning unit 726 , a display unit 727 , a memory 728 , and a control unit 729 .
- the supervisory data generation apparatus 720 is, for example, a server connected to the image capture apparatus 710 via a network so as to be capable of communication.
- the first image acquiring unit 721 a acquires a visible light image from the image capture apparatus 710 .
- the second image acquiring unit 721 b acquires an invisible light image from the image capture apparatus 710 .
- the visible light image and the invisible light image are images of the same angle of view captured simultaneously or consecutively.
- the subject detection unit 722 , the supervisory data generation determination unit 723 , the supervisory data generation unit 724 , the supervisory data storage unit 725 , and the learning unit 726 are similar to the subject detection unit 104 , the supervisory data generation determination unit 105 , the supervisory data generation unit 106 , the supervisory data storage unit 107 , and the learning unit 108 of the first embodiment.
- the display unit 727 displays a subject detection result obtained by the subject detection unit 722 in a superimposed manner on one or both of the visible light image obtained by the first image acquiring unit 721 a and the invisible light image obtained by the second image acquiring unit 721 b . Display contents are similar to those in FIG. 5 of the first embodiment.
- the memory 728 includes a non-volatile memory (ROM), a volatile memory (RAM), and the like and stores a control program for controlling the overall operation of the supervisory data generation apparatus 720 and various parameters.
- ROM non-volatile memory
- RAM volatile memory
- the control unit 729 includes a processor (CPU) for controlling the entire operation of the supervisory data generation apparatus 720 by executing the program stored in the memory 728 .
- a processor CPU for controlling the entire operation of the supervisory data generation apparatus 720 by executing the program stored in the memory 728 .
- FIG. 8 is a flowchart exemplifying supervisory data generation processing for generating supervisory data of an invisible light image according to the second embodiment.
- FIG. 8 The processing of FIG. 8 is realized by the control unit 715 of the image capture apparatus 710 and the control unit 729 of the supervisory data generation apparatus 720 of the system 700 illustrated in FIG. 7 controlling respective components. In the processing of FIG. 8 , it is assumed that the image capture apparatus 710 and the supervisory data generation apparatus 720 are connected via a network so as to be capable of communication.
- step S 801 the control unit 715 of the image capture apparatus 710 generates a visible light image and an invisible light image of the same angle of view by simultaneously or consecutively capturing the visible light image and the invisible light image using the image capture apparatus 710 and executes predetermined image processing on the visible light image and the invisible light image.
- step S 802 the control unit 715 of the image capture apparatus 710 transmits the visible light image and the invisible light image obtained in step S 801 to the supervisory data generation apparatus 720 .
- step S 803 the control unit 729 of the supervisory data generation apparatus 720 acquires the visible light image and the invisible light image transmitted from the image capture apparatus 710 in step S 802 using the first image acquiring unit 721 a and the second image acquiring unit 721 b.
- step S 804 the control unit 729 of the supervisory data generation apparatus 720 performs subject detection processing on the visible light image obtained in step S 803 using the subject detection unit 722 .
- the control unit 729 of the supervisory data generation apparatus 720 determines whether a subject has been detected by the subject detection unit 722 and advances the processing to step S 805 when the control unit 111 determines that a subject has been detected and terminates the processing when the control unit 111 determines that a subject has not been detected.
- step S 805 the control unit 729 of the supervisory data generation apparatus 720 determines whether invisible light image supervisory data can be generated using the supervisory data generation determination unit 723 and advances the processing to step S 806 when the control unit 111 determines that the supervisory data can be generated and terminates the processing when the control unit 111 determines that the supervisory data cannot be generated.
- step S 806 the control unit 729 of the supervisory data generation apparatus 720 generates invisible light image supervisory data based the subject detection result obtained in step S 804 and the invisible light image acquired in step S 803 using the supervisory data generation unit 724 .
- the image capture apparatus 710 and the supervisory data generation apparatus 720 are separate, it is possible to contribute to versatile and efficient generation of supervisory data, such as setting the supervisory data generation apparatus 720 to be a server.
- a configuration is taken such that the supervisory data generation apparatus 720 acquires a visible light image and an invisible light image generated by a single image capture apparatus 710 ; however, the supervisory data generation apparatus 720 may acquire a visible light image and an invisible light image of the same angle of view generated by a plurality of image capture apparatuses. For example, the supervisory data generation apparatus 720 may acquire a visible light image and an invisible light image of the same angle of view generated by a plurality of monitoring cameras or the like installed at different positions.
- Embodiment(s) of the present invention can also be realized by a computer of a system or apparatus that reads out and executes computer executable instructions (e.g., one or more programs) recorded on a storage medium (which may also be referred to more fully as a ‘non-transitory computer-readable storage medium’) to perform the functions of one or more of the above-described embodiment(s) and/or that includes one or more circuits (e.g., application specific integrated circuit (ASIC)) for performing the functions of one or more of the above-described embodiment(s), and by a method performed by the computer of the system or apparatus by, for example, reading out and executing the computer executable instructions from the storage medium to perform the functions of one or more of the above-described embodiment(s) and/or controlling the one or more circuits to perform the functions of one or more of the above-described embodiment(s).
- computer executable instructions e.g., one or more programs
- a storage medium which may also be referred to more fully as a
- the computer may comprise one or more processors (e.g., central processing unit (CPU), micro processing unit (MPU)) and may include a network of separate computers or separate processors to read out and execute the computer executable instructions.
- the computer executable instructions may be provided to the computer, for example, from a network or the storage medium.
- the storage medium may include, for example, one or more of a hard disk, a random-access memory (RAM), a read only memory (ROM), a storage of distributed computing systems, an optical disk (such as a compact disc (CD), digital versatile disc (DVD), or Blu-ray Disc (BD)TM), a flash memory device, a memory card, and the like.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Software Systems (AREA)
- Signal Processing (AREA)
- Image Analysis (AREA)
Abstract
A data generation apparatus comprises a first image acquiring unit that acquires a visible light image, a second image acquiring unit that acquires an invisible light image, a subject detection unit that detects a subject in the visible light image, and a supervisory data generation unit that generates supervisory data to be used in learning of a learning model for subject detection of an invisible light image based on the invisible light image and a subject detection result acquired by the subject detection unit.
Description
- The present invention relates to techniques for generating supervisory data to be used in learning of a learning model for subject detection processing of an invisible light image.
- A method in which, in an image capture apparatus, such as a digital camera, a subject is detected in an image by image analysis processing or image recognition processing in which a learning model for which learning processing has been performed by machine learning is used is known. Further, Japanese Patent Laid-Open No. 2019-118043 discloses a method of performing subject detection processing of a visible light image and an invisible light image using machine learning.
- A large amount of supervisory data is necessary for learning processing of a learning model for subject detection processing of visible light images; conventionally, learning models for visible light images are the mainstream of learning models for subject detection processing; it takes a lot of effort and man-hours to generate a large amount of supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image.
- The present invention has been made in consideration of the aforementioned problems and realizes techniques for allowing efficient and accurate generation of supervisory data to be used in learning of a learning model for detecting a subject in an invisible light image.
- In order to solve the aforementioned problems, the present invention provides a data generation apparatus comprising: a first image acquiring unit that acquires a visible light image; a second image acquiring unit that acquires an invisible light image; a subject detection unit that detects a subject in the visible light image; and a supervisory data generation unit that generates supervisory data to be used in learning of a learning model for subject detection of an invisible light image based on the invisible light image and a subject detection result acquired by the subject detection unit.
- In order to solve the aforementioned problems, the present invention provides a method of controlling a data generation apparatus, the method comprising: acquiring a visible light image and an invisible light image; detecting a subject in the visible light image; and generating supervisory data to be used in learning of a learning model for subject detection of an invisible light image based on the invisible light image and the subject detection result.
- In order to solve the aforementioned problems, the present invention provides a non-transitory computer-readable storage medium storing a program for causing a computer to function as a data generation apparatus comprising: a first image acquiring unit that acquires a visible light image; a second image acquiring unit that acquires an invisible light image; a subject detection unit that detects a subject in the visible light image; and a supervisory data generation unit that generates supervisory data to be used in learning of a learning model for subject detection of an invisible light image based on the invisible light image and a subject detection result acquired by the subject detection unit.
- According to the present invention, supervisory data to be used in learning of a learning model for subject detection of an invisible light image can be efficiently and accurately generated.
- Further features of the present invention will become apparent from the following description of exemplary embodiments (with reference to the attached drawings).
-
FIG. 1 is a block diagram illustrating an apparatus configuration according to a first embodiment. -
FIG. 2 is a diagram illustrating a relationship among a visible light image, an invisible light image, and a detection result of a subject detected in the visible light image according to the first embodiment. -
FIG. 3 is a diagram illustrating image processing for generating supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image according to the first embodiment. -
FIG. 4 is a diagram illustrating an example of display of a visible light image, an invisible light image, and a subject detection result according to the first embodiment. -
FIG. 5 is a diagram illustrating an example of display of a generation result of supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image according to the first embodiment. -
FIG. 6 is a flowchart exemplifying processing for generating supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image according to the first embodiment. -
FIG. 7 is a diagram of a system configuration according to a second embodiment. -
FIG. 8 is a flowchart exemplifying processing for generating supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image according to the second embodiment. - Hereinafter, embodiments will be described in detail with reference to the attached drawings. Note, the following embodiments are not intended to limit the scope of the claimed invention. Multiple features are described in the embodiments, but limitation is not made to an invention that requires all such features, and multiple such features may be combined as appropriate. Furthermore, in the attached drawings, the same reference numerals are given to the same or similar configurations, and redundant description thereof is omitted.
- An example in which a data generation apparatus has been applied to an image capture apparatus, such as a digital camera, will be described below.
- An
image capture apparatus 100 of the first embodiment captures a visible light image and an invisible light image and generates supervisory data for a learning model for subject detection processing of an invisible light image based on the invisible light image and a subject detection result (type, position, and size of the subject) of the visible light image. The supervisory data of the present embodiment is data to be used in machine learning of a learning model for subject detection of an invisible light image, and learning processing is executed with supervisory data as input data (an invisible light image) and output data (a detection result of a subject detected from a visible light image). -
FIG. 1 is a block diagram illustrating a configuration of theimage capture apparatus 100 according to the first embodiment. - The
image capture apparatus 100 includes a firstoptical system 101 a, a secondoptical system 101 b, a firstimage capturing unit 102 a, a secondimage capturing unit 102 b, a firstimage processing unit 103 a, a secondimage processing unit 103 b, asubject detection unit 104, a supervisory datageneration determination unit 105, a supervisorydata generation unit 106, a supervisorydata storage unit 107, alearning unit 108, adisplay unit 109, amemory 110, and acontrol unit 111. - The first
optical system 101 a includes one or more lenses and forms an image of subject image light in a visible light range on the firstimage capturing unit 102 a. The secondoptical system 101 b includes one or more lenses and forms an image of subject image light in an invisible light range on the secondimage capturing unit 102 b. An image in a visible light range is, for example, an image in a wavelength range of 400 to 800 nm. An image in an invisible light range is, for example, a near-infrared image in a near-infrared wavelength range of 800 to 2500 nm but may be a far-infrared image in a far-infrared wavelength range of 4 μm to 1000 μm or an ultraviolet image in an ultraviolet wavelength range of 380 nm or less. - The first
image capturing unit 102 a includes an image sensor, such as a CMOS sensor, for converting subject image light of a visible light range formed into an image by the firstoptical system 101 a into an electric signal. The image sensor includes, for example, color filters in an RGB Bayer array. The firstimage capturing unit 102 a includes an AD converter for converting an analog electric signal into a digital signal. - The second
image capturing unit 102 b includes an image sensor, such as a CMOS, for converting subject image light of an invisible light range formed into an image by the secondoptical system 101 b into an electric signal. The image sensor includes, for example, a color filter that passes light of a near-infrared wavelength range. The secondimage capturing unit 102 b includes an AD converter for converting an analog electric signal into a digital signal. - The first
image processing unit 103 a includes a processor (GPU) for executing predetermined image processing on a digital signal obtained by the firstimage capturing unit 102 a and generates visible light image data. The predetermined image processing includes, for example, distortion correction processing, noise removal processing, exposure correction processing, white balance processing, and edge enhancement processing. - The second
image processing unit 103 b includes a processor (GPU) for executing predetermined image processing on a digital signal obtained by the secondimage capturing unit 102 b and generates invisible light image data. The predetermined image processing includes, for example, distortion correction processing, noise removal processing, exposure correction processing, white balance processing, and edge enhancement processing. - The
subject detection unit 104 detects one or more subjects in a visible light image processed by the firstimage processing unit 103 a. Then, thesubject detection unit 104 detects class information indicating the type of the detected subject, position information indicating a position (center coordinates) of the subject, size information indicating a size of the subject, and the like. Subject detection can be realized by image analysis processing or image recognition processing in which a learning model for which learning processing has been performed by machine learning is used. A learning model is, for example, a neural network, and the class information (such as whether a person is included or a car is included), the position information, and the size information of a subject in an image is detected using a learning model trained with supervisory data. - Here, processing for generating supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image (hereinafter, abbreviated to invisible light image supervisory data) and processing for learning the learning model using the invisible light image supervisory data of the present embodiment will be described.
- The invisible light image supervisory data of the present embodiment is generated based on an invisible light image and the class information, the position information, and the size information of a subject obtained from a visible light image.
-
FIG. 2 is a diagram illustrating a relationship among a visible light image, an invisible light image, and a detection result of a subject detected in the visible light image according to the present embodiment. In the example ofFIG. 2 , avisible light image 201 is an image that has been processed by the firstimage processing unit 103 a, and thesubject detection unit 104 detects asubject 202 for which asubject region 203 has been surrounded by a rectangular frame. Asubject detection result 204 includes the class information, the size information (X, Y), and the position information (center coordinates) of the subject. - An
invisible light image 205 is an image that has been processed by the secondimage processing unit 103 b and is an image that has been captured at the same angle of view as that of thevisible light image 201. Asubject 206 is a subject that corresponds to thesubject 202 of thevisible light image 201. Asubject region 207 and asubject detection result 208 are the same as thesubject region 203 and thesubject detection result 204 of the visible light image. - The supervisory data
generation determination unit 105 determines whether invisible light image supervisory data can be generated based on a subject detection result obtained by thesubject detection unit 104 and an invisible light image obtained by the secondimage processing unit 103 b. Regarding a determination method, for example, a histogram of luminance values of an invisible light image in the same region as a subject region of a visible light image obtained by thesubject detection unit 104 is generated, and when the luminance values are within a predetermined range, it is determined that supervisory data can be generated and when the luminance values are outside of the predetermined range, it is determined that supervisory data cannot be generated. Regarding the luminance values, a histogram generated based on the entire image may be used. The supervisory datageneration determination unit 105 includes a subject type designation unit (not illustrated) and can determine whether supervisory data for a pre-designated subject can be generated by a subject type for which to generate supervisory data being set by a user operation or a subject type for which to generate supervisory data being selected by a user operation in superimposed display of an invisible light image and a subject detection result, which will be described later inFIG. 5 . - When it is determined by the supervisory data
generation determination unit 105 that supervisory data can be generated, the supervisorydata generation unit 106 generates invisible light image supervisory data based on an invisible light image and a subject detection result of thesubject detection unit 104. The supervisorydata generation unit 106 includes an image processing unit (not illustrated) and can generate a plurality of pieces of supervisory data from a single invisible light image by executing specific image processing on the invisible light image. The specific image processing includes, for example, at least one of processing for creating bokeh, processing of creating blurring, and processing for correcting luminance. -
FIG. 3 illustrates processing 301 for correcting luminance, processing 302 for creating blurring, and processing 303 for creating bokeh as examples of specific image processing to be performed on an invisible light image. In the example ofFIG. 3 , an example in which image processing is performed only on a subject region is illustrated; however, image processing may be performed on the entire image. - The supervisory
data storage unit 107 stores invisible light image supervisory data generated by the supervisorydata generation unit 106. Thelearning unit 108 performs learning processing by machine learning on a learning model for subject detection processing of an invisible light image using invisible light image supervisory data stored in the supervisorydata storage unit 107. The learning processing of thelearning unit 108 is executed with the supervisory data as input data and output data of the learning model for subject detection processing of an invisible light image. A graphics processing unit (GPU) can perform efficient computation by processing more data in parallel processing; therefore, it is useful to perform processing using the GPU when performing learning processing a plurality of times using a learning model, such as in machine learning. In the present embodiment, regarding the learning processing, computation may be performed by thecontrol unit 111, which will be described later, and the GPU cooperating with each other or computation may be performed by thecontrol unit 111 or the GPU alone. In the present embodiment, a configuration is taken such that theimage capture apparatus 100 includes thelearning unit 108; however, thelearning unit 108 may be configured to be separate from theimage capture apparatus 100. - As illustrated in
FIG. 4 , thedisplay unit 109 displays asubject detection result 401 obtained by thesubject detection unit 104 in a superimposed manner on one or both of the visible light image obtained by the firstimage processing unit 103 a and the invisible light image obtained by the secondimage processing unit 103 b. The displayedsubject detection result 401 is a rectangular frame surrounding the subject region. As illustrated inFIG. 5 , thedisplay unit 109 displays the invisible light image supervisory data obtained by the supervisorydata generation unit 106 for preview. In the example ofFIG. 5 , the class information, the size information, and the position information of the subject are displayed as invisible light imagesupervisory data 501 in a superimposed manner on the visible light image or the invisible light image for preview. This makes it possible for the user to visually recognize the subject detected in the visible light image in a state in which it is superimposed on the visible light image and the invisible light image as well as confirm the generated supervisory data. In addition, the user can change the class information of the subject, which serves as the supervisory data, and change the determination result as to whether supervisory data can be generated by an operation on the display screen. - The
memory 110 includes a non-volatile memory (ROM), a volatile memory (RAM), and the like and stores a control program for controlling the overall operation of theimage capture apparatus 100 and various parameters. - The
control unit 111 includes a processor (Central Processing Unit: CPU) for controlling the entire operation of theimage capture apparatus 100 by executing the program stored in thememory 110. - Next, the invisible light image supervisory data of the present embodiment is generated. Supervisory data generation processing will be described.
-
FIG. 6 is a flowchart illustrating processing for generating invisible light image supervisory data the present embodiment. - The processing of
FIG. 6 is realized by thecontrol unit 111 of theimage capture apparatus 100 illustrated inFIG. 1 controlling respective components by executing the program stored in thememory 110. - In step S601, the
control unit 111 generates a visible light image and an invisible light image of the same angle of view by capturing the visible light image and the invisible light image simultaneously or consecutively using the firstimage capturing unit 102 a and the secondimage capturing unit 102 b. Further, thecontrol unit 111 executes predetermined image processing on the visible light image and the invisible light image using the firstimage processing unit 103 a and the secondimage processing unit 103 b. - In step S602, the
control unit 111 performs subject detection processing on the visible light image obtained in step S601 using thesubject detection unit 104. Then, thecontrol unit 111 determines whether a subject has been detected and advances the processing to step S603 when thecontrol unit 111 determines that a subject has been detected and terminates the processing when thecontrol unit 111 determines that a subject has not been detected. - In step S603, the
control unit 111 determines whether invisible light image supervisory data can be generated by the supervisory datageneration determination unit 105 based on the subject detection result of step S602 and the invisible light image obtained in step S601. Thecontrol unit 111 advances the processing to step S604 when thecontrol unit 111 determines that the supervisory data can be generated and terminates the processing when thecontrol unit 111 determines that the supervisory data cannot be generated. - In step S604, the
control unit 111 generates invisible light image supervisory data using the supervisorydata generation unit 106 based on the subject detection result of step S602 and the invisible light image obtained in step S601. - As described above, according to the first embodiment, supervisory data to be used for machine learning of a learning model for subject detection of an invisible light image can be efficiently and accurately generated based on an invisible light image and a detection result of a subject in a visible light image. This makes it possible to generate a large amount of supervisory data for machine learning of a learning model for subject detection processing of an invisible light image.
- Next, a second embodiment will be described.
- In the second embodiment, an example in which a system for which the
image capture apparatus 100 of the first embodiment is separated into animage capture apparatus 710 for generating a visible light image and an invisible light image and a supervisorydata generation apparatus 720 for generating invisible light image supervisory data is configured will be described. - In the following, parts different from the first embodiment will be mainly described, and description of common parts will be omitted.
-
FIG. 7 is a block diagram illustrating a configuration of asystem 700 according to the second embodiment. - The
system 700 of the second embodiment includes theimage capture apparatus 710 and the supervisorydata generation apparatus 720. - The
image capture apparatus 710 includes a firstoptical system 711 a, a secondoptical system 711 b, a firstimage capturing unit 712 a, a secondimage capturing unit 712 b, a firstimage processing unit 713 a, a secondimage processing unit 713 b, amemory 714, and acontrol unit 715. - The
image capture apparatus 710 is, for example, a monitor camera or a fixed-point camera capable of simultaneously or consecutively capturing a visible light image and an invisible light image of the same angle of view. Theimage capture apparatus 710 is capable of capturing a visible light image and an invisible light image at a specific timing, continuously or at regular intervals. The firstoptical system 711 a, the secondoptical system 711 b, the firstimage capturing unit 712 a, the secondimage capturing unit 712 b, the firstimage processing unit 713 a, and the secondimage processing unit 713 b are similar to the firstoptical system 101 a, the secondoptical system 101 b, the firstimage capturing unit 102 a the secondimage capturing unit 102 b, the firstimage processing unit 103 a, and the secondimage processing unit 103 b of the first embodiment. - The
memory 714 includes a non-volatile memory (ROM), a volatile memory (RAM), and the like and stores a control program for controlling the overall operation of theimage capture apparatus 710 and various parameters. - The
control unit 715 includes a processor (CPU) for controlling the entire operation of theimage capture apparatus 710 by executing the program stored in thememory 714. - The supervisory
data generation apparatus 720 includes a firstimage acquiring unit 721 a, a secondimage acquiring unit 721 b, asubject detection unit 722, a supervisory datageneration determination unit 723, a supervisorydata generation unit 724, a supervisorydata storage unit 725, alearning unit 726, adisplay unit 727, amemory 728, and acontrol unit 729. - The supervisory
data generation apparatus 720 is, for example, a server connected to theimage capture apparatus 710 via a network so as to be capable of communication. - The first
image acquiring unit 721 a acquires a visible light image from theimage capture apparatus 710. The secondimage acquiring unit 721 b acquires an invisible light image from theimage capture apparatus 710. The visible light image and the invisible light image are images of the same angle of view captured simultaneously or consecutively. - The
subject detection unit 722, the supervisory datageneration determination unit 723, the supervisorydata generation unit 724, the supervisorydata storage unit 725, and thelearning unit 726 are similar to thesubject detection unit 104, the supervisory datageneration determination unit 105, the supervisorydata generation unit 106, the supervisorydata storage unit 107, and thelearning unit 108 of the first embodiment. - The
display unit 727 displays a subject detection result obtained by thesubject detection unit 722 in a superimposed manner on one or both of the visible light image obtained by the firstimage acquiring unit 721 a and the invisible light image obtained by the secondimage acquiring unit 721 b. Display contents are similar to those inFIG. 5 of the first embodiment. - The
memory 728 includes a non-volatile memory (ROM), a volatile memory (RAM), and the like and stores a control program for controlling the overall operation of the supervisorydata generation apparatus 720 and various parameters. - The
control unit 729 includes a processor (CPU) for controlling the entire operation of the supervisorydata generation apparatus 720 by executing the program stored in thememory 728. - Next, supervisory data generation processing for generating supervisory data of an invisible light image according to the second embodiment will be described.
-
FIG. 8 is a flowchart exemplifying supervisory data generation processing for generating supervisory data of an invisible light image according to the second embodiment. - The processing of
FIG. 8 is realized by thecontrol unit 715 of theimage capture apparatus 710 and thecontrol unit 729 of the supervisorydata generation apparatus 720 of thesystem 700 illustrated inFIG. 7 controlling respective components. In the processing ofFIG. 8 , it is assumed that theimage capture apparatus 710 and the supervisorydata generation apparatus 720 are connected via a network so as to be capable of communication. - In step S801, the
control unit 715 of theimage capture apparatus 710 generates a visible light image and an invisible light image of the same angle of view by simultaneously or consecutively capturing the visible light image and the invisible light image using theimage capture apparatus 710 and executes predetermined image processing on the visible light image and the invisible light image. - In step S802, the
control unit 715 of theimage capture apparatus 710 transmits the visible light image and the invisible light image obtained in step S801 to the supervisorydata generation apparatus 720. - In step S803, the
control unit 729 of the supervisorydata generation apparatus 720 acquires the visible light image and the invisible light image transmitted from theimage capture apparatus 710 in step S802 using the firstimage acquiring unit 721 a and the secondimage acquiring unit 721 b. - In step S804, the
control unit 729 of the supervisorydata generation apparatus 720 performs subject detection processing on the visible light image obtained in step S803 using thesubject detection unit 722. Thecontrol unit 729 of the supervisorydata generation apparatus 720 determines whether a subject has been detected by thesubject detection unit 722 and advances the processing to step S805 when thecontrol unit 111 determines that a subject has been detected and terminates the processing when thecontrol unit 111 determines that a subject has not been detected. - In step S805, the
control unit 729 of the supervisorydata generation apparatus 720 determines whether invisible light image supervisory data can be generated using the supervisory datageneration determination unit 723 and advances the processing to step S806 when thecontrol unit 111 determines that the supervisory data can be generated and terminates the processing when thecontrol unit 111 determines that the supervisory data cannot be generated. - In step S806, the
control unit 729 of the supervisorydata generation apparatus 720 generates invisible light image supervisory data based the subject detection result obtained in step S804 and the invisible light image acquired in step S803 using the supervisorydata generation unit 724. - As described above, according to the second embodiment, by configuring a system in which the
image capture apparatus 710 and the supervisorydata generation apparatus 720 are separate, it is possible to contribute to versatile and efficient generation of supervisory data, such as setting the supervisorydata generation apparatus 720 to be a server. - In the second embodiment, a configuration is taken such that the supervisory
data generation apparatus 720 acquires a visible light image and an invisible light image generated by a singleimage capture apparatus 710; however, the supervisorydata generation apparatus 720 may acquire a visible light image and an invisible light image of the same angle of view generated by a plurality of image capture apparatuses. For example, the supervisorydata generation apparatus 720 may acquire a visible light image and an invisible light image of the same angle of view generated by a plurality of monitoring cameras or the like installed at different positions. - Embodiment(s) of the present invention can also be realized by a computer of a system or apparatus that reads out and executes computer executable instructions (e.g., one or more programs) recorded on a storage medium (which may also be referred to more fully as a ‘non-transitory computer-readable storage medium’) to perform the functions of one or more of the above-described embodiment(s) and/or that includes one or more circuits (e.g., application specific integrated circuit (ASIC)) for performing the functions of one or more of the above-described embodiment(s), and by a method performed by the computer of the system or apparatus by, for example, reading out and executing the computer executable instructions from the storage medium to perform the functions of one or more of the above-described embodiment(s) and/or controlling the one or more circuits to perform the functions of one or more of the above-described embodiment(s). The computer may comprise one or more processors (e.g., central processing unit (CPU), micro processing unit (MPU)) and may include a network of separate computers or separate processors to read out and execute the computer executable instructions. The computer executable instructions may be provided to the computer, for example, from a network or the storage medium. The storage medium may include, for example, one or more of a hard disk, a random-access memory (RAM), a read only memory (ROM), a storage of distributed computing systems, an optical disk (such as a compact disc (CD), digital versatile disc (DVD), or Blu-ray Disc (BD)™), a flash memory device, a memory card, and the like.
- While the present invention has been described with reference to exemplary embodiments, it is to be understood that the invention is not limited to the disclosed exemplary embodiments. The scope of the following claims is to be accorded the broadest interpretation so as to encompass all such modifications and equivalent structures and functions.
- This application claims the benefit of Japanese Patent Application No. 2022-179090, filed Nov. 8, 2022 which is hereby incorporated by reference herein in its entirety.
Claims (13)
1. A data generation apparatus comprising:
a first image acquiring unit that acquires a visible light image;
a second image acquiring unit that acquires an invisible light image;
a subject detection unit that detects a subject in the visible light image; and
a supervisory data generation unit that generates supervisory data to be used in learning of a learning model for subject detection of an invisible light image based on the invisible light image and a subject detection result acquired by the subject detection unit.
2. The apparatus according to claim 1 , wherein
the visible light image and the invisible light image are simultaneously or consecutively captured images of the same angle of view.
3. The apparatus according to claim 1 , further comprising a supervisory data generation determination unit that determines whether the supervisory data can be generated.
4. The apparatus according to claim 3 , wherein
the supervisory data generation determination unit performs the determination based on the invisible light image.
5. The apparatus according to claim 4 , wherein
the supervisory data generation determination unit performs the determination based on luminance values of the invisible light image in the same region as a region of the subject detected in the visible light image.
6. The apparatus according to claim 3 , wherein
the supervisory data generation determination unit performs the determination based on a predesignated type of subject.
7. The apparatus according to claim 1 , wherein
the supervisory data generation unit performs at least one of processing for creating bokeh, processing for creating blurring, and processing for correcting luminance on the invisible light image.
8. The apparatus according to claim 1 , further comprising:
a display unit that displays a subject detection result acquired by the subject detection unit and a type of detected subject in a superimposed manner on one or both of the visible light image and the invisible light image.
9. The apparatus according to claim 1 , wherein
the first image acquiring unit and the second image acquiring unit acquire a visible light image and an invisible light image of the same angle of view that have been simultaneously or consecutively been captured by an image capture apparatus.
10. The apparatus according to claim 1 ,
wherein the first image acquiring unit is a first image capturing unit for capturing the visible light image and the second image acquiring unit is a second image capturing unit for capturing the invisible light image, and
wherein the first image capturing unit and the second image capturing unit capture a simultaneously or consecutively captured visible light image and invisible light image of the same angle of view.
11. The apparatus according to claim 1 , wherein
the invisible light image is one of a near-infrared image, a far-infrared image, and an ultraviolet image.
12. A method of controlling a data generation apparatus, the method comprising:
acquiring a visible light image and an invisible light image;
detecting a subject in the visible light image; and
generating supervisory data to be used in learning of a learning model for subject detection of an invisible light image based on the invisible light image and the subject detection result.
13. A non-transitory computer-readable storage medium storing a program for causing a computer to function as a data generation apparatus comprising:
a first image acquiring unit that acquires a visible light image;
a second image acquiring unit that acquires an invisible light image;
a subject detection unit that detects a subject in the visible light image; and
a supervisory data generation unit that generates supervisory data to be used in learning of a learning model for subject detection of an invisible light image based on the invisible light image and a subject detection result acquired by the subject detection unit.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2022179090A JP2024068553A (en) | 2022-11-08 | 2022-11-08 | Data creation device, control method and program |
JP2022-179090 | 2022-11-08 |
Publications (1)
Publication Number | Publication Date |
---|---|
US20240155210A1 true US20240155210A1 (en) | 2024-05-09 |
Family
ID=90928402
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US18/491,843 Pending US20240155210A1 (en) | 2022-11-08 | 2023-10-23 | Data generation apparatus and control method |
Country Status (2)
Country | Link |
---|---|
US (1) | US20240155210A1 (en) |
JP (1) | JP2024068553A (en) |
Citations (45)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6078682A (en) * | 1993-01-01 | 2000-06-20 | Canon Kabushiki Kaisha | Image processing apparatus and method, and image reading apparatus |
US20040091133A1 (en) * | 2002-09-12 | 2004-05-13 | Hitachi Ltd. | On board image processing apparatus |
US7006705B2 (en) * | 2001-03-15 | 2006-02-28 | Canon Kabushiki Kaisha | Image processing for correcting defects of read image |
US7034927B1 (en) * | 2002-06-28 | 2006-04-25 | Digeo, Inc. | System and method for identifying an object using invisible light |
US7589771B2 (en) * | 2004-12-22 | 2009-09-15 | Sony Corporation | Image processing apparatus, image processing method, image pickup apparatus, computer program and recording medium |
US8077200B2 (en) * | 2008-02-05 | 2011-12-13 | Asia Optical Co., Inc. | Image capture system and method of capturing labeled image |
US8416302B2 (en) * | 2009-02-10 | 2013-04-09 | Microsoft Corporation | Low-light imaging augmented with non-intrusive lighting |
US9392262B2 (en) * | 2014-03-07 | 2016-07-12 | Aquifi, Inc. | System and method for 3D reconstruction using multiple multi-channel cameras |
US9516295B2 (en) * | 2014-06-30 | 2016-12-06 | Aquifi, Inc. | Systems and methods for multi-channel imaging based on multiple exposure settings |
US9799122B2 (en) * | 2015-03-09 | 2017-10-24 | Canon Kabushiki Kaisha | Motion information acquiring apparatus and motion information acquiring method |
US9894255B2 (en) * | 2013-06-17 | 2018-02-13 | Industrial Technology Research Institute | Method and system for depth selective segmentation of object |
US9936145B2 (en) * | 2015-02-12 | 2018-04-03 | Canon Kabushiki Kaisha | Display control apparatus and display control method |
US20180336664A1 (en) * | 2017-05-17 | 2018-11-22 | Canon Kabushiki Kaisha | Image processing apparatus, image processing method, and storage medium |
US10209881B2 (en) * | 2012-03-15 | 2019-02-19 | Ibrahim Farid Cherradi El Fadili | Extending the free fingers typing technology and introducing the finger taps language technology |
US10291859B2 (en) * | 2014-11-13 | 2019-05-14 | Panasonic Intellectual Property Management Co., Ltd. | Imaging device and imaging method for composing a non-visible light image and a visibile light image |
US10356314B2 (en) * | 2015-11-25 | 2019-07-16 | Canon Kabushiki Kaisha | Image sensor and image capturing apparatus |
US10375357B2 (en) * | 2014-08-27 | 2019-08-06 | Apple Inc. | Method and system for providing at least one image captured by a scene camera of a vehicle |
US10395093B2 (en) * | 2016-05-16 | 2019-08-27 | Canon Kabushiki Kaisha | Image processing apparatus, image processing method, and non-transitory computer-readable storage medium |
US10489666B2 (en) * | 2017-12-18 | 2019-11-26 | Panasonic Intellectual Property Management Co., Ltd. | Imaging device and imaging system |
US10605661B2 (en) * | 2015-10-05 | 2020-03-31 | Nikon Corporation | Image capturing with filters of overlapping passbands |
US20210195128A1 (en) * | 2018-08-31 | 2021-06-24 | Sony Semiconductor Solutions Corporation | Solid-state image capturing device, information processing device, information processing system, information processing method, and computer program |
US11045271B1 (en) * | 2021-02-09 | 2021-06-29 | Bao Q Tran | Robotic medical system |
US20210217212A1 (en) * | 2017-05-24 | 2021-07-15 | HELLA GmbH & Co. KGaA | Method and system for automatically colorizing night-vision images |
US11074447B1 (en) * | 2018-07-13 | 2021-07-27 | Hana Resources, Inc. | Land analysis system using drone-captured data |
US11172180B2 (en) * | 2018-09-19 | 2021-11-09 | Canon Kabushiki Kaisha | Control apparatus, control method and non-transitory computer-readable medium |
US11195320B2 (en) * | 2019-12-12 | 2021-12-07 | Facebook Technologies, Llc | Feed-forward collision avoidance for artificial reality environments |
US11317000B2 (en) * | 2019-07-16 | 2022-04-26 | Ricoh Company, Ltd. | Image processing apparatus having invisible component remover, image processing method, and recording medium |
US11361408B2 (en) * | 2019-01-21 | 2022-06-14 | Canon Kabushiki Kaisha | Image processing apparatus, system, image processing method, and non-transitory computer-readable storage medium |
US11389059B2 (en) * | 2013-01-25 | 2022-07-19 | Wesley W. O. Krueger | Ocular-performance-based head impact measurement using a faceguard |
US11405761B2 (en) * | 2019-05-23 | 2022-08-02 | Connected Wise LLC | On-board machine vision device for activating vehicular messages from traffic signs |
US11423524B2 (en) * | 2019-03-11 | 2022-08-23 | Canon Kabushiki Kaisha | Image processing apparatus, method for controlling image processing apparatus, and non- transitory computer-readable storage medium |
US11483503B2 (en) * | 2019-01-20 | 2022-10-25 | Magik Eye Inc. | Three-dimensional sensor including bandpass filter having multiple passbands |
US11508064B2 (en) * | 2020-01-30 | 2022-11-22 | Fujitsu Limited | Computer-readable recording medium having stored therein information processing program, method for processing information, and information processing apparatus |
US20230041560A1 (en) * | 2021-08-03 | 2023-02-09 | Intrinsic Innovation Llc | Systems and methods for training pose estimators in computer vision |
US11590650B2 (en) * | 2019-06-10 | 2023-02-28 | Preferred Networks, Inc. | Generation method for training dataset, model generation method, training data generation apparatus, inference apparatus, robotic controller, model training method and robot |
US20230070798A1 (en) * | 2020-03-31 | 2023-03-09 | Sony Group Corporation | Image processing apparatus and method, program, and image processing system |
US11645864B2 (en) * | 2020-11-06 | 2023-05-09 | Hitachi, Ltd. | Imaging device, authentication device, and biometric imaging method |
US11643014B2 (en) * | 2018-07-31 | 2023-05-09 | Sony Semiconductor Solutions Corporation | Image capturing device and vehicle control system |
US20230309364A1 (en) * | 2020-09-11 | 2023-09-28 | Semiconductor Energy Laboratory Co., Ltd. | Display device, display module, and electronic device |
US11797863B2 (en) * | 2020-01-30 | 2023-10-24 | Intrinsic Innovation Llc | Systems and methods for synthesizing data for training statistical models on different imaging modalities including polarized images |
US11836179B1 (en) * | 2019-10-29 | 2023-12-05 | Meta Platforms Technologies, Llc | Multimedia query system |
US20240020968A1 (en) * | 2020-10-08 | 2024-01-18 | Edgy Bees Ltd. | Improving geo-registration using machine-learning based object identification |
US12104911B2 (en) * | 2021-03-04 | 2024-10-01 | Nec Corporation Of America | Imperceptible road markings to support automated vehicular systems |
US12150367B2 (en) * | 2020-09-11 | 2024-11-19 | Semiconductor Energy Laboratory Co., Ltd. | Display device, display module, and electronic device |
US12250489B2 (en) * | 2020-04-27 | 2025-03-11 | Sony Group Corporation | Information processing device and method for generating composite video |
-
2022
- 2022-11-08 JP JP2022179090A patent/JP2024068553A/en active Pending
-
2023
- 2023-10-23 US US18/491,843 patent/US20240155210A1/en active Pending
Patent Citations (45)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6078682A (en) * | 1993-01-01 | 2000-06-20 | Canon Kabushiki Kaisha | Image processing apparatus and method, and image reading apparatus |
US7006705B2 (en) * | 2001-03-15 | 2006-02-28 | Canon Kabushiki Kaisha | Image processing for correcting defects of read image |
US7034927B1 (en) * | 2002-06-28 | 2006-04-25 | Digeo, Inc. | System and method for identifying an object using invisible light |
US20040091133A1 (en) * | 2002-09-12 | 2004-05-13 | Hitachi Ltd. | On board image processing apparatus |
US7589771B2 (en) * | 2004-12-22 | 2009-09-15 | Sony Corporation | Image processing apparatus, image processing method, image pickup apparatus, computer program and recording medium |
US8077200B2 (en) * | 2008-02-05 | 2011-12-13 | Asia Optical Co., Inc. | Image capture system and method of capturing labeled image |
US8416302B2 (en) * | 2009-02-10 | 2013-04-09 | Microsoft Corporation | Low-light imaging augmented with non-intrusive lighting |
US10209881B2 (en) * | 2012-03-15 | 2019-02-19 | Ibrahim Farid Cherradi El Fadili | Extending the free fingers typing technology and introducing the finger taps language technology |
US11389059B2 (en) * | 2013-01-25 | 2022-07-19 | Wesley W. O. Krueger | Ocular-performance-based head impact measurement using a faceguard |
US9894255B2 (en) * | 2013-06-17 | 2018-02-13 | Industrial Technology Research Institute | Method and system for depth selective segmentation of object |
US9392262B2 (en) * | 2014-03-07 | 2016-07-12 | Aquifi, Inc. | System and method for 3D reconstruction using multiple multi-channel cameras |
US9516295B2 (en) * | 2014-06-30 | 2016-12-06 | Aquifi, Inc. | Systems and methods for multi-channel imaging based on multiple exposure settings |
US10375357B2 (en) * | 2014-08-27 | 2019-08-06 | Apple Inc. | Method and system for providing at least one image captured by a scene camera of a vehicle |
US10291859B2 (en) * | 2014-11-13 | 2019-05-14 | Panasonic Intellectual Property Management Co., Ltd. | Imaging device and imaging method for composing a non-visible light image and a visibile light image |
US9936145B2 (en) * | 2015-02-12 | 2018-04-03 | Canon Kabushiki Kaisha | Display control apparatus and display control method |
US9799122B2 (en) * | 2015-03-09 | 2017-10-24 | Canon Kabushiki Kaisha | Motion information acquiring apparatus and motion information acquiring method |
US10605661B2 (en) * | 2015-10-05 | 2020-03-31 | Nikon Corporation | Image capturing with filters of overlapping passbands |
US10356314B2 (en) * | 2015-11-25 | 2019-07-16 | Canon Kabushiki Kaisha | Image sensor and image capturing apparatus |
US10395093B2 (en) * | 2016-05-16 | 2019-08-27 | Canon Kabushiki Kaisha | Image processing apparatus, image processing method, and non-transitory computer-readable storage medium |
US20180336664A1 (en) * | 2017-05-17 | 2018-11-22 | Canon Kabushiki Kaisha | Image processing apparatus, image processing method, and storage medium |
US20210217212A1 (en) * | 2017-05-24 | 2021-07-15 | HELLA GmbH & Co. KGaA | Method and system for automatically colorizing night-vision images |
US10489666B2 (en) * | 2017-12-18 | 2019-11-26 | Panasonic Intellectual Property Management Co., Ltd. | Imaging device and imaging system |
US11074447B1 (en) * | 2018-07-13 | 2021-07-27 | Hana Resources, Inc. | Land analysis system using drone-captured data |
US11643014B2 (en) * | 2018-07-31 | 2023-05-09 | Sony Semiconductor Solutions Corporation | Image capturing device and vehicle control system |
US20210195128A1 (en) * | 2018-08-31 | 2021-06-24 | Sony Semiconductor Solutions Corporation | Solid-state image capturing device, information processing device, information processing system, information processing method, and computer program |
US11172180B2 (en) * | 2018-09-19 | 2021-11-09 | Canon Kabushiki Kaisha | Control apparatus, control method and non-transitory computer-readable medium |
US11483503B2 (en) * | 2019-01-20 | 2022-10-25 | Magik Eye Inc. | Three-dimensional sensor including bandpass filter having multiple passbands |
US11361408B2 (en) * | 2019-01-21 | 2022-06-14 | Canon Kabushiki Kaisha | Image processing apparatus, system, image processing method, and non-transitory computer-readable storage medium |
US11423524B2 (en) * | 2019-03-11 | 2022-08-23 | Canon Kabushiki Kaisha | Image processing apparatus, method for controlling image processing apparatus, and non- transitory computer-readable storage medium |
US11405761B2 (en) * | 2019-05-23 | 2022-08-02 | Connected Wise LLC | On-board machine vision device for activating vehicular messages from traffic signs |
US11590650B2 (en) * | 2019-06-10 | 2023-02-28 | Preferred Networks, Inc. | Generation method for training dataset, model generation method, training data generation apparatus, inference apparatus, robotic controller, model training method and robot |
US11317000B2 (en) * | 2019-07-16 | 2022-04-26 | Ricoh Company, Ltd. | Image processing apparatus having invisible component remover, image processing method, and recording medium |
US11836179B1 (en) * | 2019-10-29 | 2023-12-05 | Meta Platforms Technologies, Llc | Multimedia query system |
US11195320B2 (en) * | 2019-12-12 | 2021-12-07 | Facebook Technologies, Llc | Feed-forward collision avoidance for artificial reality environments |
US11797863B2 (en) * | 2020-01-30 | 2023-10-24 | Intrinsic Innovation Llc | Systems and methods for synthesizing data for training statistical models on different imaging modalities including polarized images |
US11508064B2 (en) * | 2020-01-30 | 2022-11-22 | Fujitsu Limited | Computer-readable recording medium having stored therein information processing program, method for processing information, and information processing apparatus |
US20230070798A1 (en) * | 2020-03-31 | 2023-03-09 | Sony Group Corporation | Image processing apparatus and method, program, and image processing system |
US12250489B2 (en) * | 2020-04-27 | 2025-03-11 | Sony Group Corporation | Information processing device and method for generating composite video |
US20230309364A1 (en) * | 2020-09-11 | 2023-09-28 | Semiconductor Energy Laboratory Co., Ltd. | Display device, display module, and electronic device |
US12150367B2 (en) * | 2020-09-11 | 2024-11-19 | Semiconductor Energy Laboratory Co., Ltd. | Display device, display module, and electronic device |
US20240020968A1 (en) * | 2020-10-08 | 2024-01-18 | Edgy Bees Ltd. | Improving geo-registration using machine-learning based object identification |
US11645864B2 (en) * | 2020-11-06 | 2023-05-09 | Hitachi, Ltd. | Imaging device, authentication device, and biometric imaging method |
US11045271B1 (en) * | 2021-02-09 | 2021-06-29 | Bao Q Tran | Robotic medical system |
US12104911B2 (en) * | 2021-03-04 | 2024-10-01 | Nec Corporation Of America | Imperceptible road markings to support automated vehicular systems |
US20230041560A1 (en) * | 2021-08-03 | 2023-02-09 | Intrinsic Innovation Llc | Systems and methods for training pose estimators in computer vision |
Also Published As
Publication number | Publication date |
---|---|
JP2024068553A (en) | 2024-05-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11195055B2 (en) | Image processing method, image processing apparatus, storage medium, image processing system, and manufacturing method of learnt model | |
US20150358542A1 (en) | Image processing apparatus, image capturing apparatus, image processing method, image capturing method, and non-transitory computer-readable medium for focus bracketing | |
IES20070820A2 (en) | Method for improved red eye detection in ISO images | |
US10395093B2 (en) | Image processing apparatus, image processing method, and non-transitory computer-readable storage medium | |
JP4442413B2 (en) | Image processing apparatus, image processing method, program, and recording medium | |
US10148862B2 (en) | Image capturing apparatus, method for controlling image capturing apparatus focus area display, and storage medium | |
US9706110B2 (en) | Foreign body information detection device and foreign body information detection method of imaging apparatus | |
JP5949559B2 (en) | Image processing apparatus, imaging apparatus, and image processing program | |
US20160381342A1 (en) | Image processing apparatus, imaging apparatus, image processing method, and recording medium | |
JP2008092299A (en) | Electronic camera | |
US11575841B2 (en) | Information processing apparatus, imaging apparatus, method, and storage medium | |
JP2012085083A (en) | Image processing apparatus, image pickup device, and image processing program | |
US20240155210A1 (en) | Data generation apparatus and control method | |
JP2017098900A (en) | Image processing apparatus, image processing method, and program | |
JP2008199164A (en) | Image processing method for noise reduction and imaging apparatus using the same | |
US9648232B2 (en) | Image processing apparatus, image capturing apparatus, control method and recording medium | |
US12160668B2 (en) | Image capturing system, control method, and storage medium | |
US12432457B2 (en) | Image capturing apparatus and control method thereof and storage medium | |
US20250267362A1 (en) | Image processing apparatus, image capturing apparatus, image processing method, and non-transitory computer-readable storage medium | |
JP2010219862A (en) | Face detection device, and imaging apparatus | |
US20230199319A1 (en) | Image capturing apparatus and control method thereof and storage medium | |
US20250232463A1 (en) | Image processing apparatus, control method of image processing apparatus, and non-transitory computer-readable storage medium | |
JP2010220122A (en) | Image processing apparatus, image capturing apparatus, image processing method, and image processing program | |
US20240187727A1 (en) | Image processing apparatus, image capturing apparatus, control method of image processing apparatus, and storage medium | |
JP5631181B2 (en) | Image processing apparatus, control method therefor, and program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
AS | Assignment |
Owner name: CANON KABUSHIKI KAISHA, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:KOGAI, YUKIHIRO;REEL/FRAME:065691/0252 Effective date: 20231016 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |