[go: up one dir, main page]

US20240155210A1 - Data generation apparatus and control method - Google Patents

Data generation apparatus and control method Download PDF

Info

Publication number
US20240155210A1
US20240155210A1 US18/491,843 US202318491843A US2024155210A1 US 20240155210 A1 US20240155210 A1 US 20240155210A1 US 202318491843 A US202318491843 A US 202318491843A US 2024155210 A1 US2024155210 A1 US 2024155210A1
Authority
US
United States
Prior art keywords
light image
image
unit
invisible light
subject
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US18/491,843
Inventor
Yukihiro Kogai
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Canon Inc
Original Assignee
Canon Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Canon Inc filed Critical Canon Inc
Assigned to CANON KABUSHIKI KAISHA reassignment CANON KABUSHIKI KAISHA ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: KOGAI, Yukihiro
Publication of US20240155210A1 publication Critical patent/US20240155210A1/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/10Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths
    • H04N23/11Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths for generating image signals from visible and infrared light wavelengths
    • G06T5/002
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/60Extraction of image or video features relating to illumination properties, e.g. using a reflectance or lighting model
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Definitions

  • the present invention relates to techniques for generating supervisory data to be used in learning of a learning model for subject detection processing of an invisible light image.
  • Japanese Patent Laid-Open No. 2019-118043 discloses a method of performing subject detection processing of a visible light image and an invisible light image using machine learning.
  • a large amount of supervisory data is necessary for learning processing of a learning model for subject detection processing of visible light images; conventionally, learning models for visible light images are the mainstream of learning models for subject detection processing; it takes a lot of effort and man-hours to generate a large amount of supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image.
  • the present invention has been made in consideration of the aforementioned problems and realizes techniques for allowing efficient and accurate generation of supervisory data to be used in learning of a learning model for detecting a subject in an invisible light image.
  • the present invention provides a data generation apparatus comprising: a first image acquiring unit that acquires a visible light image; a second image acquiring unit that acquires an invisible light image; a subject detection unit that detects a subject in the visible light image; and a supervisory data generation unit that generates supervisory data to be used in learning of a learning model for subject detection of an invisible light image based on the invisible light image and a subject detection result acquired by the subject detection unit.
  • the present invention provides a method of controlling a data generation apparatus, the method comprising: acquiring a visible light image and an invisible light image; detecting a subject in the visible light image; and generating supervisory data to be used in learning of a learning model for subject detection of an invisible light image based on the invisible light image and the subject detection result.
  • the present invention provides a non-transitory computer-readable storage medium storing a program for causing a computer to function as a data generation apparatus comprising: a first image acquiring unit that acquires a visible light image; a second image acquiring unit that acquires an invisible light image; a subject detection unit that detects a subject in the visible light image; and a supervisory data generation unit that generates supervisory data to be used in learning of a learning model for subject detection of an invisible light image based on the invisible light image and a subject detection result acquired by the subject detection unit.
  • supervisory data to be used in learning of a learning model for subject detection of an invisible light image can be efficiently and accurately generated.
  • FIG. 1 is a block diagram illustrating an apparatus configuration according to a first embodiment.
  • FIG. 2 is a diagram illustrating a relationship among a visible light image, an invisible light image, and a detection result of a subject detected in the visible light image according to the first embodiment.
  • FIG. 3 is a diagram illustrating image processing for generating supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image according to the first embodiment.
  • FIG. 4 is a diagram illustrating an example of display of a visible light image, an invisible light image, and a subject detection result according to the first embodiment.
  • FIG. 5 is a diagram illustrating an example of display of a generation result of supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image according to the first embodiment.
  • FIG. 6 is a flowchart exemplifying processing for generating supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image according to the first embodiment.
  • FIG. 7 is a diagram of a system configuration according to a second embodiment.
  • FIG. 8 is a flowchart exemplifying processing for generating supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image according to the second embodiment.
  • An image capture apparatus 100 of the first embodiment captures a visible light image and an invisible light image and generates supervisory data for a learning model for subject detection processing of an invisible light image based on the invisible light image and a subject detection result (type, position, and size of the subject) of the visible light image.
  • the supervisory data of the present embodiment is data to be used in machine learning of a learning model for subject detection of an invisible light image, and learning processing is executed with supervisory data as input data (an invisible light image) and output data (a detection result of a subject detected from a visible light image).
  • FIG. 1 is a block diagram illustrating a configuration of the image capture apparatus 100 according to the first embodiment.
  • the image capture apparatus 100 includes a first optical system 101 a , a second optical system 101 b , a first image capturing unit 102 a , a second image capturing unit 102 b , a first image processing unit 103 a , a second image processing unit 103 b , a subject detection unit 104 , a supervisory data generation determination unit 105 , a supervisory data generation unit 106 , a supervisory data storage unit 107 , a learning unit 108 , a display unit 109 , a memory 110 , and a control unit 111 .
  • the first optical system 101 a includes one or more lenses and forms an image of subject image light in a visible light range on the first image capturing unit 102 a .
  • the second optical system 101 b includes one or more lenses and forms an image of subject image light in an invisible light range on the second image capturing unit 102 b .
  • An image in a visible light range is, for example, an image in a wavelength range of 400 to 800 nm.
  • An image in an invisible light range is, for example, a near-infrared image in a near-infrared wavelength range of 800 to 2500 nm but may be a far-infrared image in a far-infrared wavelength range of 4 ⁇ m to 1000 ⁇ m or an ultraviolet image in an ultraviolet wavelength range of 380 nm or less.
  • the first image capturing unit 102 a includes an image sensor, such as a CMOS sensor, for converting subject image light of a visible light range formed into an image by the first optical system 101 a into an electric signal.
  • the image sensor includes, for example, color filters in an RGB Bayer array.
  • the first image capturing unit 102 a includes an AD converter for converting an analog electric signal into a digital signal.
  • the second image capturing unit 102 b includes an image sensor, such as a CMOS, for converting subject image light of an invisible light range formed into an image by the second optical system 101 b into an electric signal.
  • the image sensor includes, for example, a color filter that passes light of a near-infrared wavelength range.
  • the second image capturing unit 102 b includes an AD converter for converting an analog electric signal into a digital signal.
  • the first image processing unit 103 a includes a processor (GPU) for executing predetermined image processing on a digital signal obtained by the first image capturing unit 102 a and generates visible light image data.
  • the predetermined image processing includes, for example, distortion correction processing, noise removal processing, exposure correction processing, white balance processing, and edge enhancement processing.
  • the second image processing unit 103 b includes a processor (GPU) for executing predetermined image processing on a digital signal obtained by the second image capturing unit 102 b and generates invisible light image data.
  • the predetermined image processing includes, for example, distortion correction processing, noise removal processing, exposure correction processing, white balance processing, and edge enhancement processing.
  • the subject detection unit 104 detects one or more subjects in a visible light image processed by the first image processing unit 103 a . Then, the subject detection unit 104 detects class information indicating the type of the detected subject, position information indicating a position (center coordinates) of the subject, size information indicating a size of the subject, and the like.
  • Subject detection can be realized by image analysis processing or image recognition processing in which a learning model for which learning processing has been performed by machine learning is used.
  • a learning model is, for example, a neural network, and the class information (such as whether a person is included or a car is included), the position information, and the size information of a subject in an image is detected using a learning model trained with supervisory data.
  • the invisible light image supervisory data of the present embodiment is generated based on an invisible light image and the class information, the position information, and the size information of a subject obtained from a visible light image.
  • FIG. 2 is a diagram illustrating a relationship among a visible light image, an invisible light image, and a detection result of a subject detected in the visible light image according to the present embodiment.
  • a visible light image 201 is an image that has been processed by the first image processing unit 103 a
  • the subject detection unit 104 detects a subject 202 for which a subject region 203 has been surrounded by a rectangular frame.
  • a subject detection result 204 includes the class information, the size information (X, Y), and the position information (center coordinates) of the subject.
  • An invisible light image 205 is an image that has been processed by the second image processing unit 103 b and is an image that has been captured at the same angle of view as that of the visible light image 201 .
  • a subject 206 is a subject that corresponds to the subject 202 of the visible light image 201 .
  • a subject region 207 and a subject detection result 208 are the same as the subject region 203 and the subject detection result 204 of the visible light image.
  • the supervisory data generation determination unit 105 determines whether invisible light image supervisory data can be generated based on a subject detection result obtained by the subject detection unit 104 and an invisible light image obtained by the second image processing unit 103 b .
  • a determination method for example, a histogram of luminance values of an invisible light image in the same region as a subject region of a visible light image obtained by the subject detection unit 104 is generated, and when the luminance values are within a predetermined range, it is determined that supervisory data can be generated and when the luminance values are outside of the predetermined range, it is determined that supervisory data cannot be generated.
  • a histogram generated based on the entire image may be used.
  • the supervisory data generation determination unit 105 includes a subject type designation unit (not illustrated) and can determine whether supervisory data for a pre-designated subject can be generated by a subject type for which to generate supervisory data being set by a user operation or a subject type for which to generate supervisory data being selected by a user operation in superimposed display of an invisible light image and a subject detection result, which will be described later in FIG. 5 .
  • the supervisory data generation unit 106 When it is determined by the supervisory data generation determination unit 105 that supervisory data can be generated, the supervisory data generation unit 106 generates invisible light image supervisory data based on an invisible light image and a subject detection result of the subject detection unit 104 .
  • the supervisory data generation unit 106 includes an image processing unit (not illustrated) and can generate a plurality of pieces of supervisory data from a single invisible light image by executing specific image processing on the invisible light image.
  • the specific image processing includes, for example, at least one of processing for creating bokeh, processing of creating blurring, and processing for correcting luminance.
  • FIG. 3 illustrates processing 301 for correcting luminance, processing 302 for creating blurring, and processing 303 for creating bokeh as examples of specific image processing to be performed on an invisible light image.
  • processing 301 for correcting luminance processing 302 for creating blurring
  • processing 303 for creating bokeh processing 301 for correcting luminance
  • processing 302 for creating blurring processing 302 for creating blurring
  • processing 303 for creating bokeh processing 301 for correcting luminance
  • processing 302 for creating blurring processing 303 for creating bokeh
  • the supervisory data storage unit 107 stores invisible light image supervisory data generated by the supervisory data generation unit 106 .
  • the learning unit 108 performs learning processing by machine learning on a learning model for subject detection processing of an invisible light image using invisible light image supervisory data stored in the supervisory data storage unit 107 .
  • the learning processing of the learning unit 108 is executed with the supervisory data as input data and output data of the learning model for subject detection processing of an invisible light image.
  • a graphics processing unit GPU can perform efficient computation by processing more data in parallel processing; therefore, it is useful to perform processing using the GPU when performing learning processing a plurality of times using a learning model, such as in machine learning.
  • computation may be performed by the control unit 111 , which will be described later, and the GPU cooperating with each other or computation may be performed by the control unit 111 or the GPU alone.
  • a configuration is taken such that the image capture apparatus 100 includes the learning unit 108 ; however, the learning unit 108 may be configured to be separate from the image capture apparatus 100 .
  • the display unit 109 displays a subject detection result 401 obtained by the subject detection unit 104 in a superimposed manner on one or both of the visible light image obtained by the first image processing unit 103 a and the invisible light image obtained by the second image processing unit 103 b .
  • the displayed subject detection result 401 is a rectangular frame surrounding the subject region.
  • the display unit 109 displays the invisible light image supervisory data obtained by the supervisory data generation unit 106 for preview.
  • the class information, the size information, and the position information of the subject are displayed as invisible light image supervisory data 501 in a superimposed manner on the visible light image or the invisible light image for preview.
  • the user can change the class information of the subject, which serves as the supervisory data, and change the determination result as to whether supervisory data can be generated by an operation on the display screen.
  • the memory 110 includes a non-volatile memory (ROM), a volatile memory (RAM), and the like and stores a control program for controlling the overall operation of the image capture apparatus 100 and various parameters.
  • ROM non-volatile memory
  • RAM volatile memory
  • the control unit 111 includes a processor (Central Processing Unit: CPU) for controlling the entire operation of the image capture apparatus 100 by executing the program stored in the memory 110 .
  • CPU Central Processing Unit
  • FIG. 6 is a flowchart illustrating processing for generating invisible light image supervisory data the present embodiment.
  • FIG. 6 The processing of FIG. 6 is realized by the control unit 111 of the image capture apparatus 100 illustrated in FIG. 1 controlling respective components by executing the program stored in the memory 110 .
  • step S 601 the control unit 111 generates a visible light image and an invisible light image of the same angle of view by capturing the visible light image and the invisible light image simultaneously or consecutively using the first image capturing unit 102 a and the second image capturing unit 102 b . Further, the control unit 111 executes predetermined image processing on the visible light image and the invisible light image using the first image processing unit 103 a and the second image processing unit 103 b.
  • step S 602 the control unit 111 performs subject detection processing on the visible light image obtained in step S 601 using the subject detection unit 104 . Then, the control unit 111 determines whether a subject has been detected and advances the processing to step S 603 when the control unit 111 determines that a subject has been detected and terminates the processing when the control unit 111 determines that a subject has not been detected.
  • step S 603 the control unit 111 determines whether invisible light image supervisory data can be generated by the supervisory data generation determination unit 105 based on the subject detection result of step S 602 and the invisible light image obtained in step S 601 .
  • the control unit 111 advances the processing to step S 604 when the control unit 111 determines that the supervisory data can be generated and terminates the processing when the control unit 111 determines that the supervisory data cannot be generated.
  • step S 604 the control unit 111 generates invisible light image supervisory data using the supervisory data generation unit 106 based on the subject detection result of step S 602 and the invisible light image obtained in step S 601 .
  • supervisory data to be used for machine learning of a learning model for subject detection of an invisible light image can be efficiently and accurately generated based on an invisible light image and a detection result of a subject in a visible light image. This makes it possible to generate a large amount of supervisory data for machine learning of a learning model for subject detection processing of an invisible light image.
  • FIG. 7 is a block diagram illustrating a configuration of a system 700 according to the second embodiment.
  • the system 700 of the second embodiment includes the image capture apparatus 710 and the supervisory data generation apparatus 720 .
  • the image capture apparatus 710 includes a first optical system 711 a , a second optical system 711 b , a first image capturing unit 712 a , a second image capturing unit 712 b , a first image processing unit 713 a , a second image processing unit 713 b , a memory 714 , and a control unit 715 .
  • the image capture apparatus 710 is, for example, a monitor camera or a fixed-point camera capable of simultaneously or consecutively capturing a visible light image and an invisible light image of the same angle of view.
  • the image capture apparatus 710 is capable of capturing a visible light image and an invisible light image at a specific timing, continuously or at regular intervals.
  • the first optical system 711 a , the second optical system 711 b , the first image capturing unit 712 a , the second image capturing unit 712 b , the first image processing unit 713 a , and the second image processing unit 713 b are similar to the first optical system 101 a , the second optical system 101 b , the first image capturing unit 102 a the second image capturing unit 102 b , the first image processing unit 103 a , and the second image processing unit 103 b of the first embodiment.
  • the memory 714 includes a non-volatile memory (ROM), a volatile memory (RAM), and the like and stores a control program for controlling the overall operation of the image capture apparatus 710 and various parameters.
  • ROM non-volatile memory
  • RAM volatile memory
  • the control unit 715 includes a processor (CPU) for controlling the entire operation of the image capture apparatus 710 by executing the program stored in the memory 714 .
  • a processor CPU
  • the supervisory data generation apparatus 720 includes a first image acquiring unit 721 a , a second image acquiring unit 721 b , a subject detection unit 722 , a supervisory data generation determination unit 723 , a supervisory data generation unit 724 , a supervisory data storage unit 725 , a learning unit 726 , a display unit 727 , a memory 728 , and a control unit 729 .
  • the supervisory data generation apparatus 720 is, for example, a server connected to the image capture apparatus 710 via a network so as to be capable of communication.
  • the first image acquiring unit 721 a acquires a visible light image from the image capture apparatus 710 .
  • the second image acquiring unit 721 b acquires an invisible light image from the image capture apparatus 710 .
  • the visible light image and the invisible light image are images of the same angle of view captured simultaneously or consecutively.
  • the subject detection unit 722 , the supervisory data generation determination unit 723 , the supervisory data generation unit 724 , the supervisory data storage unit 725 , and the learning unit 726 are similar to the subject detection unit 104 , the supervisory data generation determination unit 105 , the supervisory data generation unit 106 , the supervisory data storage unit 107 , and the learning unit 108 of the first embodiment.
  • the display unit 727 displays a subject detection result obtained by the subject detection unit 722 in a superimposed manner on one or both of the visible light image obtained by the first image acquiring unit 721 a and the invisible light image obtained by the second image acquiring unit 721 b . Display contents are similar to those in FIG. 5 of the first embodiment.
  • the memory 728 includes a non-volatile memory (ROM), a volatile memory (RAM), and the like and stores a control program for controlling the overall operation of the supervisory data generation apparatus 720 and various parameters.
  • ROM non-volatile memory
  • RAM volatile memory
  • the control unit 729 includes a processor (CPU) for controlling the entire operation of the supervisory data generation apparatus 720 by executing the program stored in the memory 728 .
  • a processor CPU for controlling the entire operation of the supervisory data generation apparatus 720 by executing the program stored in the memory 728 .
  • FIG. 8 is a flowchart exemplifying supervisory data generation processing for generating supervisory data of an invisible light image according to the second embodiment.
  • FIG. 8 The processing of FIG. 8 is realized by the control unit 715 of the image capture apparatus 710 and the control unit 729 of the supervisory data generation apparatus 720 of the system 700 illustrated in FIG. 7 controlling respective components. In the processing of FIG. 8 , it is assumed that the image capture apparatus 710 and the supervisory data generation apparatus 720 are connected via a network so as to be capable of communication.
  • step S 801 the control unit 715 of the image capture apparatus 710 generates a visible light image and an invisible light image of the same angle of view by simultaneously or consecutively capturing the visible light image and the invisible light image using the image capture apparatus 710 and executes predetermined image processing on the visible light image and the invisible light image.
  • step S 802 the control unit 715 of the image capture apparatus 710 transmits the visible light image and the invisible light image obtained in step S 801 to the supervisory data generation apparatus 720 .
  • step S 803 the control unit 729 of the supervisory data generation apparatus 720 acquires the visible light image and the invisible light image transmitted from the image capture apparatus 710 in step S 802 using the first image acquiring unit 721 a and the second image acquiring unit 721 b.
  • step S 804 the control unit 729 of the supervisory data generation apparatus 720 performs subject detection processing on the visible light image obtained in step S 803 using the subject detection unit 722 .
  • the control unit 729 of the supervisory data generation apparatus 720 determines whether a subject has been detected by the subject detection unit 722 and advances the processing to step S 805 when the control unit 111 determines that a subject has been detected and terminates the processing when the control unit 111 determines that a subject has not been detected.
  • step S 805 the control unit 729 of the supervisory data generation apparatus 720 determines whether invisible light image supervisory data can be generated using the supervisory data generation determination unit 723 and advances the processing to step S 806 when the control unit 111 determines that the supervisory data can be generated and terminates the processing when the control unit 111 determines that the supervisory data cannot be generated.
  • step S 806 the control unit 729 of the supervisory data generation apparatus 720 generates invisible light image supervisory data based the subject detection result obtained in step S 804 and the invisible light image acquired in step S 803 using the supervisory data generation unit 724 .
  • the image capture apparatus 710 and the supervisory data generation apparatus 720 are separate, it is possible to contribute to versatile and efficient generation of supervisory data, such as setting the supervisory data generation apparatus 720 to be a server.
  • a configuration is taken such that the supervisory data generation apparatus 720 acquires a visible light image and an invisible light image generated by a single image capture apparatus 710 ; however, the supervisory data generation apparatus 720 may acquire a visible light image and an invisible light image of the same angle of view generated by a plurality of image capture apparatuses. For example, the supervisory data generation apparatus 720 may acquire a visible light image and an invisible light image of the same angle of view generated by a plurality of monitoring cameras or the like installed at different positions.
  • Embodiment(s) of the present invention can also be realized by a computer of a system or apparatus that reads out and executes computer executable instructions (e.g., one or more programs) recorded on a storage medium (which may also be referred to more fully as a ‘non-transitory computer-readable storage medium’) to perform the functions of one or more of the above-described embodiment(s) and/or that includes one or more circuits (e.g., application specific integrated circuit (ASIC)) for performing the functions of one or more of the above-described embodiment(s), and by a method performed by the computer of the system or apparatus by, for example, reading out and executing the computer executable instructions from the storage medium to perform the functions of one or more of the above-described embodiment(s) and/or controlling the one or more circuits to perform the functions of one or more of the above-described embodiment(s).
  • computer executable instructions e.g., one or more programs
  • a storage medium which may also be referred to more fully as a
  • the computer may comprise one or more processors (e.g., central processing unit (CPU), micro processing unit (MPU)) and may include a network of separate computers or separate processors to read out and execute the computer executable instructions.
  • the computer executable instructions may be provided to the computer, for example, from a network or the storage medium.
  • the storage medium may include, for example, one or more of a hard disk, a random-access memory (RAM), a read only memory (ROM), a storage of distributed computing systems, an optical disk (such as a compact disc (CD), digital versatile disc (DVD), or Blu-ray Disc (BD)TM), a flash memory device, a memory card, and the like.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Signal Processing (AREA)
  • Image Analysis (AREA)

Abstract

A data generation apparatus comprises a first image acquiring unit that acquires a visible light image, a second image acquiring unit that acquires an invisible light image, a subject detection unit that detects a subject in the visible light image, and a supervisory data generation unit that generates supervisory data to be used in learning of a learning model for subject detection of an invisible light image based on the invisible light image and a subject detection result acquired by the subject detection unit.

Description

    BACKGROUND OF THE INVENTION Field of the Invention
  • The present invention relates to techniques for generating supervisory data to be used in learning of a learning model for subject detection processing of an invisible light image.
  • Description of the Related Art
  • A method in which, in an image capture apparatus, such as a digital camera, a subject is detected in an image by image analysis processing or image recognition processing in which a learning model for which learning processing has been performed by machine learning is used is known. Further, Japanese Patent Laid-Open No. 2019-118043 discloses a method of performing subject detection processing of a visible light image and an invisible light image using machine learning.
  • A large amount of supervisory data is necessary for learning processing of a learning model for subject detection processing of visible light images; conventionally, learning models for visible light images are the mainstream of learning models for subject detection processing; it takes a lot of effort and man-hours to generate a large amount of supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image.
  • SUMMARY OF THE INVENTION
  • The present invention has been made in consideration of the aforementioned problems and realizes techniques for allowing efficient and accurate generation of supervisory data to be used in learning of a learning model for detecting a subject in an invisible light image.
  • In order to solve the aforementioned problems, the present invention provides a data generation apparatus comprising: a first image acquiring unit that acquires a visible light image; a second image acquiring unit that acquires an invisible light image; a subject detection unit that detects a subject in the visible light image; and a supervisory data generation unit that generates supervisory data to be used in learning of a learning model for subject detection of an invisible light image based on the invisible light image and a subject detection result acquired by the subject detection unit.
  • In order to solve the aforementioned problems, the present invention provides a method of controlling a data generation apparatus, the method comprising: acquiring a visible light image and an invisible light image; detecting a subject in the visible light image; and generating supervisory data to be used in learning of a learning model for subject detection of an invisible light image based on the invisible light image and the subject detection result.
  • In order to solve the aforementioned problems, the present invention provides a non-transitory computer-readable storage medium storing a program for causing a computer to function as a data generation apparatus comprising: a first image acquiring unit that acquires a visible light image; a second image acquiring unit that acquires an invisible light image; a subject detection unit that detects a subject in the visible light image; and a supervisory data generation unit that generates supervisory data to be used in learning of a learning model for subject detection of an invisible light image based on the invisible light image and a subject detection result acquired by the subject detection unit.
  • According to the present invention, supervisory data to be used in learning of a learning model for subject detection of an invisible light image can be efficiently and accurately generated.
  • Further features of the present invention will become apparent from the following description of exemplary embodiments (with reference to the attached drawings).
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a block diagram illustrating an apparatus configuration according to a first embodiment.
  • FIG. 2 is a diagram illustrating a relationship among a visible light image, an invisible light image, and a detection result of a subject detected in the visible light image according to the first embodiment.
  • FIG. 3 is a diagram illustrating image processing for generating supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image according to the first embodiment.
  • FIG. 4 is a diagram illustrating an example of display of a visible light image, an invisible light image, and a subject detection result according to the first embodiment.
  • FIG. 5 is a diagram illustrating an example of display of a generation result of supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image according to the first embodiment.
  • FIG. 6 is a flowchart exemplifying processing for generating supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image according to the first embodiment.
  • FIG. 7 is a diagram of a system configuration according to a second embodiment.
  • FIG. 8 is a flowchart exemplifying processing for generating supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image according to the second embodiment.
  • DESCRIPTION OF THE EMBODIMENTS
  • Hereinafter, embodiments will be described in detail with reference to the attached drawings. Note, the following embodiments are not intended to limit the scope of the claimed invention. Multiple features are described in the embodiments, but limitation is not made to an invention that requires all such features, and multiple such features may be combined as appropriate. Furthermore, in the attached drawings, the same reference numerals are given to the same or similar configurations, and redundant description thereof is omitted.
  • An example in which a data generation apparatus has been applied to an image capture apparatus, such as a digital camera, will be described below.
  • An image capture apparatus 100 of the first embodiment captures a visible light image and an invisible light image and generates supervisory data for a learning model for subject detection processing of an invisible light image based on the invisible light image and a subject detection result (type, position, and size of the subject) of the visible light image. The supervisory data of the present embodiment is data to be used in machine learning of a learning model for subject detection of an invisible light image, and learning processing is executed with supervisory data as input data (an invisible light image) and output data (a detection result of a subject detected from a visible light image).
  • FIG. 1 is a block diagram illustrating a configuration of the image capture apparatus 100 according to the first embodiment.
  • The image capture apparatus 100 includes a first optical system 101 a, a second optical system 101 b, a first image capturing unit 102 a, a second image capturing unit 102 b, a first image processing unit 103 a, a second image processing unit 103 b, a subject detection unit 104, a supervisory data generation determination unit 105, a supervisory data generation unit 106, a supervisory data storage unit 107, a learning unit 108, a display unit 109, a memory 110, and a control unit 111.
  • The first optical system 101 a includes one or more lenses and forms an image of subject image light in a visible light range on the first image capturing unit 102 a. The second optical system 101 b includes one or more lenses and forms an image of subject image light in an invisible light range on the second image capturing unit 102 b. An image in a visible light range is, for example, an image in a wavelength range of 400 to 800 nm. An image in an invisible light range is, for example, a near-infrared image in a near-infrared wavelength range of 800 to 2500 nm but may be a far-infrared image in a far-infrared wavelength range of 4 μm to 1000 μm or an ultraviolet image in an ultraviolet wavelength range of 380 nm or less.
  • The first image capturing unit 102 a includes an image sensor, such as a CMOS sensor, for converting subject image light of a visible light range formed into an image by the first optical system 101 a into an electric signal. The image sensor includes, for example, color filters in an RGB Bayer array. The first image capturing unit 102 a includes an AD converter for converting an analog electric signal into a digital signal.
  • The second image capturing unit 102 b includes an image sensor, such as a CMOS, for converting subject image light of an invisible light range formed into an image by the second optical system 101 b into an electric signal. The image sensor includes, for example, a color filter that passes light of a near-infrared wavelength range. The second image capturing unit 102 b includes an AD converter for converting an analog electric signal into a digital signal.
  • The first image processing unit 103 a includes a processor (GPU) for executing predetermined image processing on a digital signal obtained by the first image capturing unit 102 a and generates visible light image data. The predetermined image processing includes, for example, distortion correction processing, noise removal processing, exposure correction processing, white balance processing, and edge enhancement processing.
  • The second image processing unit 103 b includes a processor (GPU) for executing predetermined image processing on a digital signal obtained by the second image capturing unit 102 b and generates invisible light image data. The predetermined image processing includes, for example, distortion correction processing, noise removal processing, exposure correction processing, white balance processing, and edge enhancement processing.
  • The subject detection unit 104 detects one or more subjects in a visible light image processed by the first image processing unit 103 a. Then, the subject detection unit 104 detects class information indicating the type of the detected subject, position information indicating a position (center coordinates) of the subject, size information indicating a size of the subject, and the like. Subject detection can be realized by image analysis processing or image recognition processing in which a learning model for which learning processing has been performed by machine learning is used. A learning model is, for example, a neural network, and the class information (such as whether a person is included or a car is included), the position information, and the size information of a subject in an image is detected using a learning model trained with supervisory data.
  • Here, processing for generating supervisory data to be used in machine learning of a learning model for subject detection of an invisible light image (hereinafter, abbreviated to invisible light image supervisory data) and processing for learning the learning model using the invisible light image supervisory data of the present embodiment will be described.
  • The invisible light image supervisory data of the present embodiment is generated based on an invisible light image and the class information, the position information, and the size information of a subject obtained from a visible light image.
  • FIG. 2 is a diagram illustrating a relationship among a visible light image, an invisible light image, and a detection result of a subject detected in the visible light image according to the present embodiment. In the example of FIG. 2 , a visible light image 201 is an image that has been processed by the first image processing unit 103 a, and the subject detection unit 104 detects a subject 202 for which a subject region 203 has been surrounded by a rectangular frame. A subject detection result 204 includes the class information, the size information (X, Y), and the position information (center coordinates) of the subject.
  • An invisible light image 205 is an image that has been processed by the second image processing unit 103 b and is an image that has been captured at the same angle of view as that of the visible light image 201. A subject 206 is a subject that corresponds to the subject 202 of the visible light image 201. A subject region 207 and a subject detection result 208 are the same as the subject region 203 and the subject detection result 204 of the visible light image.
  • The supervisory data generation determination unit 105 determines whether invisible light image supervisory data can be generated based on a subject detection result obtained by the subject detection unit 104 and an invisible light image obtained by the second image processing unit 103 b. Regarding a determination method, for example, a histogram of luminance values of an invisible light image in the same region as a subject region of a visible light image obtained by the subject detection unit 104 is generated, and when the luminance values are within a predetermined range, it is determined that supervisory data can be generated and when the luminance values are outside of the predetermined range, it is determined that supervisory data cannot be generated. Regarding the luminance values, a histogram generated based on the entire image may be used. The supervisory data generation determination unit 105 includes a subject type designation unit (not illustrated) and can determine whether supervisory data for a pre-designated subject can be generated by a subject type for which to generate supervisory data being set by a user operation or a subject type for which to generate supervisory data being selected by a user operation in superimposed display of an invisible light image and a subject detection result, which will be described later in FIG. 5 .
  • When it is determined by the supervisory data generation determination unit 105 that supervisory data can be generated, the supervisory data generation unit 106 generates invisible light image supervisory data based on an invisible light image and a subject detection result of the subject detection unit 104. The supervisory data generation unit 106 includes an image processing unit (not illustrated) and can generate a plurality of pieces of supervisory data from a single invisible light image by executing specific image processing on the invisible light image. The specific image processing includes, for example, at least one of processing for creating bokeh, processing of creating blurring, and processing for correcting luminance.
  • FIG. 3 illustrates processing 301 for correcting luminance, processing 302 for creating blurring, and processing 303 for creating bokeh as examples of specific image processing to be performed on an invisible light image. In the example of FIG. 3 , an example in which image processing is performed only on a subject region is illustrated; however, image processing may be performed on the entire image.
  • The supervisory data storage unit 107 stores invisible light image supervisory data generated by the supervisory data generation unit 106. The learning unit 108 performs learning processing by machine learning on a learning model for subject detection processing of an invisible light image using invisible light image supervisory data stored in the supervisory data storage unit 107. The learning processing of the learning unit 108 is executed with the supervisory data as input data and output data of the learning model for subject detection processing of an invisible light image. A graphics processing unit (GPU) can perform efficient computation by processing more data in parallel processing; therefore, it is useful to perform processing using the GPU when performing learning processing a plurality of times using a learning model, such as in machine learning. In the present embodiment, regarding the learning processing, computation may be performed by the control unit 111, which will be described later, and the GPU cooperating with each other or computation may be performed by the control unit 111 or the GPU alone. In the present embodiment, a configuration is taken such that the image capture apparatus 100 includes the learning unit 108; however, the learning unit 108 may be configured to be separate from the image capture apparatus 100.
  • As illustrated in FIG. 4 , the display unit 109 displays a subject detection result 401 obtained by the subject detection unit 104 in a superimposed manner on one or both of the visible light image obtained by the first image processing unit 103 a and the invisible light image obtained by the second image processing unit 103 b. The displayed subject detection result 401 is a rectangular frame surrounding the subject region. As illustrated in FIG. 5 , the display unit 109 displays the invisible light image supervisory data obtained by the supervisory data generation unit 106 for preview. In the example of FIG. 5 , the class information, the size information, and the position information of the subject are displayed as invisible light image supervisory data 501 in a superimposed manner on the visible light image or the invisible light image for preview. This makes it possible for the user to visually recognize the subject detected in the visible light image in a state in which it is superimposed on the visible light image and the invisible light image as well as confirm the generated supervisory data. In addition, the user can change the class information of the subject, which serves as the supervisory data, and change the determination result as to whether supervisory data can be generated by an operation on the display screen.
  • The memory 110 includes a non-volatile memory (ROM), a volatile memory (RAM), and the like and stores a control program for controlling the overall operation of the image capture apparatus 100 and various parameters.
  • The control unit 111 includes a processor (Central Processing Unit: CPU) for controlling the entire operation of the image capture apparatus 100 by executing the program stored in the memory 110.
  • Next, the invisible light image supervisory data of the present embodiment is generated. Supervisory data generation processing will be described.
  • FIG. 6 is a flowchart illustrating processing for generating invisible light image supervisory data the present embodiment.
  • The processing of FIG. 6 is realized by the control unit 111 of the image capture apparatus 100 illustrated in FIG. 1 controlling respective components by executing the program stored in the memory 110.
  • In step S601, the control unit 111 generates a visible light image and an invisible light image of the same angle of view by capturing the visible light image and the invisible light image simultaneously or consecutively using the first image capturing unit 102 a and the second image capturing unit 102 b. Further, the control unit 111 executes predetermined image processing on the visible light image and the invisible light image using the first image processing unit 103 a and the second image processing unit 103 b.
  • In step S602, the control unit 111 performs subject detection processing on the visible light image obtained in step S601 using the subject detection unit 104. Then, the control unit 111 determines whether a subject has been detected and advances the processing to step S603 when the control unit 111 determines that a subject has been detected and terminates the processing when the control unit 111 determines that a subject has not been detected.
  • In step S603, the control unit 111 determines whether invisible light image supervisory data can be generated by the supervisory data generation determination unit 105 based on the subject detection result of step S602 and the invisible light image obtained in step S601. The control unit 111 advances the processing to step S604 when the control unit 111 determines that the supervisory data can be generated and terminates the processing when the control unit 111 determines that the supervisory data cannot be generated.
  • In step S604, the control unit 111 generates invisible light image supervisory data using the supervisory data generation unit 106 based on the subject detection result of step S602 and the invisible light image obtained in step S601.
  • As described above, according to the first embodiment, supervisory data to be used for machine learning of a learning model for subject detection of an invisible light image can be efficiently and accurately generated based on an invisible light image and a detection result of a subject in a visible light image. This makes it possible to generate a large amount of supervisory data for machine learning of a learning model for subject detection processing of an invisible light image.
  • Second Embodiment
  • Next, a second embodiment will be described.
  • In the second embodiment, an example in which a system for which the image capture apparatus 100 of the first embodiment is separated into an image capture apparatus 710 for generating a visible light image and an invisible light image and a supervisory data generation apparatus 720 for generating invisible light image supervisory data is configured will be described.
  • In the following, parts different from the first embodiment will be mainly described, and description of common parts will be omitted.
  • FIG. 7 is a block diagram illustrating a configuration of a system 700 according to the second embodiment.
  • The system 700 of the second embodiment includes the image capture apparatus 710 and the supervisory data generation apparatus 720.
  • The image capture apparatus 710 includes a first optical system 711 a, a second optical system 711 b, a first image capturing unit 712 a, a second image capturing unit 712 b, a first image processing unit 713 a, a second image processing unit 713 b, a memory 714, and a control unit 715.
  • The image capture apparatus 710 is, for example, a monitor camera or a fixed-point camera capable of simultaneously or consecutively capturing a visible light image and an invisible light image of the same angle of view. The image capture apparatus 710 is capable of capturing a visible light image and an invisible light image at a specific timing, continuously or at regular intervals. The first optical system 711 a, the second optical system 711 b, the first image capturing unit 712 a, the second image capturing unit 712 b, the first image processing unit 713 a, and the second image processing unit 713 b are similar to the first optical system 101 a, the second optical system 101 b, the first image capturing unit 102 a the second image capturing unit 102 b, the first image processing unit 103 a, and the second image processing unit 103 b of the first embodiment.
  • The memory 714 includes a non-volatile memory (ROM), a volatile memory (RAM), and the like and stores a control program for controlling the overall operation of the image capture apparatus 710 and various parameters.
  • The control unit 715 includes a processor (CPU) for controlling the entire operation of the image capture apparatus 710 by executing the program stored in the memory 714.
  • The supervisory data generation apparatus 720 includes a first image acquiring unit 721 a, a second image acquiring unit 721 b, a subject detection unit 722, a supervisory data generation determination unit 723, a supervisory data generation unit 724, a supervisory data storage unit 725, a learning unit 726, a display unit 727, a memory 728, and a control unit 729.
  • The supervisory data generation apparatus 720 is, for example, a server connected to the image capture apparatus 710 via a network so as to be capable of communication.
  • The first image acquiring unit 721 a acquires a visible light image from the image capture apparatus 710. The second image acquiring unit 721 b acquires an invisible light image from the image capture apparatus 710. The visible light image and the invisible light image are images of the same angle of view captured simultaneously or consecutively.
  • The subject detection unit 722, the supervisory data generation determination unit 723, the supervisory data generation unit 724, the supervisory data storage unit 725, and the learning unit 726 are similar to the subject detection unit 104, the supervisory data generation determination unit 105, the supervisory data generation unit 106, the supervisory data storage unit 107, and the learning unit 108 of the first embodiment.
  • The display unit 727 displays a subject detection result obtained by the subject detection unit 722 in a superimposed manner on one or both of the visible light image obtained by the first image acquiring unit 721 a and the invisible light image obtained by the second image acquiring unit 721 b. Display contents are similar to those in FIG. 5 of the first embodiment.
  • The memory 728 includes a non-volatile memory (ROM), a volatile memory (RAM), and the like and stores a control program for controlling the overall operation of the supervisory data generation apparatus 720 and various parameters.
  • The control unit 729 includes a processor (CPU) for controlling the entire operation of the supervisory data generation apparatus 720 by executing the program stored in the memory 728.
  • Next, supervisory data generation processing for generating supervisory data of an invisible light image according to the second embodiment will be described.
  • FIG. 8 is a flowchart exemplifying supervisory data generation processing for generating supervisory data of an invisible light image according to the second embodiment.
  • The processing of FIG. 8 is realized by the control unit 715 of the image capture apparatus 710 and the control unit 729 of the supervisory data generation apparatus 720 of the system 700 illustrated in FIG. 7 controlling respective components. In the processing of FIG. 8 , it is assumed that the image capture apparatus 710 and the supervisory data generation apparatus 720 are connected via a network so as to be capable of communication.
  • In step S801, the control unit 715 of the image capture apparatus 710 generates a visible light image and an invisible light image of the same angle of view by simultaneously or consecutively capturing the visible light image and the invisible light image using the image capture apparatus 710 and executes predetermined image processing on the visible light image and the invisible light image.
  • In step S802, the control unit 715 of the image capture apparatus 710 transmits the visible light image and the invisible light image obtained in step S801 to the supervisory data generation apparatus 720.
  • In step S803, the control unit 729 of the supervisory data generation apparatus 720 acquires the visible light image and the invisible light image transmitted from the image capture apparatus 710 in step S802 using the first image acquiring unit 721 a and the second image acquiring unit 721 b.
  • In step S804, the control unit 729 of the supervisory data generation apparatus 720 performs subject detection processing on the visible light image obtained in step S803 using the subject detection unit 722. The control unit 729 of the supervisory data generation apparatus 720 determines whether a subject has been detected by the subject detection unit 722 and advances the processing to step S805 when the control unit 111 determines that a subject has been detected and terminates the processing when the control unit 111 determines that a subject has not been detected.
  • In step S805, the control unit 729 of the supervisory data generation apparatus 720 determines whether invisible light image supervisory data can be generated using the supervisory data generation determination unit 723 and advances the processing to step S806 when the control unit 111 determines that the supervisory data can be generated and terminates the processing when the control unit 111 determines that the supervisory data cannot be generated.
  • In step S806, the control unit 729 of the supervisory data generation apparatus 720 generates invisible light image supervisory data based the subject detection result obtained in step S804 and the invisible light image acquired in step S803 using the supervisory data generation unit 724.
  • As described above, according to the second embodiment, by configuring a system in which the image capture apparatus 710 and the supervisory data generation apparatus 720 are separate, it is possible to contribute to versatile and efficient generation of supervisory data, such as setting the supervisory data generation apparatus 720 to be a server.
  • In the second embodiment, a configuration is taken such that the supervisory data generation apparatus 720 acquires a visible light image and an invisible light image generated by a single image capture apparatus 710; however, the supervisory data generation apparatus 720 may acquire a visible light image and an invisible light image of the same angle of view generated by a plurality of image capture apparatuses. For example, the supervisory data generation apparatus 720 may acquire a visible light image and an invisible light image of the same angle of view generated by a plurality of monitoring cameras or the like installed at different positions.
  • OTHER EMBODIMENTS
  • Embodiment(s) of the present invention can also be realized by a computer of a system or apparatus that reads out and executes computer executable instructions (e.g., one or more programs) recorded on a storage medium (which may also be referred to more fully as a ‘non-transitory computer-readable storage medium’) to perform the functions of one or more of the above-described embodiment(s) and/or that includes one or more circuits (e.g., application specific integrated circuit (ASIC)) for performing the functions of one or more of the above-described embodiment(s), and by a method performed by the computer of the system or apparatus by, for example, reading out and executing the computer executable instructions from the storage medium to perform the functions of one or more of the above-described embodiment(s) and/or controlling the one or more circuits to perform the functions of one or more of the above-described embodiment(s). The computer may comprise one or more processors (e.g., central processing unit (CPU), micro processing unit (MPU)) and may include a network of separate computers or separate processors to read out and execute the computer executable instructions. The computer executable instructions may be provided to the computer, for example, from a network or the storage medium. The storage medium may include, for example, one or more of a hard disk, a random-access memory (RAM), a read only memory (ROM), a storage of distributed computing systems, an optical disk (such as a compact disc (CD), digital versatile disc (DVD), or Blu-ray Disc (BD)™), a flash memory device, a memory card, and the like.
  • While the present invention has been described with reference to exemplary embodiments, it is to be understood that the invention is not limited to the disclosed exemplary embodiments. The scope of the following claims is to be accorded the broadest interpretation so as to encompass all such modifications and equivalent structures and functions.
  • This application claims the benefit of Japanese Patent Application No. 2022-179090, filed Nov. 8, 2022 which is hereby incorporated by reference herein in its entirety.

Claims (13)

What is claimed is:
1. A data generation apparatus comprising:
a first image acquiring unit that acquires a visible light image;
a second image acquiring unit that acquires an invisible light image;
a subject detection unit that detects a subject in the visible light image; and
a supervisory data generation unit that generates supervisory data to be used in learning of a learning model for subject detection of an invisible light image based on the invisible light image and a subject detection result acquired by the subject detection unit.
2. The apparatus according to claim 1, wherein
the visible light image and the invisible light image are simultaneously or consecutively captured images of the same angle of view.
3. The apparatus according to claim 1, further comprising a supervisory data generation determination unit that determines whether the supervisory data can be generated.
4. The apparatus according to claim 3, wherein
the supervisory data generation determination unit performs the determination based on the invisible light image.
5. The apparatus according to claim 4, wherein
the supervisory data generation determination unit performs the determination based on luminance values of the invisible light image in the same region as a region of the subject detected in the visible light image.
6. The apparatus according to claim 3, wherein
the supervisory data generation determination unit performs the determination based on a predesignated type of subject.
7. The apparatus according to claim 1, wherein
the supervisory data generation unit performs at least one of processing for creating bokeh, processing for creating blurring, and processing for correcting luminance on the invisible light image.
8. The apparatus according to claim 1, further comprising:
a display unit that displays a subject detection result acquired by the subject detection unit and a type of detected subject in a superimposed manner on one or both of the visible light image and the invisible light image.
9. The apparatus according to claim 1, wherein
the first image acquiring unit and the second image acquiring unit acquire a visible light image and an invisible light image of the same angle of view that have been simultaneously or consecutively been captured by an image capture apparatus.
10. The apparatus according to claim 1,
wherein the first image acquiring unit is a first image capturing unit for capturing the visible light image and the second image acquiring unit is a second image capturing unit for capturing the invisible light image, and
wherein the first image capturing unit and the second image capturing unit capture a simultaneously or consecutively captured visible light image and invisible light image of the same angle of view.
11. The apparatus according to claim 1, wherein
the invisible light image is one of a near-infrared image, a far-infrared image, and an ultraviolet image.
12. A method of controlling a data generation apparatus, the method comprising:
acquiring a visible light image and an invisible light image;
detecting a subject in the visible light image; and
generating supervisory data to be used in learning of a learning model for subject detection of an invisible light image based on the invisible light image and the subject detection result.
13. A non-transitory computer-readable storage medium storing a program for causing a computer to function as a data generation apparatus comprising:
a first image acquiring unit that acquires a visible light image;
a second image acquiring unit that acquires an invisible light image;
a subject detection unit that detects a subject in the visible light image; and
a supervisory data generation unit that generates supervisory data to be used in learning of a learning model for subject detection of an invisible light image based on the invisible light image and a subject detection result acquired by the subject detection unit.
US18/491,843 2022-11-08 2023-10-23 Data generation apparatus and control method Pending US20240155210A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2022179090A JP2024068553A (en) 2022-11-08 2022-11-08 Data creation device, control method and program
JP2022-179090 2022-11-08

Publications (1)

Publication Number Publication Date
US20240155210A1 true US20240155210A1 (en) 2024-05-09

Family

ID=90928402

Family Applications (1)

Application Number Title Priority Date Filing Date
US18/491,843 Pending US20240155210A1 (en) 2022-11-08 2023-10-23 Data generation apparatus and control method

Country Status (2)

Country Link
US (1) US20240155210A1 (en)
JP (1) JP2024068553A (en)

Citations (45)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6078682A (en) * 1993-01-01 2000-06-20 Canon Kabushiki Kaisha Image processing apparatus and method, and image reading apparatus
US20040091133A1 (en) * 2002-09-12 2004-05-13 Hitachi Ltd. On board image processing apparatus
US7006705B2 (en) * 2001-03-15 2006-02-28 Canon Kabushiki Kaisha Image processing for correcting defects of read image
US7034927B1 (en) * 2002-06-28 2006-04-25 Digeo, Inc. System and method for identifying an object using invisible light
US7589771B2 (en) * 2004-12-22 2009-09-15 Sony Corporation Image processing apparatus, image processing method, image pickup apparatus, computer program and recording medium
US8077200B2 (en) * 2008-02-05 2011-12-13 Asia Optical Co., Inc. Image capture system and method of capturing labeled image
US8416302B2 (en) * 2009-02-10 2013-04-09 Microsoft Corporation Low-light imaging augmented with non-intrusive lighting
US9392262B2 (en) * 2014-03-07 2016-07-12 Aquifi, Inc. System and method for 3D reconstruction using multiple multi-channel cameras
US9516295B2 (en) * 2014-06-30 2016-12-06 Aquifi, Inc. Systems and methods for multi-channel imaging based on multiple exposure settings
US9799122B2 (en) * 2015-03-09 2017-10-24 Canon Kabushiki Kaisha Motion information acquiring apparatus and motion information acquiring method
US9894255B2 (en) * 2013-06-17 2018-02-13 Industrial Technology Research Institute Method and system for depth selective segmentation of object
US9936145B2 (en) * 2015-02-12 2018-04-03 Canon Kabushiki Kaisha Display control apparatus and display control method
US20180336664A1 (en) * 2017-05-17 2018-11-22 Canon Kabushiki Kaisha Image processing apparatus, image processing method, and storage medium
US10209881B2 (en) * 2012-03-15 2019-02-19 Ibrahim Farid Cherradi El Fadili Extending the free fingers typing technology and introducing the finger taps language technology
US10291859B2 (en) * 2014-11-13 2019-05-14 Panasonic Intellectual Property Management Co., Ltd. Imaging device and imaging method for composing a non-visible light image and a visibile light image
US10356314B2 (en) * 2015-11-25 2019-07-16 Canon Kabushiki Kaisha Image sensor and image capturing apparatus
US10375357B2 (en) * 2014-08-27 2019-08-06 Apple Inc. Method and system for providing at least one image captured by a scene camera of a vehicle
US10395093B2 (en) * 2016-05-16 2019-08-27 Canon Kabushiki Kaisha Image processing apparatus, image processing method, and non-transitory computer-readable storage medium
US10489666B2 (en) * 2017-12-18 2019-11-26 Panasonic Intellectual Property Management Co., Ltd. Imaging device and imaging system
US10605661B2 (en) * 2015-10-05 2020-03-31 Nikon Corporation Image capturing with filters of overlapping passbands
US20210195128A1 (en) * 2018-08-31 2021-06-24 Sony Semiconductor Solutions Corporation Solid-state image capturing device, information processing device, information processing system, information processing method, and computer program
US11045271B1 (en) * 2021-02-09 2021-06-29 Bao Q Tran Robotic medical system
US20210217212A1 (en) * 2017-05-24 2021-07-15 HELLA GmbH & Co. KGaA Method and system for automatically colorizing night-vision images
US11074447B1 (en) * 2018-07-13 2021-07-27 Hana Resources, Inc. Land analysis system using drone-captured data
US11172180B2 (en) * 2018-09-19 2021-11-09 Canon Kabushiki Kaisha Control apparatus, control method and non-transitory computer-readable medium
US11195320B2 (en) * 2019-12-12 2021-12-07 Facebook Technologies, Llc Feed-forward collision avoidance for artificial reality environments
US11317000B2 (en) * 2019-07-16 2022-04-26 Ricoh Company, Ltd. Image processing apparatus having invisible component remover, image processing method, and recording medium
US11361408B2 (en) * 2019-01-21 2022-06-14 Canon Kabushiki Kaisha Image processing apparatus, system, image processing method, and non-transitory computer-readable storage medium
US11389059B2 (en) * 2013-01-25 2022-07-19 Wesley W. O. Krueger Ocular-performance-based head impact measurement using a faceguard
US11405761B2 (en) * 2019-05-23 2022-08-02 Connected Wise LLC On-board machine vision device for activating vehicular messages from traffic signs
US11423524B2 (en) * 2019-03-11 2022-08-23 Canon Kabushiki Kaisha Image processing apparatus, method for controlling image processing apparatus, and non- transitory computer-readable storage medium
US11483503B2 (en) * 2019-01-20 2022-10-25 Magik Eye Inc. Three-dimensional sensor including bandpass filter having multiple passbands
US11508064B2 (en) * 2020-01-30 2022-11-22 Fujitsu Limited Computer-readable recording medium having stored therein information processing program, method for processing information, and information processing apparatus
US20230041560A1 (en) * 2021-08-03 2023-02-09 Intrinsic Innovation Llc Systems and methods for training pose estimators in computer vision
US11590650B2 (en) * 2019-06-10 2023-02-28 Preferred Networks, Inc. Generation method for training dataset, model generation method, training data generation apparatus, inference apparatus, robotic controller, model training method and robot
US20230070798A1 (en) * 2020-03-31 2023-03-09 Sony Group Corporation Image processing apparatus and method, program, and image processing system
US11645864B2 (en) * 2020-11-06 2023-05-09 Hitachi, Ltd. Imaging device, authentication device, and biometric imaging method
US11643014B2 (en) * 2018-07-31 2023-05-09 Sony Semiconductor Solutions Corporation Image capturing device and vehicle control system
US20230309364A1 (en) * 2020-09-11 2023-09-28 Semiconductor Energy Laboratory Co., Ltd. Display device, display module, and electronic device
US11797863B2 (en) * 2020-01-30 2023-10-24 Intrinsic Innovation Llc Systems and methods for synthesizing data for training statistical models on different imaging modalities including polarized images
US11836179B1 (en) * 2019-10-29 2023-12-05 Meta Platforms Technologies, Llc Multimedia query system
US20240020968A1 (en) * 2020-10-08 2024-01-18 Edgy Bees Ltd. Improving geo-registration using machine-learning based object identification
US12104911B2 (en) * 2021-03-04 2024-10-01 Nec Corporation Of America Imperceptible road markings to support automated vehicular systems
US12150367B2 (en) * 2020-09-11 2024-11-19 Semiconductor Energy Laboratory Co., Ltd. Display device, display module, and electronic device
US12250489B2 (en) * 2020-04-27 2025-03-11 Sony Group Corporation Information processing device and method for generating composite video

Patent Citations (45)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6078682A (en) * 1993-01-01 2000-06-20 Canon Kabushiki Kaisha Image processing apparatus and method, and image reading apparatus
US7006705B2 (en) * 2001-03-15 2006-02-28 Canon Kabushiki Kaisha Image processing for correcting defects of read image
US7034927B1 (en) * 2002-06-28 2006-04-25 Digeo, Inc. System and method for identifying an object using invisible light
US20040091133A1 (en) * 2002-09-12 2004-05-13 Hitachi Ltd. On board image processing apparatus
US7589771B2 (en) * 2004-12-22 2009-09-15 Sony Corporation Image processing apparatus, image processing method, image pickup apparatus, computer program and recording medium
US8077200B2 (en) * 2008-02-05 2011-12-13 Asia Optical Co., Inc. Image capture system and method of capturing labeled image
US8416302B2 (en) * 2009-02-10 2013-04-09 Microsoft Corporation Low-light imaging augmented with non-intrusive lighting
US10209881B2 (en) * 2012-03-15 2019-02-19 Ibrahim Farid Cherradi El Fadili Extending the free fingers typing technology and introducing the finger taps language technology
US11389059B2 (en) * 2013-01-25 2022-07-19 Wesley W. O. Krueger Ocular-performance-based head impact measurement using a faceguard
US9894255B2 (en) * 2013-06-17 2018-02-13 Industrial Technology Research Institute Method and system for depth selective segmentation of object
US9392262B2 (en) * 2014-03-07 2016-07-12 Aquifi, Inc. System and method for 3D reconstruction using multiple multi-channel cameras
US9516295B2 (en) * 2014-06-30 2016-12-06 Aquifi, Inc. Systems and methods for multi-channel imaging based on multiple exposure settings
US10375357B2 (en) * 2014-08-27 2019-08-06 Apple Inc. Method and system for providing at least one image captured by a scene camera of a vehicle
US10291859B2 (en) * 2014-11-13 2019-05-14 Panasonic Intellectual Property Management Co., Ltd. Imaging device and imaging method for composing a non-visible light image and a visibile light image
US9936145B2 (en) * 2015-02-12 2018-04-03 Canon Kabushiki Kaisha Display control apparatus and display control method
US9799122B2 (en) * 2015-03-09 2017-10-24 Canon Kabushiki Kaisha Motion information acquiring apparatus and motion information acquiring method
US10605661B2 (en) * 2015-10-05 2020-03-31 Nikon Corporation Image capturing with filters of overlapping passbands
US10356314B2 (en) * 2015-11-25 2019-07-16 Canon Kabushiki Kaisha Image sensor and image capturing apparatus
US10395093B2 (en) * 2016-05-16 2019-08-27 Canon Kabushiki Kaisha Image processing apparatus, image processing method, and non-transitory computer-readable storage medium
US20180336664A1 (en) * 2017-05-17 2018-11-22 Canon Kabushiki Kaisha Image processing apparatus, image processing method, and storage medium
US20210217212A1 (en) * 2017-05-24 2021-07-15 HELLA GmbH & Co. KGaA Method and system for automatically colorizing night-vision images
US10489666B2 (en) * 2017-12-18 2019-11-26 Panasonic Intellectual Property Management Co., Ltd. Imaging device and imaging system
US11074447B1 (en) * 2018-07-13 2021-07-27 Hana Resources, Inc. Land analysis system using drone-captured data
US11643014B2 (en) * 2018-07-31 2023-05-09 Sony Semiconductor Solutions Corporation Image capturing device and vehicle control system
US20210195128A1 (en) * 2018-08-31 2021-06-24 Sony Semiconductor Solutions Corporation Solid-state image capturing device, information processing device, information processing system, information processing method, and computer program
US11172180B2 (en) * 2018-09-19 2021-11-09 Canon Kabushiki Kaisha Control apparatus, control method and non-transitory computer-readable medium
US11483503B2 (en) * 2019-01-20 2022-10-25 Magik Eye Inc. Three-dimensional sensor including bandpass filter having multiple passbands
US11361408B2 (en) * 2019-01-21 2022-06-14 Canon Kabushiki Kaisha Image processing apparatus, system, image processing method, and non-transitory computer-readable storage medium
US11423524B2 (en) * 2019-03-11 2022-08-23 Canon Kabushiki Kaisha Image processing apparatus, method for controlling image processing apparatus, and non- transitory computer-readable storage medium
US11405761B2 (en) * 2019-05-23 2022-08-02 Connected Wise LLC On-board machine vision device for activating vehicular messages from traffic signs
US11590650B2 (en) * 2019-06-10 2023-02-28 Preferred Networks, Inc. Generation method for training dataset, model generation method, training data generation apparatus, inference apparatus, robotic controller, model training method and robot
US11317000B2 (en) * 2019-07-16 2022-04-26 Ricoh Company, Ltd. Image processing apparatus having invisible component remover, image processing method, and recording medium
US11836179B1 (en) * 2019-10-29 2023-12-05 Meta Platforms Technologies, Llc Multimedia query system
US11195320B2 (en) * 2019-12-12 2021-12-07 Facebook Technologies, Llc Feed-forward collision avoidance for artificial reality environments
US11797863B2 (en) * 2020-01-30 2023-10-24 Intrinsic Innovation Llc Systems and methods for synthesizing data for training statistical models on different imaging modalities including polarized images
US11508064B2 (en) * 2020-01-30 2022-11-22 Fujitsu Limited Computer-readable recording medium having stored therein information processing program, method for processing information, and information processing apparatus
US20230070798A1 (en) * 2020-03-31 2023-03-09 Sony Group Corporation Image processing apparatus and method, program, and image processing system
US12250489B2 (en) * 2020-04-27 2025-03-11 Sony Group Corporation Information processing device and method for generating composite video
US20230309364A1 (en) * 2020-09-11 2023-09-28 Semiconductor Energy Laboratory Co., Ltd. Display device, display module, and electronic device
US12150367B2 (en) * 2020-09-11 2024-11-19 Semiconductor Energy Laboratory Co., Ltd. Display device, display module, and electronic device
US20240020968A1 (en) * 2020-10-08 2024-01-18 Edgy Bees Ltd. Improving geo-registration using machine-learning based object identification
US11645864B2 (en) * 2020-11-06 2023-05-09 Hitachi, Ltd. Imaging device, authentication device, and biometric imaging method
US11045271B1 (en) * 2021-02-09 2021-06-29 Bao Q Tran Robotic medical system
US12104911B2 (en) * 2021-03-04 2024-10-01 Nec Corporation Of America Imperceptible road markings to support automated vehicular systems
US20230041560A1 (en) * 2021-08-03 2023-02-09 Intrinsic Innovation Llc Systems and methods for training pose estimators in computer vision

Also Published As

Publication number Publication date
JP2024068553A (en) 2024-05-20

Similar Documents

Publication Publication Date Title
US11195055B2 (en) Image processing method, image processing apparatus, storage medium, image processing system, and manufacturing method of learnt model
US20150358542A1 (en) Image processing apparatus, image capturing apparatus, image processing method, image capturing method, and non-transitory computer-readable medium for focus bracketing
IES20070820A2 (en) Method for improved red eye detection in ISO images
US10395093B2 (en) Image processing apparatus, image processing method, and non-transitory computer-readable storage medium
JP4442413B2 (en) Image processing apparatus, image processing method, program, and recording medium
US10148862B2 (en) Image capturing apparatus, method for controlling image capturing apparatus focus area display, and storage medium
US9706110B2 (en) Foreign body information detection device and foreign body information detection method of imaging apparatus
JP5949559B2 (en) Image processing apparatus, imaging apparatus, and image processing program
US20160381342A1 (en) Image processing apparatus, imaging apparatus, image processing method, and recording medium
JP2008092299A (en) Electronic camera
US11575841B2 (en) Information processing apparatus, imaging apparatus, method, and storage medium
JP2012085083A (en) Image processing apparatus, image pickup device, and image processing program
US20240155210A1 (en) Data generation apparatus and control method
JP2017098900A (en) Image processing apparatus, image processing method, and program
JP2008199164A (en) Image processing method for noise reduction and imaging apparatus using the same
US9648232B2 (en) Image processing apparatus, image capturing apparatus, control method and recording medium
US12160668B2 (en) Image capturing system, control method, and storage medium
US12432457B2 (en) Image capturing apparatus and control method thereof and storage medium
US20250267362A1 (en) Image processing apparatus, image capturing apparatus, image processing method, and non-transitory computer-readable storage medium
JP2010219862A (en) Face detection device, and imaging apparatus
US20230199319A1 (en) Image capturing apparatus and control method thereof and storage medium
US20250232463A1 (en) Image processing apparatus, control method of image processing apparatus, and non-transitory computer-readable storage medium
JP2010220122A (en) Image processing apparatus, image capturing apparatus, image processing method, and image processing program
US20240187727A1 (en) Image processing apparatus, image capturing apparatus, control method of image processing apparatus, and storage medium
JP5631181B2 (en) Image processing apparatus, control method therefor, and program

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

AS Assignment

Owner name: CANON KABUSHIKI KAISHA, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:KOGAI, YUKIHIRO;REEL/FRAME:065691/0252

Effective date: 20231016

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED