US20200084362A1 - Image processing method and image processing device - Google Patents
Image processing method and image processing device Download PDFInfo
- Publication number
- US20200084362A1 US20200084362A1 US16/292,335 US201916292335A US2020084362A1 US 20200084362 A1 US20200084362 A1 US 20200084362A1 US 201916292335 A US201916292335 A US 201916292335A US 2020084362 A1 US2020084362 A1 US 2020084362A1
- Authority
- US
- United States
- Prior art keywords
- subject
- brightness
- image
- dynamic range
- high dynamic
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/741—Circuitry for compensating brightness variation in the scene by increasing the dynamic range of the image compared to the dynamic range of the electronic image sensors
-
- H04N5/2355—
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration using two or more images, e.g. averaging or subtraction
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/90—Dynamic range modification of images or parts thereof
- G06T5/94—Dynamic range modification of images or parts thereof based on local image properties, e.g. for local contrast enhancement
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/71—Circuitry for evaluating the brightness variation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/743—Bracketing, i.e. taking a series of images with varying exposure conditions
-
- H04N5/2351—
-
- H04N5/2356—
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10141—Special mode during image acquisition
- G06T2207/10144—Varying exposure
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20172—Image enhancement details
- G06T2207/20208—High dynamic range [HDR] image processing
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20212—Image combination
- G06T2207/20221—Image fusion; Image merging
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
Definitions
- the present disclosure relates to an image processing method and an image processing device, and more particularly to an image processing method and an image processing device that can optimize a subject photographed in an image.
- the quality of the captured image may be unsatisfying due to the exposure mechanism which is difficult to be calibrated under the limitations of the dynamic range of the sensors.
- the subject e.g., the portrait
- conventional practice may select a longer exposure setting in order to increase the image brightness of the subject area.
- this method may cause overexposure in other areas of higher brightness in the image due to an increase in exposure time. That is, although this method allows the subject to have a relatively normal brightness in the generated image, it may cause the background to be overexposed, as shown in FIG. 1A .
- the well-known high dynamic range (HDR) technique is a technique for processing multiple images by superimposing images taken based on different exposure values and then adjusting by tone reproduction to produce an HDR image with both highlight and shadow details.
- HDR high dynamic range
- the adjustment of the regional tone reproduction is not only time consuming, but also prone to lead to glows (i.e., an abnormal white or black border at the black and white boundary), as shown in FIG. 1C .
- the brightness of the subject may be lowered as the image is superimposed, which may cause a decrease in the contrast of the subject and an unnatural brightness change.
- a high-HDR image synthesis algorithm is proposed to directly acquire a plurality of original images with different exposures by using a machine learning and image synthesis model, and then synthesize a reference high dynamic range image according to network parameters.
- this prior art could not achieve the effect of improving the image quality of the subject in the HDR image.
- a method for operating a device to capture an HDR image has been proposed to take multiple photos in a normal exposure manner before accepting an HDR command, and obtain an overexposed and underexposed photo after accepting the HDR command. Afterward, the normally exposed, overexposed and underexposed photos are aligned.
- this prior are also failed to achieve the effect of improving the image quality of the subject in the HDR image.
- a conventional imaging technique for skin sectioning has been proposed, which can take a non-HDR photo first and then take an HDR photo of the same location.
- the previous case also failed to achieve the effect of improving the image quality of the subject in the HDR image.
- a conventional method and system for implementing partial HDR has been proposed, in which a user can manually select a partially exposed region and adjust an exposure value for a particular region.
- the mechanism for adjusting the exposure value in this prior art is highly related to the manual operation content of the user, the effect of improving the image quality of the subject in the HDR image cannot be effectively achieved.
- the present disclosure provides an image processing method and an image processing device, which can be used to solve the above technical problems.
- the disclosure provides an image processing method, including: retrieving a first image of a specific scene, and extracting a first subject from the first image; determining a first brightness based on the first subject; using the first brightness as a reference value of a bracketing mechanism, and executing the bracketing mechanism to capture a plurality of second images for the specific scene; synthesizing the second images as a first high dynamic range image; extracting a second subject from the first high dynamic range image, wherein the second subject has a second brightness; optimizing the second subject when the second brightness is lower than the first brightness by a predetermined threshold; and synthesizing the optimized second subject and the first high dynamic range image as a second high dynamic range image.
- the disclosure provides an image processing device, including a non-transitory storage circuit, an image capturing circuit, and a processor.
- the non-transitory storage circuit stores a plurality of modules.
- the processor is coupled to the storage circuit and the image capturing circuit and accesses the modules to perform the following steps: controlling the image capturing circuit to retrieve a first image of a specific scene, and extracting a first subject from the first image; determining a first brightness based on the first subject; using the first brightness as a reference value of a bracketing mechanism, and controlling the image capturing circuit to perform the bracketing mechanism to capture a plurality of second images for the specific scene; synthesizing the second images as a first high dynamic range image; extracting a second subject from the first high dynamic range image, wherein the second subject has a second brightness; optimizing the second subject when the second brightness is lower than the first brightness by a predetermined threshold; synthesizing the optimized second subject and the first high dynamic range image as a second high dynamic range image.
- FIG. 1A is a schematic diagram showing overexposed background.
- FIG. 1B is a schematic diagram showing the subject being too dark.
- FIG. 1C is a schematic diagram showing the glow phenomenon.
- FIG. 2 is a functional block diagram of an image processing device according to an embodiment of the disclosure.
- FIG. 3 is a flowchart of an image processing method according to an embodiment of the present disclosure.
- FIG. 4 is an application scenario for performing subject optimization using the method of FIG. 3 according to an embodiment of the disclosure.
- FIG. 2 is a functional block diagram of an image processing device according to an embodiment of the disclosure.
- the image processing device 200 includes a storage circuit 212 , an image capturing circuit 214 , and a processor 216 .
- the image processing device 200 can be a mobile phone, a smart phone, a personal computer (PC), a notebook PC, a netbook PC, a tablet PC, a digital camera or other similar smart devices, but the disclosure is not limited thereto.
- the storage circuit 212 can be any type of fixed or removable random access memory (RAM), read-only memory (ROM), flash memory, a hard disc or other similar devices that can be used to store computer readable and programmable software, images, programming modules and other information.
- RAM random access memory
- ROM read-only memory
- flash memory a hard disc or other similar devices that can be used to store computer readable and programmable software, images, programming modules and other information.
- the image capturing circuit 214 can be any camera having a charge coupled device (CCD) lens, a complementary metal oxide semiconductor transistor (CMOS) lens, or an infrared lens, or can be an image capturing device that may obtain depth information, such as a depth camera or a stereo camera.
- CCD charge coupled device
- CMOS complementary metal oxide semiconductor transistor
- infrared lens or can be an image capturing device that may obtain depth information, such as a depth camera or a stereo camera.
- the processor 216 is coupled to the storage circuit 212 and the image capturing circuit 214 , and can be a general purpose processor, a special purpose processor, a conventional processor, a digital signal processor, a plurality of microprocessors, or one or more microprocessors combined with digital signal processing cores, controllers, microcontrollers, Application Specific Integrated Circuits (ASICs), Field Programmable Gate Arrays (FPGAs), integrated circuits, state machines, processors based on Advanced RISC Machine (ARM), and the like.
- ASICs Application Specific Integrated Circuits
- FPGAs Field Programmable Gate Arrays
- ARM Advanced RISC Machine
- FIG. 3 is a flowchart of an image processing method according to an embodiment of the present disclosure
- FIG. 4 is an application scenario for performing subject optimization using the method of FIG. 3 according to an embodiment of the present disclosure.
- the method of FIG. 3 can be performed by the image processing device 200 of FIG. 2 , and the details of the steps of FIG. 3 will be described in accompanying with the elements shown in FIG. 2 and the scenario shown in FIG. 4 .
- step S 311 the processor 216 controls the image capturing circuit 214 to retrieve the first image 411 of the specific scene and extracts the first subject 411 a (for example, the human body) from the first image 411 .
- the processor 216 controls the image capturing circuit 214 to retrieve the first image 411 of the specific scene and extracts the first subject 411 a (for example, the human body) from the first image 411 .
- the processor 216 can input the first image 411 into an artificial intelligence module to capture the first subject 411 a based on the artificial intelligence module.
- the artificial intelligence module can train an artificial intelligence model based on a certain amount of training data (for example, training images) in advance.
- the aspects of the subject such as a human body, an animal, a plant, etc.
- the aforementioned artificial intelligence model may be trained to learn that what kind of features (e.g., colors, shapes) are necessary for objects or regions to be recognized as the subject in an image based on the marked aspects of the subjects in the training images.
- the artificial intelligence model when the artificial intelligence model receives an unknown image, it can analyze the features of each object and region in the unknown image to find out which objects and regions may correspond to the subject of the unknown image (e.g., a human body), but the present disclosure is not limited thereto. Therefore, in an embodiment, after the first image 411 is received by the artificial intelligence module, the first subject 411 a in the first image 411 can be identified by the artificial intelligence model.
- the processor 216 can also generate a depth map of the first image 411 and analyze the depth map to find the first subject 411 a therein.
- the image capturing circuit 214 is implemented as a depth camera (such as a time-of-flight (TOF) camera), a stereo camera, a dual lens camera, an infrared depth sensing module, or other devices capable of obtaining the depth of images
- the processor 216 can control the image capturing circuit 214 to generate a depth map of the first image 411 .
- the processor 216 can find an object region that is closer to the image processing device 200 (i.e., shallower in depth) from the depth map.
- the processor 216 may mark the object area as the subject area 411 b as shown in FIG. 4 , that is, the area corresponding to the first subject 411 a , but the disclosure is not limited thereto.
- the processor 216 may determine the first brightness based on the first subject 411 a .
- the processor 216 can retrieve a plurality of pixels included in the first subject 411 a and the brightness of each of the pixels.
- the processor 216 can calculate a brightness average of the brightness of each of the pixels and use the brightness average as the first brightness.
- the first brightness is the average brightness of the first subject 411 a , but the present disclosure is not limited thereto.
- the processor 411 may also use the average brightness of these subjects as the first brightness or the highest brightness among the subjects as the first brightness.
- the first brightness may also be determined by an artificial intelligence module.
- the artificial intelligence module can use a classification module to determine the image classification to which the first image 411 belongs, and then perform a ranking module based on the result of the classification and the first training data (including features and colors). Then, based on the ranked result and the second training data (including, for example, a result of a network vote, a field of view (FOV)), which value is more suitable as the first brightness corresponding to the first image 411 may be determined, but the disclosure is not limited thereto.
- a classification module to determine the image classification to which the first image 411 belongs
- a ranking module based on the result of the classification and the first training data (including features and colors). Then, based on the ranked result and the second training data (including, for example, a result of a network vote, a field of view (FOV)
- FOV field of view
- the processor 216 may use the first brightness as a reference value of a bracketing mechanism and control the image capturing circuit 214 to perform the bracketing mechanism to capture a plurality of second images 412 for the specific scene.
- processor 216 may use the first brightness as the 0 exposure value (EV) (i.e., 0 EV) in the bracketing mechanism. Thereafter, the processor 216 can capture the specific scene based on the ⁇ N exposure value to the +N exposure value to generate the second images 412 , where N is a positive integer.
- EV 0 exposure value
- the processor 216 can respectively capture images based on ⁇ 2EV, ⁇ 1EV, 0EV (which corresponds to the first brightness, that is, the average brightness of the first subject 411 a ), 1EV and 2EV and take the captured images as the second images 412 .
- the second images 412 shown in FIG. 4 are merely examples, which should not be interpreted as limiting possible implementations of the disclosure.
- step S 314 the processor 216 can synthesize the second images 412 as a first HDR image (not shown).
- the processor 216 may superimpose the second images 412 as the first HDR image based on a known HDR image processing technology, and details would not be further discussed herein.
- the processor 216 can extract a second subject (not shown) from the first HDR image.
- the processor 216 can input the first HDR image to the previously mentioned artificial intelligence module to identify the second subject in the first HDR image through the artificial intelligence model therein.
- the processor 216 can also identify the subject region corresponding to the second subject from the depth map corresponding to the first HDR image based on the previously taught manner, thereby defining the second subject in the first HDR image, and the details of which would not be repeated herein.
- the second subject found through the artificial intelligence module and the previously found first subject 411 a should substantially correspond to the same subject (i.e., the human body shown in FIG. 4 ).
- the present disclosure can determine whether the difference between the second brightness of the second subject (for example, the brightness average of the pixels in the region of the second subject) and the first brightness of the first subject 411 a is too large and take it as a reference for determining whether to optimize the second subject.
- the second brightness of the second subject for example, the brightness average of the pixels in the region of the second subject
- the first brightness of the first subject 411 a is too large and take it as a reference for determining whether to optimize the second subject.
- the processor 216 may determine whether the second brightness is lower than the first brightness by a predetermined threshold.
- the designer may set the predetermined threshold value to a desired value (for example, 10%, 20%, 30%, etc.), but the disclosure is not limited thereto.
- the processor 216 can proceed to perform step S 317 to optimize the second subject.
- the processor 216 can adjust the second brightness of the second subject to be the first brightness of the first subject 411 a to optimize the second subject, thereby making the second subject become brighter.
- the processor 216 may also adjust the second brightness of the second subject by any scale value set by the designer (e.g., 10%, 20%, 30%, etc.) to optimize the second subject in a manner of quantitatively increasing the brightness of the second subject, but the disclosure is not limited thereto.
- the processor 216 may proceed to perform step S 318 to synthesize the optimized second subject 413 a and the first HDR image as the second HDR image 413 .
- the processor 216 may superimpose the optimized second subject 413 a on the second subject of the first HDR image to generate the second HDR image 413 . That is, the image processing device 200 can use the second HDR image 413 as the final output image for the user to view.
- the image quality of the second HDR image 413 can be further improved due to the optimized second subject 413 a for having improved brightness.
- the processor 216 can proceed to perform step S 319 to output the first HDR image. In other words, if the brightness of the second subject does not decrease significantly, the processor 216 may directly use the first HDR image as the final output image without optimizing the second subject.
- the present disclosure provides an image processing method and an image processing device, which can find a first subject from a first image based on an artificial intelligence module or a depth map, and use the average brightness of the first subject as the reference value of the bracketing mechanism. Then, after generating the first HDR image based on the bracketing mechanism, the present disclosure may optimize the second subject when the difference between the brightness of the second subject in the first HDR image and the brightness of the first subject is too large, and accordingly generate a second HDR image. Thereby, in addition to providing clear highlight and shadow details, the second subject of the second HDR image can provide a better stereoscopic effect and a more natural brightness change after being optimized.
- the present disclosure can automatically recognize the first subject and the second subject through artificial intelligence, not only the operation process can become more efficient, but the outputted HDR image can be guaranteed to have a satisfying image quality.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Studio Devices (AREA)
- Image Processing (AREA)
Abstract
Description
- This application claims the priority benefit of Taiwan application serial no. 107131335, filed on Sep. 6, 2018. The entirety of the above-mentioned patent application is hereby incorporated by reference herein and made a part of this specification.
- The present disclosure relates to an image processing method and an image processing device, and more particularly to an image processing method and an image processing device that can optimize a subject photographed in an image.
- When shooting a scene with high contrast brightness (for example, backlighting scene) in the traditional single image capturing technology, the quality of the captured image may be unsatisfying due to the exposure mechanism which is difficult to be calibrated under the limitations of the dynamic range of the sensors. For example, in the case where the subject (e.g., the portrait) is dark in the image, conventional practice may select a longer exposure setting in order to increase the image brightness of the subject area. However, this method may cause overexposure in other areas of higher brightness in the image due to an increase in exposure time. That is, although this method allows the subject to have a relatively normal brightness in the generated image, it may cause the background to be overexposed, as shown in
FIG. 1A . - On the other hand, if the exposure setting is changed to focus on the bright area, in order to avoid overexposure in the bright area, a shorter exposure time will be selected accordingly. However, this may instead cause the image subject area to be too dark due to underexposure, as shown in
FIG. 1B . In other words, no matter what exposure setting is used, it is difficult to produce images with clear subjects and backgrounds. - The well-known high dynamic range (HDR) technique is a technique for processing multiple images by superimposing images taken based on different exposure values and then adjusting by tone reproduction to produce an HDR image with both highlight and shadow details. Through the HDR mechanism, the dynamic range of the scene can be restored more accurately while avoiding areas where the image is overexposed or too dark and retaining details of the bright and dark areas.
- However, the adjustment of the regional tone reproduction is not only time consuming, but also prone to lead to glows (i.e., an abnormal white or black border at the black and white boundary), as shown in
FIG. 1C . Moreover, in the HDR image generated by the conventional HDR image processing technology, the brightness of the subject may be lowered as the image is superimposed, which may cause a decrease in the contrast of the subject and an unnatural brightness change. - Conventionally, a high-HDR image synthesis algorithm is proposed to directly acquire a plurality of original images with different exposures by using a machine learning and image synthesis model, and then synthesize a reference high dynamic range image according to network parameters. However, this prior art could not achieve the effect of improving the image quality of the subject in the HDR image.
- Also, a method for operating a device to capture an HDR image has been proposed to take multiple photos in a normal exposure manner before accepting an HDR command, and obtain an overexposed and underexposed photo after accepting the HDR command. Afterward, the normally exposed, overexposed and underexposed photos are aligned. However, this prior are also failed to achieve the effect of improving the image quality of the subject in the HDR image.
- A conventional imaging technique for skin sectioning has been proposed, which can take a non-HDR photo first and then take an HDR photo of the same location. However, the previous case also failed to achieve the effect of improving the image quality of the subject in the HDR image.
- A conventional method and system for implementing partial HDR has been proposed, in which a user can manually select a partially exposed region and adjust an exposure value for a particular region. However, since the mechanism for adjusting the exposure value in this prior art is highly related to the manual operation content of the user, the effect of improving the image quality of the subject in the HDR image cannot be effectively achieved.
- Accordingly, the present disclosure provides an image processing method and an image processing device, which can be used to solve the above technical problems.
- The disclosure provides an image processing method, including: retrieving a first image of a specific scene, and extracting a first subject from the first image; determining a first brightness based on the first subject; using the first brightness as a reference value of a bracketing mechanism, and executing the bracketing mechanism to capture a plurality of second images for the specific scene; synthesizing the second images as a first high dynamic range image; extracting a second subject from the first high dynamic range image, wherein the second subject has a second brightness; optimizing the second subject when the second brightness is lower than the first brightness by a predetermined threshold; and synthesizing the optimized second subject and the first high dynamic range image as a second high dynamic range image.
- The disclosure provides an image processing device, including a non-transitory storage circuit, an image capturing circuit, and a processor. The non-transitory storage circuit stores a plurality of modules. The processor is coupled to the storage circuit and the image capturing circuit and accesses the modules to perform the following steps: controlling the image capturing circuit to retrieve a first image of a specific scene, and extracting a first subject from the first image; determining a first brightness based on the first subject; using the first brightness as a reference value of a bracketing mechanism, and controlling the image capturing circuit to perform the bracketing mechanism to capture a plurality of second images for the specific scene; synthesizing the second images as a first high dynamic range image; extracting a second subject from the first high dynamic range image, wherein the second subject has a second brightness; optimizing the second subject when the second brightness is lower than the first brightness by a predetermined threshold; synthesizing the optimized second subject and the first high dynamic range image as a second high dynamic range image.
- The accompanying drawings are included to provide a further understanding of the invention, and are incorporated in and constitute a part of this specification. The drawings illustrate embodiments of the invention and, together with the description, serve to explain the principles of the invention.
-
FIG. 1A is a schematic diagram showing overexposed background. -
FIG. 1B is a schematic diagram showing the subject being too dark. -
FIG. 1C is a schematic diagram showing the glow phenomenon. -
FIG. 2 is a functional block diagram of an image processing device according to an embodiment of the disclosure. -
FIG. 3 is a flowchart of an image processing method according to an embodiment of the present disclosure. -
FIG. 4 is an application scenario for performing subject optimization using the method ofFIG. 3 according to an embodiment of the disclosure. - Reference will now be made in detail to the present preferred embodiments of the invention, examples of which are illustrated in the accompanying drawings. Wherever possible, the same reference numbers are used in the drawings and the description to refer to the same or like parts.
- Please refer to
FIG. 2 , which is a functional block diagram of an image processing device according to an embodiment of the disclosure. In the embodiment, theimage processing device 200 includes astorage circuit 212, an image capturingcircuit 214, and aprocessor 216. - In various embodiments, the
image processing device 200 can be a mobile phone, a smart phone, a personal computer (PC), a notebook PC, a netbook PC, a tablet PC, a digital camera or other similar smart devices, but the disclosure is not limited thereto. - The
storage circuit 212 can be any type of fixed or removable random access memory (RAM), read-only memory (ROM), flash memory, a hard disc or other similar devices that can be used to store computer readable and programmable software, images, programming modules and other information. - The image capturing
circuit 214 can be any camera having a charge coupled device (CCD) lens, a complementary metal oxide semiconductor transistor (CMOS) lens, or an infrared lens, or can be an image capturing device that may obtain depth information, such as a depth camera or a stereo camera. - The
processor 216 is coupled to thestorage circuit 212 and theimage capturing circuit 214, and can be a general purpose processor, a special purpose processor, a conventional processor, a digital signal processor, a plurality of microprocessors, or one or more microprocessors combined with digital signal processing cores, controllers, microcontrollers, Application Specific Integrated Circuits (ASICs), Field Programmable Gate Arrays (FPGAs), integrated circuits, state machines, processors based on Advanced RISC Machine (ARM), and the like. - See
FIG. 3 andFIG. 4 , whereinFIG. 3 is a flowchart of an image processing method according to an embodiment of the present disclosure, andFIG. 4 is an application scenario for performing subject optimization using the method ofFIG. 3 according to an embodiment of the present disclosure. The method ofFIG. 3 can be performed by theimage processing device 200 ofFIG. 2 , and the details of the steps ofFIG. 3 will be described in accompanying with the elements shown inFIG. 2 and the scenario shown inFIG. 4 . - In step S311, the
processor 216 controls theimage capturing circuit 214 to retrieve thefirst image 411 of the specific scene and extracts thefirst subject 411 a (for example, the human body) from thefirst image 411. - In an embodiment, the
processor 216 can input thefirst image 411 into an artificial intelligence module to capture thefirst subject 411 a based on the artificial intelligence module. The artificial intelligence module can train an artificial intelligence model based on a certain amount of training data (for example, training images) in advance. In different embodiments, the aspects of the subject (such as a human body, an animal, a plant, etc.) that the designer wants to consider may be marked manually or otherwise in the aforementioned training images, so that the aforementioned artificial intelligence model may be trained to learn that what kind of features (e.g., colors, shapes) are necessary for objects or regions to be recognized as the subject in an image based on the marked aspects of the subjects in the training images. After the above training process is completed, when the artificial intelligence model receives an unknown image, it can analyze the features of each object and region in the unknown image to find out which objects and regions may correspond to the subject of the unknown image (e.g., a human body), but the present disclosure is not limited thereto. Therefore, in an embodiment, after thefirst image 411 is received by the artificial intelligence module, the first subject 411 a in thefirst image 411 can be identified by the artificial intelligence model. - In another embodiment, the
processor 216 can also generate a depth map of thefirst image 411 and analyze the depth map to find the first subject 411 a therein. For example, when theimage capturing circuit 214 is implemented as a depth camera (such as a time-of-flight (TOF) camera), a stereo camera, a dual lens camera, an infrared depth sensing module, or other devices capable of obtaining the depth of images, theprocessor 216 can control theimage capturing circuit 214 to generate a depth map of thefirst image 411. Thereafter, theprocessor 216 can find an object region that is closer to the image processing device 200 (i.e., shallower in depth) from the depth map. Next, theprocessor 216 may mark the object area as thesubject area 411 b as shown inFIG. 4 , that is, the area corresponding to the first subject 411 a, but the disclosure is not limited thereto. - In step S312, the
processor 216 may determine the first brightness based on the first subject 411 a. In an embodiment, theprocessor 216 can retrieve a plurality of pixels included in the first subject 411 a and the brightness of each of the pixels. Next, theprocessor 216 can calculate a brightness average of the brightness of each of the pixels and use the brightness average as the first brightness. In other words, the first brightness is the average brightness of the first subject 411 a, but the present disclosure is not limited thereto. In other embodiments, if there are multiple subjects in thefirst image 411, theprocessor 411 may also use the average brightness of these subjects as the first brightness or the highest brightness among the subjects as the first brightness. - In other embodiments, the first brightness may also be determined by an artificial intelligence module. Specifically, the artificial intelligence module can use a classification module to determine the image classification to which the
first image 411 belongs, and then perform a ranking module based on the result of the classification and the first training data (including features and colors). Then, based on the ranked result and the second training data (including, for example, a result of a network vote, a field of view (FOV)), which value is more suitable as the first brightness corresponding to thefirst image 411 may be determined, but the disclosure is not limited thereto. - In step S313, the
processor 216 may use the first brightness as a reference value of a bracketing mechanism and control theimage capturing circuit 214 to perform the bracketing mechanism to capture a plurality ofsecond images 412 for the specific scene. In an embodiment,processor 216 may use the first brightness as the 0 exposure value (EV) (i.e., 0 EV) in the bracketing mechanism. Thereafter, theprocessor 216 can capture the specific scene based on the −N exposure value to the +N exposure value to generate thesecond images 412, where N is a positive integer. For example, if N is 2, theprocessor 216 can respectively capture images based on −2EV, −1EV, 0EV (which corresponds to the first brightness, that is, the average brightness of the first subject 411 a), 1EV and 2EV and take the captured images as thesecond images 412. It should be understood that thesecond images 412 shown inFIG. 4 are merely examples, which should not be interpreted as limiting possible implementations of the disclosure. - In step S314, the
processor 216 can synthesize thesecond images 412 as a first HDR image (not shown). In this embodiment, theprocessor 216 may superimpose thesecond images 412 as the first HDR image based on a known HDR image processing technology, and details would not be further discussed herein. - In step S315, the
processor 216 can extract a second subject (not shown) from the first HDR image. In an embodiment, theprocessor 216 can input the first HDR image to the previously mentioned artificial intelligence module to identify the second subject in the first HDR image through the artificial intelligence model therein. In another embodiment, theprocessor 216 can also identify the subject region corresponding to the second subject from the depth map corresponding to the first HDR image based on the previously taught manner, thereby defining the second subject in the first HDR image, and the details of which would not be repeated herein. - Since the first HDR image and the
first image 411 substantially correspond to the same specific scene, the second subject found through the artificial intelligence module and the previously found first subject 411 a should substantially correspond to the same subject (i.e., the human body shown inFIG. 4 ). - However, as mentioned previously, the second subject in the first HDR image may experience a decrease in contrast due to the superposition of the images, which may cause a decrease in the stereoscopic effect of the second subject and an unnatural change in brightness. Therefore, the present disclosure can determine whether the difference between the second brightness of the second subject (for example, the brightness average of the pixels in the region of the second subject) and the first brightness of the first subject 411 a is too large and take it as a reference for determining whether to optimize the second subject.
- Specifically, in step S316, the
processor 216 may determine whether the second brightness is lower than the first brightness by a predetermined threshold. In different embodiments, the designer may set the predetermined threshold value to a desired value (for example, 10%, 20%, 30%, etc.), but the disclosure is not limited thereto. - If the second brightness is lower than the first brightness by the predetermined threshold value (for example, 10%), it means that the second subject in the first HDR image may have poor image quality such as a decrease in contrast and a decrease in stereoscopic effect. Therefore, the
processor 216 can proceed to perform step S317 to optimize the second subject. - In an embodiment, the
processor 216 can adjust the second brightness of the second subject to be the first brightness of the first subject 411 a to optimize the second subject, thereby making the second subject become brighter. In other embodiments, theprocessor 216 may also adjust the second brightness of the second subject by any scale value set by the designer (e.g., 10%, 20%, 30%, etc.) to optimize the second subject in a manner of quantitatively increasing the brightness of the second subject, but the disclosure is not limited thereto. - After optimizing the second subject, the
processor 216 may proceed to perform step S318 to synthesize the optimized second subject 413 a and the first HDR image as thesecond HDR image 413. In an embodiment, theprocessor 216 may superimpose the optimized second subject 413 a on the second subject of the first HDR image to generate thesecond HDR image 413. That is, theimage processing device 200 can use thesecond HDR image 413 as the final output image for the user to view. - As can be observed in the
second HDR image 413 shown inFIG. 4 , not only the highlight and shadow details can be enhanced in thesecond HDR image 413, but the image quality of thesecond HDR image 413 can be further improved due to the optimized second subject 413 a for having improved brightness. - On the other hand, if the second brightness is determined to be not lower than the first brightness by the predetermined threshold in step S316, it represents that the second subject in the first HDR image may not have issues such as a decrease in contrast or a decrease in stereoscopic effect. Therefore, the
processor 216 can proceed to perform step S319 to output the first HDR image. In other words, if the brightness of the second subject does not decrease significantly, theprocessor 216 may directly use the first HDR image as the final output image without optimizing the second subject. - In summary, the present disclosure provides an image processing method and an image processing device, which can find a first subject from a first image based on an artificial intelligence module or a depth map, and use the average brightness of the first subject as the reference value of the bracketing mechanism. Then, after generating the first HDR image based on the bracketing mechanism, the present disclosure may optimize the second subject when the difference between the brightness of the second subject in the first HDR image and the brightness of the first subject is too large, and accordingly generate a second HDR image. Thereby, in addition to providing clear highlight and shadow details, the second subject of the second HDR image can provide a better stereoscopic effect and a more natural brightness change after being optimized.
- In addition, since the present disclosure can automatically recognize the first subject and the second subject through artificial intelligence, not only the operation process can become more efficient, but the outputted HDR image can be guaranteed to have a satisfying image quality.
- It will be apparent to those skilled in the art that various modifications and variations can be made to the structure of the present invention without departing from the scope or spirit of the invention. In view of the foregoing, it is intended that the present invention cover modifications and variations of this invention provided they fall within the scope of the following claims and their equivalents.
Claims (20)
Applications Claiming Priority (3)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| TW107131335 | 2018-09-06 | ||
| TW107131335A | 2018-09-06 | ||
| TW107131335A TWI695622B (en) | 2018-09-06 | 2018-09-06 | Image processing method and image processing device |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20200084362A1 true US20200084362A1 (en) | 2020-03-12 |
| US10630911B2 US10630911B2 (en) | 2020-04-21 |
Family
ID=69718856
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US16/292,335 Active US10630911B2 (en) | 2018-09-06 | 2019-03-05 | Image processing method and image processing device |
Country Status (3)
| Country | Link |
|---|---|
| US (1) | US10630911B2 (en) |
| CN (1) | CN110881108B (en) |
| TW (1) | TWI695622B (en) |
Cited By (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN114363525A (en) * | 2022-03-17 | 2022-04-15 | 杭州灵西机器人智能科技有限公司 | HDR fast automatic imaging method, structured light camera, electronic device, and storage medium |
| US20230017607A1 (en) * | 2021-07-12 | 2023-01-19 | Panasonic Intellectual Property Management Co., Ltd. | Imaging apparatus |
| US12542975B2 (en) * | 2022-03-07 | 2026-02-03 | Canon Kabushiki Kaisha | Image processing apparatus to generate composite image, control method, and recording medium |
Families Citing this family (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10863105B1 (en) * | 2017-06-27 | 2020-12-08 | Amazon Technologies, Inc. | High dynamic range imaging for event detection and inventory management |
| CN111835982B (en) * | 2020-07-02 | 2021-08-17 | 维沃移动通信有限公司 | Image acquisition method, image acquisition device, electronic device and storage medium |
| CN112822404B (en) * | 2021-01-12 | 2022-08-09 | Oppo广东移动通信有限公司 | Image processing method and device and storage medium |
Family Cites Families (31)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP4461937B2 (en) * | 2003-09-30 | 2010-05-12 | セイコーエプソン株式会社 | Generation of high-resolution images based on multiple low-resolution images |
| JP2007028236A (en) * | 2005-07-19 | 2007-02-01 | Hitachi Ltd | Imaging device |
| JP2007081550A (en) * | 2005-09-12 | 2007-03-29 | Konica Minolta Photo Imaging Inc | Imaging apparatus, and image processing unit, method, and program |
| JP4325703B2 (en) * | 2007-05-24 | 2009-09-02 | ソニー株式会社 | Solid-state imaging device, signal processing device and signal processing method for solid-state imaging device, and imaging device |
| JP2009017078A (en) * | 2007-07-03 | 2009-01-22 | Fujifilm Corp | Digital still camera and operation control method thereof |
| JP5083090B2 (en) * | 2007-09-18 | 2012-11-28 | ソニー株式会社 | Display control apparatus, imaging apparatus, display control method, and program |
| JP5412953B2 (en) * | 2009-05-20 | 2014-02-12 | リコーイメージング株式会社 | Imaging device |
| US9288428B2 (en) * | 2010-05-11 | 2016-03-15 | Olympus Corporation | Shooting apparatus and method for controlling shooting apparatus |
| US8699821B2 (en) | 2010-07-05 | 2014-04-15 | Apple Inc. | Aligning images |
| JP5762756B2 (en) * | 2011-01-20 | 2015-08-12 | オリンパス株式会社 | Image processing apparatus, image processing method, image processing program, and photographing apparatus |
| CN103324033A (en) * | 2012-03-23 | 2013-09-25 | 盛乐信息技术(上海)有限公司 | Exposure method and exposure system |
| US9648248B2 (en) * | 2012-12-17 | 2017-05-09 | Sony Corporation | Methods, systems, and media for high dynamic range imaging |
| JP6259185B2 (en) * | 2012-12-21 | 2018-01-10 | キヤノン株式会社 | IMAGING DEVICE, ITS CONTROL METHOD, PROGRAM, AND STORAGE MEDIUM |
| US8902328B2 (en) * | 2013-03-14 | 2014-12-02 | Konica Minolta Laboratory U.S.A., Inc. | Method of selecting a subset from an image set for generating high dynamic range image |
| KR102145201B1 (en) * | 2013-08-12 | 2020-08-18 | 삼성전자주식회사 | Method and apparatus for dynamic range enhancement of an image |
| JP6267502B2 (en) * | 2013-12-10 | 2018-01-24 | キヤノン株式会社 | IMAGING DEVICE, IMAGING DEVICE CONTROL METHOD, AND PROGRAM |
| US9852499B2 (en) * | 2013-12-13 | 2017-12-26 | Konica Minolta Laboratory U.S.A., Inc. | Automatic selection of optimum algorithms for high dynamic range image processing based on scene classification |
| JP6119719B2 (en) | 2014-02-05 | 2017-04-26 | カシオ計算機株式会社 | Medical skin inspection apparatus, method and program for inspecting skin lesions |
| KR102149187B1 (en) * | 2014-02-21 | 2020-08-28 | 삼성전자주식회사 | Electronic device and control method of the same |
| US9258490B2 (en) * | 2014-02-28 | 2016-02-09 | Konica Minolta Laboratory U.S.A., Inc. | Smoothing of ghost maps in a ghost artifact detection method for HDR image creation |
| US9185270B2 (en) * | 2014-02-28 | 2015-11-10 | Konia Minolta Laboratory U.S.A., Inc. | Ghost artifact detection and removal in HDR image creation using graph based selection of local reference |
| US10216996B2 (en) * | 2014-09-29 | 2019-02-26 | Sony Interactive Entertainment Inc. | Schemes for retrieving and associating content items with real-world objects using augmented reality and object recognition |
| JP6570918B2 (en) * | 2015-08-19 | 2019-09-04 | オリンパス株式会社 | Imaging apparatus and imaging method |
| CN105959591A (en) | 2016-05-30 | 2016-09-21 | 广东欧珀移动通信有限公司 | Implementation method and system of partial HDR |
| TWI590192B (en) | 2016-07-27 | 2017-07-01 | 國立中興大學 | Adaptive high dynamic range image fusion algorithm |
| CN106161967B (en) * | 2016-09-13 | 2020-03-17 | 维沃移动通信有限公司 | Backlight scene panoramic shooting method and mobile terminal |
| US9955085B2 (en) * | 2016-09-22 | 2018-04-24 | Apple Inc. | Adaptive bracketing techniques |
| US9883119B1 (en) * | 2016-09-22 | 2018-01-30 | Qualcomm Incorporated | Method and system for hardware-based motion sensitive HDR image processing |
| CN107888840B (en) * | 2017-10-30 | 2019-10-15 | Oppo广东移动通信有限公司 | High dynamic range image acquisition method and device |
| CN108337447A (en) * | 2018-04-12 | 2018-07-27 | Oppo广东移动通信有限公司 | High dynamic range image exposure compensation value acquisition method, device, equipment and medium |
| US10193627B1 (en) * | 2018-05-31 | 2019-01-29 | Ford Global Technologies, Llc | Detection of visible light communication sources over a high dynamic range |
-
2018
- 2018-09-06 TW TW107131335A patent/TWI695622B/en not_active IP Right Cessation
-
2019
- 2019-03-05 US US16/292,335 patent/US10630911B2/en active Active
- 2019-03-21 CN CN201910216129.9A patent/CN110881108B/en not_active Expired - Fee Related
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20230017607A1 (en) * | 2021-07-12 | 2023-01-19 | Panasonic Intellectual Property Management Co., Ltd. | Imaging apparatus |
| US12236623B2 (en) * | 2021-07-12 | 2025-02-25 | Panasonic Intellectual Property Management Co., Ltd. | Imaging apparatus with subject region detection |
| US12542975B2 (en) * | 2022-03-07 | 2026-02-03 | Canon Kabushiki Kaisha | Image processing apparatus to generate composite image, control method, and recording medium |
| CN114363525A (en) * | 2022-03-17 | 2022-04-15 | 杭州灵西机器人智能科技有限公司 | HDR fast automatic imaging method, structured light camera, electronic device, and storage medium |
Also Published As
| Publication number | Publication date |
|---|---|
| US10630911B2 (en) | 2020-04-21 |
| TW202011728A (en) | 2020-03-16 |
| CN110881108A (en) | 2020-03-13 |
| TWI695622B (en) | 2020-06-01 |
| CN110881108B (en) | 2021-05-04 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US10630911B2 (en) | Image processing method and image processing device | |
| CN113888437B (en) | Image processing method, device, electronic device and computer readable storage medium | |
| US10250819B2 (en) | Image processing apparatus and image processing method | |
| US10074165B2 (en) | Image composition device, image composition method, and recording medium | |
| CN108337445B (en) | Photographing method, related equipment and computer storage medium | |
| EP3793188A1 (en) | Image processing method, electronic device, and computer readable storage medium | |
| US20220094838A1 (en) | Method, Electronic Device and Computer-Readable Storage Medium for Generating a High Dynamic Range Image | |
| CN110225248A (en) | Image acquisition method and apparatus, electronic device, computer-readable storage medium | |
| CN107809582A (en) | Image processing method, electronic device, and computer-readable storage medium | |
| WO2017096866A1 (en) | Method and apparatus for generating high dynamic range image | |
| CN110650288B (en) | Focus control method and apparatus, electronic device, computer-readable storage medium | |
| CN103428410B (en) | Image capture device and image synthesis method thereof | |
| US20160019681A1 (en) | Image processing method and electronic device using the same | |
| US10972676B2 (en) | Image processing method and electronic device capable of optimizing hdr image by using depth information | |
| CN110049240B (en) | Camera control method, apparatus, electronic device and computer-readable storage medium | |
| CN115278069A (en) | Image processing method and device, computer readable storage medium and terminal | |
| CN110881103B (en) | Focusing control method and device, electronic equipment and computer readable storage medium | |
| US11869224B2 (en) | Method and system for establishing light source information prediction model | |
| CN110365897B (en) | Image correction method and device, electronic equipment and computer readable storage medium | |
| CN110392211B (en) | Image processing method and apparatus, electronic device, computer-readable storage medium | |
| CN110213499A (en) | Image processing method and device, electronic equipment and computer readable storage medium | |
| TWI797528B (en) | Dual sensor imaging system and privacy protection imaging method thereof | |
| CN104113701B (en) | Image pickup method and device | |
| CN113824890B (en) | Image processing method, device, electronic device and storage medium | |
| CN110460773B (en) | Image processing method and device, electronic equipment and computer readable storage medium |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: ALTEK CORPORATION, TAIWAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:HUANG, HSU-LIEN;CHAN, KAI-LIN;REEL/FRAME:048511/0339 Effective date: 20190225 |
|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |