US11151731B2 - Apparatus and method for efficient regularized image alignment for multi-frame fusion - Google Patents
Apparatus and method for efficient regularized image alignment for multi-frame fusion Download PDFInfo
- Publication number
- US11151731B2 US11151731B2 US16/727,751 US201916727751A US11151731B2 US 11151731 B2 US11151731 B2 US 11151731B2 US 201916727751 A US201916727751 A US 201916727751A US 11151731 B2 US11151731 B2 US 11151731B2
- Authority
- US
- United States
- Prior art keywords
- motion vector
- reference image
- tiles
- image
- tile
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/90—Dynamic range modification of images or parts thereof
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/28—Determining representative reference patterns, e.g. by averaging or distorting; Generating dictionaries
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/18—Image warping, e.g. rearranging pixels individually
-
- G06T5/008—
-
- G06T5/009—
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration using two or more images, e.g. averaging or subtraction
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/90—Dynamic range modification of images or parts thereof
- G06T5/92—Dynamic range modification of images or parts thereof based on global image properties
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/90—Dynamic range modification of images or parts thereof
- G06T5/94—Dynamic range modification of images or parts thereof based on local image properties, e.g. for local contrast enhancement
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/207—Analysis of motion for motion estimation over a hierarchy of resolutions
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/215—Motion-based segmentation
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/223—Analysis of motion using block-matching
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/223—Analysis of motion using block-matching
- G06T7/238—Analysis of motion using block-matching using non-full search, e.g. three-step search
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/246—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
- G06T7/248—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments involving reference images or patches
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/30—Determination of transform parameters for the alignment of images, i.e. image registration
- G06T7/33—Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
- G06T7/337—Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods involving reference images or patches
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/741—Circuitry for compensating brightness variation in the scene by increasing the dynamic range of the image compared to the dynamic range of the electronic image sensors
-
- H04N5/2355—
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20016—Hierarchical, coarse-to-fine, multiscale or multiresolution image processing; Pyramid transform
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20021—Dividing image into blocks, subimages or windows
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20172—Image enhancement details
- G06T2207/20208—High dynamic range [HDR] image processing
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20212—Image combination
- G06T2207/20221—Image fusion; Image merging
Definitions
- This disclosure relates generally to image capturing systems. More specifically, this disclosure relates to an apparatus and method for regularized image alignment for multi-frame fusion.
- aligning each non-reference frame with a selected reference frame is a crucial step. If this step has low quality, it directly affects the following image blending step and may lead to insufficient blending level or even ghost artifacts.
- Global image registration algorithms using a global transformation matrix is a common and efficient way to achieve alignment. But using a global transformation matrix only reduces misalignment due to camera motion and sometimes even fails to find a reliable solution in the absence of matched features. This situation frequently happens in high dynamic range (HDR) applications as input frames are under-exposed or over-exposed.
- HDR high dynamic range
- An alternative is to use methods, such as optical flow, to find a dense correspondence between frames. Though these methods produce high-quality alignment, they require a significant computational cost posing a great challenge for mobile platforms.
- This disclosure provides an apparatus and method for efficient regularized image alignment for multi-frame fusion.
- a method includes: receiving a reference image and a non-reference image; dividing the reference image into a plurality of tiles; determining, using an electronic device, a motion vector map using coarse-to-fine based motion vector estimation; and generating an output frame using the motion vector map with the reference image and the non-reference image.
- an electronic device in a second embodiment, includes at least one sensor and at least one processing device.
- the at least one processing device is configured to receive a reference image and a non-reference image; divide the reference image into a plurality of tiles; determine a motion vector map using coarse-to-fine based motion vector estimation; and generate an output frame using the motion vector map with the reference image and the non-reference image.
- a non-transitory machine-readable medium contains instructions that when executed cause at least one processor of an electronic device to: receive a reference image and a non-reference image; divide the reference image into a plurality of tiles; determine a motion vector map using coarse-to-fine based motion vector estimation; and generate an output frame using the motion vector map with the reference image and the non-reference image.
- the term “or” is inclusive, meaning and/or.
- various functions described below can be implemented or supported by one or more computer programs, each of which is formed from computer readable program code and embodied in a computer readable medium.
- application and “program” refer to one or more computer programs, software components, sets of instructions, procedures, functions, objects, classes, instances, related data, or a portion thereof adapted for implementation in a suitable computer readable program code.
- computer readable program code includes any type of computer code, including source code, object code, and executable code.
- computer readable medium includes any type of medium capable of being accessed by a computer, such as read only memory (ROM), random access memory (RAM), a hard disk drive, a compact disc (CD), a digital video disc (DVD), or any other type of memory.
- ROM read only memory
- RAM random access memory
- CD compact disc
- DVD digital video disc
- a “non-transitory” computer readable medium excludes wired, wireless, optical, or other communication links that transport transitory electrical or other signals.
- a non-transitory computer readable medium includes media where data can be permanently stored and media where data can be stored and later overwritten, such as a rewritable optical disc or an erasable memory device.
- phrases such as “have,” “may have,” “include,” or “may include” a feature indicate the existence of the feature and do not exclude the existence of other features.
- the phrases “A or B,” “at least one of A and/or B,” or “one or more of A and/or B” may include all possible combinations of A and B.
- “A or B,” “at least one of A and B,” and “at least one of A or B” may indicate all of (1) including at least one A, (2) including at least one B, or (3) including at least one A and at least one B.
- first and second may modify various components regardless of importance and do not limit the components. These terms are only used to distinguish one component from another.
- a first user device and a second user device may indicate different user devices from each other, regardless of the order or importance of the devices.
- a first component may be denoted a second component and vice versa without departing from the scope of this disclosure.
- the phrase “configured (or set) to” may be interchangeably used with the phrases “suitable for,” “having the capacity to,” “designed to,” “adapted to,” “made to,” or “capable of” depending on the circumstances.
- the phrase “configured (or set) to” does not essentially mean “specifically designed in hardware to.” Rather, the phrase “configured to” may mean that a device can perform an operation together with another device or parts.
- the phrase “processor configured (or set) to perform A, B, and C” may mean a generic-purpose processor (such as a CPU or application processor) that may perform the operations by executing one or more software programs stored in a memory device or a dedicated processor (such as an embedded processor) for performing the operations.
- Examples of an “electronic device” may include at least one of a smartphone, a tablet personal computer (PC), a mobile phone, a video phone, an e-book reader, a desktop PC, a laptop computer, a netbook computer, a workstation, a personal digital assistant (PDA), a portable multimedia player (PMP), an MP3 player, a mobile medical device, a camera, or a wearable device (such as smart glasses, a head-mounted device (HMD), electronic clothes, an electronic bracelet, an electronic necklace, an electronic appcessory, an electronic tattoo, a smart mirror, or a smart watch).
- PDA personal digital assistant
- PMP portable multimedia player
- MP3 player MP3 player
- a mobile medical device such as smart glasses, a head-mounted device (HMD), electronic clothes, an electronic bracelet, an electronic necklace, an electronic appcessory, an electronic tattoo, a smart mirror, or a smart watch.
- FIG. 1 illustrates an example network configuration including an electronic device in accordance with this disclosure
- FIGS. 2A and 2B illustrate an example process for efficient regularized image alignment using a multi-frame fusion algorithm in accordance with this disclosure
- FIG. 3 illustrates an example coarse-to-fine tile-based motion vector estimation in accordance with this disclosure
- FIGS. 4A, 4B, 4C, and 4D illustrate an example outlier removal in accordance with this disclosure
- FIGS. 5A and 5B illustrate an example structure-preserving refinement in accordance with this disclosure
- FIG. 6 illustrates an example improvement to ghost artifacts for HDR applications in accordance with this disclosure
- FIG. 7 illustrates an example blending issue reduction for MBR applications in accordance with this disclosure.
- FIG. 8 illustrates an example method for efficient regularized image alignment for multi-frame fusion in accordance with this disclosure.
- FIGS. 1 through 8 discussed below, and the various embodiments of this disclosure are described with reference to the accompanying drawings. However, it should be appreciated that this disclosure is not limited to these embodiments, and all changes and/or equivalents or replacements thereto also belong to the scope of this disclosure. The same or similar reference denotations may be used to refer to the same or similar elements throughout the specification and the drawings.
- a coarse-to-fine alignment on four-level Gaussian pyramids of the input frames is first performed to find similarities between image tiles. Then an outlier rejection step followed by a quadratic structure-preserving constraint is adopted to reduce image content distortion from the previous step.
- aligning each non-reference frame with a selected reference frame is a crucial step. If this step has a low quality, it directly affects the following image blending step and may lead to insufficient blending level or even ghost artifacts.
- Global image registration algorithms that achieve alignment using a global transformation matrix are a common and efficient way, but these algorithms only reduce misalignment due to camera motion, and sometimes even fail to find a reliable solution in the absence of matched features (This situation frequently happens in HDR applications as input frames are under-exposed or over-exposed).
- An alternative is to find a dense correspondence between frames, such as optical flow. Though it produces high-quality alignment, its significant computational cost poses a great challenge for mobile platforms.
- One or more embodiments of the present disclosure based on the requirement of speed/memory, provide a simple algorithm that strikes a balance between computational cost and correspondence quality.
- a coarse-to-fine alignment on four-level Gaussian pyramids of the input frames is first performed to find correspondences between image tiles. Then an outlier rejection step followed by a quadratic structure-preserving constraint is adopted to reduce image content distortion from the previous step. Its effectiveness and efficiency has been demonstrated via large number of input frames used for HDR and MBR applications.
- One or more embodiments of the present disclosure provide an algorithm capable of aligning multiple images/frames in the presence of camera motion or small object motion without introducing significant image distortion. It is a necessary component in the pipeline of any multi-frame blending algorithms, such as high dynamic range imaging and motion blur reduction technique, both of which fuse several images captured with different exposure/ISO settings.
- FIG. 1 illustrates an example network configuration 100 including an electronic device in accordance with this disclosure.
- the embodiment of the network configuration 100 shown in FIG. 1 is for illustration only. Other embodiments of the network configuration 100 could be used without departing from the scope of this disclosure.
- an electronic device 101 is included in the network configuration 100 .
- the electronic device 101 can include at least one of a bus 110 , a processor 120 , a memory 130 , an input/output (I/O) interface 150 , a display 160 , a communication interface 170 , or a sensor 180 .
- the electronic device 101 may exclude at least one of these components or may add at least one other component.
- the bus 110 includes a circuit for connecting the components 120 - 180 with one another and for transferring communications (such as control messages and/or data) between the components.
- the processor 120 includes one or more of a central processing unit (CPU), an application processor (AP), or a communication processor (CP).
- the processor 120 is able to control on at least one of the other components of the electronic device 101 and/or perform an operation or data processing relating to communication.
- the processor 120 is a graphics processor unit (GPU).
- the processor 120 can receive image data captured by at least one camera during a capture event.
- the processor 120 can process the image data (as discussed in more detail below) to generate HDR images of dynamic scenes using graph cut-based labeling.
- the memory 130 can include a volatile and/or non-volatile memory.
- the memory 130 can store commands or data related to at least one other component of the electronic device 101 .
- the memory 130 can store software and/or a program 140 .
- the program 140 includes, for example, a kernel 141 , middleware 143 , an application programming interface (API) 145 , and/or an application program (or “application”) 147 .
- At least a portion of the kernel 141 , middleware 143 , or API 145 may be denoted an operating system (OS).
- OS operating system
- the kernel 141 can control or manage system resources (such as the bus 110 , processor 120 , or memory 130 ) used to perform operations or functions implemented in other programs (such as the middleware 143 , API 145 , or application 147 ).
- the kernel 141 provides an interface that allows the middleware 143 , the API 145 , or the application 147 to access the individual components of the electronic device 101 to control or manage the system resources.
- the application 147 includes one or more applications for image capture as discussed below. These functions can be performed by a single application or by multiple applications that each carries out one or more of these functions.
- the middleware 143 can function as a relay to allow the API 145 or the application 147 to communicate data with the kernel 141 , for instance.
- a plurality of applications 147 can be provided.
- the middleware 143 is able to control work requests received from the applications 147 , such as by allocating the priority of using the system resources of the electronic device 101 (like the bus 110 , the processor 120 , or the memory 130 ) to at least one of the plurality of applications 147 .
- the API 145 is an interface allowing the application 147 to control functions provided from the kernel 141 or the middleware 143 .
- the API 145 includes at least one interface or function (such as a command) for filing control, window control, image processing, or text control.
- the I/O interface 150 serves as an interface that can, for example, transfer commands or data input from a user or other external devices to other component(s) of the electronic device 101 .
- the I/O interface 150 can also output commands or data received from other component(s) of the electronic device 101 to the user or the other external device.
- the display 160 includes, for example, a liquid crystal display (LCD), a light emitting diode (LED) display, an organic light emitting diode (OLED) display, a quantum-dot light emitting diode (QLED) display, a microelectromechanical systems (MEMS) display, or an electronic paper display.
- the display 160 can also be a depth-aware display, such as a multi-focal display.
- the display 160 is able to display, for example, various contents (such as text, images, videos, icons, or symbols) to the user.
- the display 160 can include a touchscreen and may receive, for example, a touch, gesture, proximity, or hovering input using an electronic pen or a body portion of the user.
- the communication interface 170 is able to set up communication between the electronic device 101 and an external electronic device (such as a first electronic device 102 , a second electronic device 104 , or a server 106 ).
- the communication interface 170 can be connected with a network 162 or 164 through wireless or wired communication to communicate with the external electronic device.
- the communication interface 170 can be a wired or wireless transceiver or any other component for transmitting and receiving signals, such as images.
- the electronic device 101 further includes one or more sensors 180 that can meter a physical quantity or detect an activation state of the electronic device 101 and convert metered or detected information into an electrical signal.
- one or more sensors 180 can include one or more buttons for touch input, one or more cameras, a gesture sensor, a gyroscope or gyro sensor, an air pressure sensor, a magnetic sensor or magnetometer, an acceleration sensor or accelerometer, a grip sensor, a proximity sensor, a color sensor (such as a red green blue (RGB) sensor), a bio-physical sensor, a temperature sensor, a humidity sensor, an illumination sensor, an ultraviolet (UV) sensor, an electromyography (EMG) sensor, an electroencephalogram (EEG) sensor, an electrocardiogram (ECG) sensor, an infrared (IR) sensor, an ultrasound sensor, an iris sensor, or a fingerprint sensor.
- UV ultraviolet
- EMG electromyography
- EEG electroencephalogram
- ECG electrocardiogram
- IR infrared
- the sensor(s) 180 can also include an inertial measurement unit, which can include one or more accelerometers, gyroscopes, and other components.
- the sensor(s) 180 can further include a control circuit for controlling at least one of the sensors included here. Any of these sensor(s) 180 can be located within the electronic device 101 .
- the first external electronic device 102 or the second external electronic device 104 can be a wearable device or an electronic device-mountable wearable device (such as an HMD).
- the electronic device 101 can communicate with the electronic device 102 through the communication interface 170 .
- the electronic device 101 can be directly connected with the electronic device 102 to communicate with the electronic device 102 without involving with a separate network.
- the electronic device 101 can also be an augmented reality wearable device, such as eyeglasses, that include one or more cameras.
- the wireless communication is able to use at least one of, for example, long term evolution (LTE), long term evolution-advanced (LTE-A), 5th generation wireless system (5G), millimeter-wave or 60 GHz wireless communication, Wireless USB, code division multiple access (CDMA), wideband code division multiple access (WCDMA), universal mobile telecommunication system (UMTS), wireless broadband (WiBro), or global system for mobile communication (GSM), as a cellular communication protocol.
- the wired connection can include, for example, at least one of a universal serial bus (USB), high definition multimedia interface (HDMI), recommended standard 232 (RS-232), or plain old telephone service (POTS).
- the network 162 includes at least one communication network, such as a computer network (like a local area network (LAN) or wide area network (WAN)), Internet, or a telephone network.
- the first and second external electronic devices 102 and 104 and server 106 each can be a device of the same or a different type from the electronic device 101 .
- the server 106 includes a group of one or more servers.
- all or some of the operations executed on the electronic device 101 can be executed on another or multiple other electronic devices (such as the electronic devices 102 and 104 or server 106 ).
- the electronic device 101 when the electronic device 101 should perform some function or service automatically or at a request, the electronic device 101 , instead of executing the function or service on its own or additionally, can request another device (such as electronic devices 102 and 104 or server 106 ) to perform at least some functions associated therewith.
- the other electronic device (such as electronic devices 102 and 104 or server 106 ) is able to execute the requested functions or additional functions and transfer a result of the execution to the electronic device 101 .
- the electronic device 101 can provide a requested function or service by processing the received result as it is or additionally.
- a cloud computing, distributed computing, or client-server computing technique may be used, for example. While FIG. 1 shows that the electronic device 101 includes the communication interface 170 to communicate with the external electronic device 104 or server 106 via the network 162 , the electronic device 101 may be independently operated without a separate communication function according to some embodiments of this disclosure.
- the server 106 can optionally support the electronic device 101 by performing or supporting at least one of the operations (or functions) implemented on the electronic device 101 .
- the server 106 can include a processing module or processor that may support the processor 120 implemented in the electronic device 101 .
- FIG. 1 illustrates one example of a network configuration 100 including an electronic device 101
- the network configuration 100 could include any number of each component in any suitable arrangement.
- computing and communication systems come in a wide variety of configurations, and FIG. 1 does not limit the scope of this disclosure to any particular configuration.
- FIG. 1 illustrates one operational environment in which various features disclosed in this patent document can be used, these features could be used in any other suitable system.
- FIGS. 2A and 2B illustrate an example process for efficient regularized image alignment using a multi-frame fusion algorithm in accordance with this disclosure.
- the process 200 shown in FIG. 2A is described as being performed using the electronic device 101 shown in FIG. 1 .
- the process 200 shown in FIG. 2A could be used with any other suitable electronic device and in any suitable system.
- the process 200 includes steps to capture multiple image frames of a scene at different exposures and process the image frames to generate a fused output.
- the fused imaged is blended using an additional chroma processing to reduce ghosting and blurring in images.
- the process 200 involves the capture of multiple image frames 205 .
- the image frames 205 are captured at different exposures, such as when one image frame 205 is captured using an automatic exposure (“auto-exposure”) or other longer exposure and the image frame 205 is captured using a shorter exposure (compared to the automatic or a longer exposure).
- An auto-exposure generally refers to an exposure that is automatically, such as without intervening human action, determined by a camera or other device, typically with little or no user input.
- a user is allowed to specify an exposure mode (such as portrait, landscape, sports, or other modes), and the automatic exposure can be generated based on the selected exposure mode without any other user input.
- Each exposure setting is typically associated with different settings for the camera, such as different apertures, shutter speeds, and camera sensor sensitivities.
- a shorter-exposure image frame is typically darker, lacks image details, and has more noise compared to an auto-exposure or other longer-exposure image frame.
- the shorter-exposure image frame may include one or more regions that are under-exposed, while the auto-exposure or other longer-exposure image frame may include one or more regions that are over-exposed.
- the short-exposure frame may only have a shorter exposure time, but has a higher ISO to match the overall image brightness of the auto-exposure or long-exposure frame. Note that while often described below as involving the use of an auto-exposure image frame and at least one shorter-exposure image frame, embodiments of the present disclosure may be used with any suitable combination of image frames captured using different exposures.
- the processor 120 can control the camera of the electronic device 101 so that the image frames 205 are captured rapidly, such as in a burst mode.
- a capture request that triggers the capture of the image frames 205 represents any suitable command or input indicating a need or desire to capture an image of a scene using the electronic device 101 .
- the capture request could be initiated in response to a user's pressing of a “soft” button presented on the display 160 or the user's pressing of a “hard” button.
- two image frames 205 are captured in response to the capture request, although more than two images could be captured.
- the image frames 205 may be produced in any suitable manner, such as where each image frame is simply captured by a camera or where a multiple-frame fusion technique is used to capture multiple initial image frames and combine them into one or more of the image frames 205 .
- one image frame 205 can be used as a reference image frame, and the other image frame 205 can be used as a non-reference image frame.
- the reference image frame may represent the auto-exposure or other longer-exposure image frame, or the reference image frame may represent the shorter-exposure image frame.
- the auto-exposure or other longer-exposure image frame may be used as the reference image frame by default, since this typically allows the image frame with greater image details to be used more when generating a composite or final image of a scene.
- this is not desirable (such as due to the creation of image artifacts), in which case the shorter-exposure image frame may be selected as the reference image frame.
- the raw image frames are pre-processed in some manner to provide part of the image processing.
- the pre-processing operation 210 can perform a white balancing function to change or correct the color balance in the raw image frames.
- the pre-processing operation 210 also can perform a function to reconstruct full color image frames from incomplete color samples contained in the raw image frames using a mask (such as a CFA mask).
- the image frames 205 are provided to an image alignment operation 215 .
- the image alignment operation aligns the image frames 205 and produce aligned image frames.
- the image alignment operation 215 can modify the non-reference image frame so that particular features in the non-reference image frame align with corresponding features in the reference image frame.
- the aligned image frame represents an aligned version of the reference image frame
- the aligned image frame represents an aligned version of the non-reference image frame. Alignment may be needed to compensate for misalignment caused by the electronic device 101 moving or rotating between image capture events, which causes objects in the image frames 205 to move or rotate slightly (as is common with handheld devices).
- the image frames 205 can be aligned both geometrically and photometrically.
- the image alignment operation 215 can use global Oriented FAST and Rotated BRIEF (ORB) features and local features from a block search to align the image frames, although other implementations of the image registration operation could also be used.
- ORB Rotated BRIEF
- the reference image frame here may or may not be modified during the alignment, and the non-reference image frame could represent the only image frame that is modified during the alignment.
- the aligned image frame (the non-reference image frame) can be histogram matched as part of the pre-processing operation 210 .
- Histogram matches a histogram of the non-reference image frame to a histogram of the reference image frame, such as by applying a suitable transfer function to the aligned image frame.
- histogram matching can operate to make the brightness level generally equal for both aligned image frames. Histogram matching may involve increasing the brightness of the shorter-exposure image frame to substantially match the brightness of the auto-exposure or other longer-exposure image frame, although the converse may occur. This also results in the generation of a pre-processed aligned image frame associated with the aligned image frame. More details regarding the image alignment operation 215 are described in relation to FIG. 2B .
- the aligned image frame is then blended in a blending operation 220 .
- the blending operation 220 blends or otherwise combines the pixels from the image frames based on the label map(s) in order to produce at least one final image of a scene.
- the final image generally represents a blend of the image frames, where each pixel in the final image is extracted from either the reference image frame or the non-reference image frame (depending on the corresponding value in the label map). Additional image processing operations can occur once the proper pixels are extracted from the image frames and used to form an image. Ideally, the final image has little or no artifacts and improved image details, even in areas where at least one of the image frames 205 were over-exposed or under-exposed.
- the blended frames then are put through a post-processing operation 225 .
- the post-processing operation 225 can perform any processing to the blended image to complete the fused output image 230 .
- FIG. 2B illustrate an example local image alignment operation 215 in accordance with this disclosure. If multiple frames are being captured, the cameras move when the lens is not kept perfectly still. Between different images captured, slight movement in the camera causes the images to be misaligned. The images need to be aligned.
- Global alignment is the determination of how much the camera moves between each of the frames.
- a model would be assigned for the movement of the image based on the movement of the camera.
- the images are corrected based on the model assigned based on the determined movement of the camera from the reference frame.
- This method would only capture the camera motion and the models are imperfect. Even though only the camera is determined to move in a single frame, this is not always true for a real camera as there are always objects moving in the scene, or other secondary effects may be present such as depth-related distortion from motion. This leads to approximating camera motion. If it appears that two images are aligned, global alignment would still have features in the images that are not correctly aligned.
- Certain alignments start with a simple global alignment and then allow the alignment to adapt to local content.
- One method to accomplish this is through optical flow methods.
- An optical flow method tries to confirm or find that every location in a non-reference image is also found in the reference image. If this is performed for every pixel, a full map of how everything has moved can be generated. This method is costly for processing time and expensive, which is not practical for mobile electronics.
- the flow techniques also have complications for images if there is no real image content (such as a wall or sky) where the output would be completely noisy or just wrong.
- Certain alignments provide for gaining the benefits of optical flow quality in the alignment without the costs and adding in more regularization to ensure that the objects align appropriately in the final image.
- Small motions are also a consideration that can be corrected or situations where the camera moves; but the objects in the images move independently. Small motions tend to go with the camera and appear as camera motion. Large motion will look like real scene motion and is outside the determination of alignment.
- the image alignment operation can include a histogram matching operation 235 , a coarse-to-fine tile-based motion vector estimation operation 240 , an outlier removal operation 245 , and a structure-guided refinement operation 250 .
- the image alignment operation 215 receives input reference frame 255 and non-reference frame 260 that have been processed in the pre-processed operation 210 .
- one image frame can be used as a reference image frame 255
- the other image frame can be used as a non-reference image frame 260 .
- the reference image frame 255 may represent the auto-exposure or other longer-exposure image frame, or the reference image frame 255 may represent the shorter-exposure image frame.
- the auto-exposure or other longer-exposure image frame may be used as the reference image frame 255 by default, since this typically allows the image frame with greater image details to be used more when generating a composite or final image of a scene. As described below, there may be some instances in which this is not desirable (such as due to the creation of image artifacts), in which case the shorter-exposure image frame may be selected as the reference image frame 255 .
- the image alignment operation 215 includes a histogram matching operation 235 .
- the histogram match occurs first to bring multiple images or frames with different capture configurations to a same brightness levels. Histogram matching is necessary for the following search of motion vectors.
- the purpose of a histogram match is to take two images that may not be exposed the same way and adjust accordingly. An image that is under exposed is compared to an image that is correctly exposed, which would cause one image to be darker than the other. The darker image would need to be adjusted to the correctly exposed image to properly compare the two images. A short exposure time with high gain to achieve a same effect as a longer exposure with low gain would also need histogram matching. These pictures should look the same, but will have slightly different images, which can use the histogram matching for normalize differences between images to make the images look as similar as possible.
- the histogram match also occurs first to bring multiple image/frames with different capture configuration to the same brightness level.
- This histogram matching can be used for searching of motion vectors, which is unnecessary for sparse features such as oriented FAST and rotated BRIEF (ORB).
- a histogram of the non-reference frame 260 is compared to a histogram of the reference frame 255 .
- the histogram of the non-reference frame 260 is transformed to match the histogram of the reference frame 255 .
- the transformed histogram of the non-reference frame can be used later in the image alignment operation 215 , as well as in the image blending operation 220 or the post processing operation 225 .
- a tile-based motion vector image breaks up the image into multiple tiles and tries to find motion map like a flow. The goal would be to find for each tile the motion vector to find the content of tile in another frame. The result is a two-dimensional (2D) map of motion vectors of the reference frame.
- the tile-based motion vectors can generate evenly distributed features. For example, sparse features such as ORB are sparse and sometimes biased towards local areas. The distribution property reduces the probability of registration failure.
- the coarse-to-fine tile-based motion vector estimation (MV estimation) 240 can determine motion vectors for a patch between the reference frame 255 and the non-reference frame 260 .
- the tile-based search of motion vectors can generate evenly distributed features. Sparse features such as ORB are sparse and sometimes biased towards local areas. The distribution property reduces the probability of registration failure.
- the search of motion vectors (features) can be performed in a coarse-to-fine scheme to reduce a search radius, which in turn decreases processing time. The selection of tile size and search radius guarantees that most common cases could be covered. Sub-pixel search improves alignment accuracy. Normalized cross correlation is adopted to be robust against poor histogram matching or different noise levels.
- the output of the MV estimation 240 is a motion vector for each of the patches into which the reference frame 255 is divided.
- the set of motion vectors for the reference frame 255 and non-reference frame 260 are output for use later in the image alignment operation 215 , image blending operation 220 , or the post processing operation 225 .
- the MV estimation 240 can use a hierarchy implementation to reduce a search radius for faster results.
- the MV estimation 240 can search the motion vector on a sub-pixel lever for greater accuracy.
- the MV estimation 240 can search for a motion vector for each tile providing sufficient matched features for a more robust motion vector estimation.
- the MV estimation 240 can search using an L2 distance for images with the same exposure. For images with different exposures, the L2 distance can be replaced by normalized cross-correlation. More details regarding the MV estimation 240 are described with respect to FIG. 3 .
- the outlier removal operation 245 can receive the locally aligned non-reference frame 260 and can determine if any outliers were produced during the motion vector estimation 240 .
- the outlier removal is the concept of removing items that do not match to the global motion. For example, a person moving across the frame might show a motion vector in a substantially different direction. This motion would be labeled as an outlier.
- Outlier rejection can reject undesirable motion vectors (motion vectors on flat areas and large motion areas) in some multi-frame fusion modes.
- HDR multi-frame higher dynamic range
- motion vectors are not reliable on flat areas due to an absence of features.
- other multi-frame applications such as multi-frame noise reduction
- motion vectors on large moving objects are not constrained during the search. Both situations could lead to non-reference image distortion, which is generally not a problem as the blending operation 220 can reject inconsistent pixels between the reference image 255 and the warped non-reference image 265 .
- the image contents in the non-reference images 260 are unique and can appear in the final composite image.
- the outlier removal 245 can reject erroneous motion vector outliers to avoid potential image distortion for a more robust image. More details regarding the outlier removal operation 245 are described with respect to FIG. 4 .
- the structure-guided refinement operation 250 can preserve an image structure from issues such as warping.
- a quadratic optimization problem is set up to try to maintain structure in the image. For example, if a building with edge is in the image, the edges of the building should not to be distorted.
- the refinement operation 250 can preserve the objects, like the edges in the building, in the scene.
- the structure-guided mesh warping can be used in multi-frame fusion modes.
- the refinement operation 250 can impose constraints on flat areas and motion areas with a global transform, while deforming the remaining image area with searched features. Compared to common optical flow, the quadratic optimization equation can be solved in a closed-form and the processing time is feasible on mobile platforms.
- the refinement operation 250 can pose a constraint as a quadratic optimization problem and solve the problem using linear equations for faster results.
- the refinement operation 250 can add similarity constraints and global constraints to reduce image contents distortion for enhanced structure preservation.
- the refinement operation can output a warped non-reference frame 265 to the image blending operation 220 or the post processing operation 225 . More details regarding the structure-guided refinement operation 250 are described with respect to FIG. 5 .
- FIGS. 2A and 2B illustrate examples of a process for efficient regularized image alignment using a multi-frame fusion algorithm
- various changes may be made to FIGS. 2A and 2B .
- various operations shown in FIGS. 2A and 2B could overlap, occur in parallel, occur in a different order, or occur any number of times.
- the specific operations shown in FIGS. 2A and 2B are examples only, and other techniques could be used to perform each of the operations shown in FIGS. 2A and 2B .
- FIG. 3 illustrates an example coarse-to-fine tile-based motion vector estimation 240 in accordance with this disclosure.
- FIG. 3 illustrates how the coarse-to-fine tile based motion vector estimation 240 shown in FIG. 2B may be used to quickly and accurately produce motion vectors for each tile divided from the reference frame 255 .
- the generation of the motion vectors is described as being performed using the electronic device 101 of FIG. 1 , although any other suitable electronic device in any suitable system may be used.
- the reference frame 325 and non-reference frame 330 correspond respectively to reference frame 255 and non-reference frame 260 shown in FIG. 2B .
- the frames 325 and 330 are different resolutions of a man waving his left hand while sitting on a ledge in an office building. Behind the man is a window with a view of the city including many natural (trees, clouds, etc.) and manmade objects (other buildings, roads, etc.). The movement of the hand is being lifted from the reference frame 325 to the non-reference frame 330 . The movement of the arm causes smaller movements in the overall posture of the man.
- the motion vector maps can accurately detect the different motions for enhanced local alignment.
- the coarse-to-fine tile-based motion vector basically takes a high-resolution image and the resolution is reduced in steps 305 a - 305 d to a low-resolution. Each step is then broken into tiles 310 . Starting at the low-resolution step 305 a , the tile 310 from the reference frame 325 is found in another frame, such as the non-reference frame 330 . In other words, the reference image frame is reduced in resolution and then split into frames. Starting at the low-resolution 305 a , the movement of each tile in the reference image is then found in the non-reference image. This allows a search over a large range with low-resolution covering more content.
- the motion is found in the low-resolution first step 305 a and then the search is moved to a higher resolution second step 305 b .
- the tile 310 in the low-resolution map 305 a is used as a search area 315 in the second step 305 b . That search area 315 is divided into different frames 320 for searching for the current tile 310 in the different step 305 a - 305 d for matching.
- the search is moved up again to another higher-level resolution third step 305 c.
- the frame in which the tile 310 was found to match is used as the search area 315 in the third step 305 c .
- the electronic device 101 can divide the search area 315 in the third step 305 c into multiple frames. The frames are then searched for the tile 310 . Once the tile 310 is found in a frame of the third step 305 c , the search is moved up to the highest resolution fourth step 305 d.
- the electronic device 101 can search the area in the highest resolution fourth step 305 d based on the frame 320 of the third step 305 c .
- the tile 310 is considered to be located in the non-reference image and a motion vector can be developed to determined how far the location of the frame 320 in the final image is compared to the tile 310 in the reference image 325 .
- the movement vectors 340 can be labeled in a motion vector map 335 .
- the motion vector maps 335 for each resolution show movement vectors 340 based on color and intensity of the color.
- the color indicates the direction of the movement vector and the intensity of the color indicates the amount of movement in the specified direction. The stronger or darker colors indicate significant movement between the image frames.
- the search frame can be located at any frame of the underlying resolution step 305 b - 305 d .
- the example illustrated in FIG. 3 shows that movement for the tile 310 in the top left corner is negligible in the non-reference frame.
- the corresponding movement vector 340 in the motion vector map 335 would be very low intensity, if there is any color at all.
- the auto-exposure or other longer-exposure image frame is used as the reference image frame 325 during the coarse-to-fine tile-based motion vector estimation 240 , and any areas of the image frame 325 containing motion could be replaced using corresponding areas of the non-reference image frame 330 . It is noted that there may be various situations in which one or more saturated areas in the longer-exposure image frame are partially occluded by at least one moving object in the shorter-exposure image frame.
- the motion vector estimation 240 can perform a coarse-to-fine alignment on four Gaussian pyramids of the input frames.
- the electronic device 101 can split the reference image frame into multiple tiles 310 . On each pyramid, the electronic device 101 can search for a corresponding tile in a neighborhood of a non-reference image frame 330 for each tile in the reference image frame.
- the tile size and search radius can vary with the different levels 305 a - 305 d.
- the electronic device 101 can evaluate multiple hypotheses when up sampling the coarser level of motion vectors to avoid boundary issue. For same-exposed images, the electronic device 101 can search the movement vector 340 by minimizing an L2 norm distance. For different-exposed images, a search is performed by maximizing normalized cross correlation.
- the above searching method can only generate pixel-level alignment.
- the electronic device 101 can use a quadratic function to fit near the pixel minimum and directly compute a subpixel minimum.
- FIGS. 4A through 4D illustrate example outlier removal of a non-reference image for use in producing an HDR image of a dynamic scene in accordance with this disclosure.
- FIGS. 4A through 4D illustrate how the image outlier removal 245 shown in FIG. 2B may be used to generate an aligned non-reference image for use in producing an HDR image of a dynamic scene.
- the generation of the non-reference image 260 with outliers removed here is described as being performed using the electronic device 101 of FIG. 1 , although any other suitable electronic device in any suitable system may be used.
- the image 405 shows the non-reference image before local alignment.
- the background of the image appears as a featureless area.
- the hand appears blurry due to movement and the arm appears to be wavy.
- Motion vectors may not be reliable on “flat areas,” such as a saturated area or a featureless area.
- the “flat areas” are detected by comparing a gradient magnitude accumulation in a tile with a pre-defined threshold. “Large motion vectors” can lead to significant image distortion if they are directly used for image warping.
- the electronic device 101 can use motion vectors from the motion vector estimation 2400 to compute a global geometric transformation, which could be described by an affine matrix. Although a transformation is described by an affine matrix, any type of global geometric transformation could be applied.
- the affine matrix “records” or preserves straight lines in a two-dimensional image. If a distance between a motion vector and the global affine matrix exceeds a threshold, it is referred to as a “large motion vector.” If this threshold is too small, the fine alignment results can be close to the global registration. When the threshold is increased, the strength of the local alignment is correspondingly increased.
- image 410 local alignment has been performed, but no outlier removal has been applied. While the arm appears to be aligned with slight misalignment in the fingers and arm, the background is significantly warped. The straight edge of the street is disjointed due to flaws with local alignment procedures experienced by the lack of details in the alignment image.
- the outlier removal process corrects straight edges in the image 415 when edges are detected. The image 415 still has artifacts due to motion.
- the electronic device 101 can apply outlier removal in motion areas.
- image 420 the artifacts due to motion in the arm and fingers are corrected. Due to the size of the fingers, the motion may not be fully corrected. The accuracy of small details in the finger is a trade off with computing complexity.
- FIGS. 5A and 5B illustrate an example structure-guided refinement operation 255 in accordance with this disclosure.
- FIG. 5A illustrates a quadratic mesh of a final image generated used in the structure-guided refinement 250 shown in FIG. 2B .
- FIG. 5B illustrates an example before refinement image and after refinement image generated using the structure-preserving refinement operation 255 shown in FIG. 2B .
- the electronic device 101 can preserve image structure by imposing quadratic constraints on a mesh vertex 505 .
- the local alignment term Ep can represent the error term for feature points 515 (represented by a bilinear combination of vertexes 520 ) in the non-reference frames 510 after the features points are warped to align with the corresponding feature points in the reference image frame.
- the feature points 515 in the reference image frame 505 are centers of the tiles on the finest scale.
- the corresponding feature points are features points in the reference frames 530 shifted by the computed motion vectors.
- the similarity term Es represents the error in similarity of the triangle coordinates 525 (formed by three vertexes) after warping, meaning that the shape of the triangles should be similar before warping and after warping to keep Es low.
- the global constraint term Eg can enhance “flat area” and “large motion areas” to take the global affine transform.
- artifacts 535 still exist.
- the bump on the knuckle area of the hand of the person is not actually meant to be in the image; but is an artifact of the local alignment.
- artifacts 535 that exist on the edge of the ledge at the base of the window.
- the after-refinement image has these “bump” artifacts 535 corrected.
- FIGS. 3 through 5B illustrate various examples of image alignment operations
- various changes may be made to FIGS. 3 through 5B .
- FIGS. 3A through 5B are merely meant to illustrate examples of the types of results that could be obtained using the approaches described in this disclosure.
- images of scenes can vary widely, and the results obtained using the approaches described in this patent document can also vary widely depending on the circumstances.
- FIGS. 6 and 7 illustrate an example enhance results 600 , 700 of the image alignment operation 215 in accordance with this disclosure.
- the embodiments of results 600 and results 700 illustrated in FIGS. 6 and 7 are for illustration only FIGS. 6 and 7 do not limit the scope of this disclosure to any particular results of an image alignment operation.
- FIG. 6 illustrates an example ghost artifact correction 600 for ghost artifacts 615 in accordance with this disclosure.
- the images capture a side of a person's head and a background through a window. In the background, several buildings can be seen. Above the buildings appear to be lights reflected in the window. A ghost artifact appears around the hair of the person of the global aligned image 605 .
- the ghosting artifact can be significantly reduced, if not corrected altogether as seen in the local aligned image 610 .
- FIG. 7 illustrates an example blending issue correction 700 in accordance with this disclosure.
- the blending issues remaining in the global aligned image 175 can be seen in the trees of the global aligned blending map 715 and the globally aligned image 705 .
- the locally aligned blending map 720 has more detail in the trees that can be deciphered.
- the locally aligned image 710 has fewer blending issues compared to the globally aligned image.
- FIG. 8 illustrates an example method 800 for efficient regularized image alignment for multi-frame fusion in accordance with this disclosure.
- the method 800 shown in FIG. 8 is described as involving the performance of the process 200 shown in FIG. 2A using the electronic device 101 shown in FIG. 1 .
- the method 800 shown in FIG. 8 could be used with any other suitable electronic device and in any suitable system.
- the electronic device 101 can receive a reference image and a non-reference image.
- “Receiving” in this context can mean capture using an image sensor, receive from an external device, or load from a memory.
- the reference image and the non-reference image can be captured using a same lens at different exposures, different lenses at different exposures, different lenses at different resolutions, and the like.
- the reference image and the non-reference image capture the same subject matter at different resolutions, exposures, offsets, and the like.
- the electronic device 101 can divide the reference image into a plurality of tiles.
- the reference frame is divided into a plurality of tiles in order to search for each tile in one or more non-reference frames.
- the tiles can be the same size, which means that the image would be evenly divided horizontally and vertically.
- the electronic device 101 can determine a motion vector map using a local alignment for each tile using a Gaussian pyramid of the non-reference image.
- the electronic device 101 can split a lower resolution frame of the Gaussian pyramid corresponding to the non-reference image into a plurality of search tiles. Then for each of the plurality of tiles in the reference image, the electronic device 101 can locate a matching tile from the plurality of search tiles that corresponds to a tile of the plurality of tiles and determine a low-resolution movement vector based on a change in a location of the tile from the reference image to the matching tile in the lower resolution frame of the non-reference image. The result is a generation of a low-resolution motion map based on the low-resolution movement vectors for the plurality of tiles in the reference image. This sub-process can be performed for the lowest resolution level of the Gaussian pyramid.
- the electronic device 101 can split a search area in the non-reference image into a plurality of second search tiles, wherein the search area corresponds to the matching tile.
- the search area in the non-reference image is split into a plurality of second search tiles, where the search area corresponds to the matching tile.
- a second matching tile from the plurality of second search tiles is located that corresponds to the tile of the plurality of tiles.
- the electronic device 101 determines a motion vector based on a change in location of the tile in the reference image to the second matching tile in the non-reference image.
- the motion vector map is generated based on the movement vectors for the plurality of tiles in the reference image.
- the electronic device can determine outlier motion vectors in the motion vector map.
- a global affine matrix is computed using motion vectors in the motion vector map.
- a difference is determined by comparing each of the motion vectors to the global affine matrix.
- a large motion vector is determined when the difference is greater than a threshold and the determined large motion vector is removed from the motion vector map.
- the electronic device 101 can also determine flat areas to remove from the motion vector map.
- the non-reference image can be divided into a plurality of non-reference tiles.
- a gradient magnitude accumulation in a non-reference tile can be compared with a pre-defined threshold and a flat area can be determined based on the gradient magnitude accumulation exceeding the pre-defined threshold.
- the electronic device 101 can remove motion vectors from the motion vector map that correspond to the flat area.
- the electronic device 101 can generate an output frame using the motion vector map with the reference image and the non-reference image.
- an image blending operation uses the alignment output to generate a blending map.
- the electronic device 101 performs post-processing using the reference image, the non-reference image, the motion vector map, and the blending map to generate the output image.
- FIG. 8 illustrates one example of a method 800 for efficient regularized image alignment for multi-frame fusion
- various changes may be made to FIG. 8 .
- steps in FIG. 8 could overlap, occur in parallel, occur in a different order, or occur any number of times.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Multimedia (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Signal Processing (AREA)
- Artificial Intelligence (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Studio Devices (AREA)
Abstract
Description
E=E p+Δ1 E g+Δ2 E s (1)
Claims (20)
E=E p+λ1 E g+λ2 E s,
E=E p+λ1 E g+λ2 E s,
E=E p+λ1 E g+λ2 E s, and
Priority Applications (5)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US16/727,751 US11151731B2 (en) | 2019-08-06 | 2019-12-26 | Apparatus and method for efficient regularized image alignment for multi-frame fusion |
| KR1020200094683A KR20210018084A (en) | 2019-08-06 | 2020-07-29 | Efficient Regularized Image Alignment for Multi-Frame Fusion |
| EP20849055.7A EP3956863A4 (en) | 2019-08-06 | 2020-07-30 | APPARATUS AND METHOD FOR EFFECTIVE REGULARIZED IMAGE ALIGNMENT FOR MULTI-FRAME MERGING |
| CN202080054064.3A CN114503541B (en) | 2019-08-06 | 2020-07-30 | Apparatus and method for efficient regularized image alignment for multi-frame fusion |
| PCT/KR2020/010083 WO2021025375A1 (en) | 2019-08-06 | 2020-07-30 | Apparatus and method for efficient regularized image alignment for multi-frame fusion |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US201962883306P | 2019-08-06 | 2019-08-06 | |
| US16/727,751 US11151731B2 (en) | 2019-08-06 | 2019-12-26 | Apparatus and method for efficient regularized image alignment for multi-frame fusion |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20210042941A1 US20210042941A1 (en) | 2021-02-11 |
| US11151731B2 true US11151731B2 (en) | 2021-10-19 |
Family
ID=74498973
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US16/727,751 Active US11151731B2 (en) | 2019-08-06 | 2019-12-26 | Apparatus and method for efficient regularized image alignment for multi-frame fusion |
Country Status (2)
| Country | Link |
|---|---|
| US (1) | US11151731B2 (en) |
| KR (1) | KR20210018084A (en) |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US11373281B1 (en) * | 2021-02-23 | 2022-06-28 | Qualcomm Incorporated | Techniques for anchor frame switching |
| US12518364B2 (en) | 2023-01-26 | 2026-01-06 | Samsung Electronics Co., Ltd. | Machine learning segmentation-based tone mapping in high noise and high dynamic range environments or other environments |
Families Citing this family (11)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10863105B1 (en) * | 2017-06-27 | 2020-12-08 | Amazon Technologies, Inc. | High dynamic range imaging for event detection and inventory management |
| WO2020055196A1 (en) * | 2018-09-11 | 2020-03-19 | Samsung Electronics Co., Ltd. | Apparatus and methods for generating high dynamic range media, based on multi-stage compensation of motion |
| CN113191994B (en) * | 2021-04-26 | 2023-11-21 | 北京小米移动软件有限公司 | Image processing method, device and storage medium |
| CN115499559B (en) * | 2021-06-18 | 2025-03-04 | 伟光有限公司 | Image processing device and method, processing chip and electronic device |
| CN115696066A (en) * | 2021-07-29 | 2023-02-03 | 荣耀终端有限公司 | Image processing method and electronic device |
| CN114049288B (en) * | 2021-11-05 | 2025-09-02 | Oppo广东移动通信有限公司 | Image generation method, device, electronic device, and computer-readable storage medium |
| US12079971B2 (en) | 2022-02-07 | 2024-09-03 | Samsung Electronics Co., Ltd. | Hand motion pattern modeling and motion blur synthesizing techniques |
| US11863880B2 (en) | 2022-05-31 | 2024-01-02 | Microsoft Technology Licensing, Llc | Image frame selection for multi-frame fusion |
| US12530787B2 (en) * | 2022-12-21 | 2026-01-20 | AIDash, Inc. | Techniques for digital image registration |
| CN118279311B (en) * | 2024-06-03 | 2024-10-01 | 先临三维科技股份有限公司 | Scanning rod state monitoring method, device, storage medium and electronic equipment |
| CN119603564A (en) * | 2024-12-06 | 2025-03-11 | 维沃移动通信有限公司 | Image processing method, device, equipment and readable storage medium |
Citations (12)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN1109243A (en) * | 1993-09-09 | 1995-09-27 | 索尼公司 | Motion vector detecting apparatus and method |
| US6173087B1 (en) | 1996-11-13 | 2001-01-09 | Sarnoff Corporation | Multi-view image registration with application to mosaicing and lens distortion correction |
| WO2002002173A1 (en) * | 2000-06-30 | 2002-01-10 | Cedars Sinai Medical Center | Image guidance method coronary stent deployment |
| FR2817697A1 (en) * | 2000-12-05 | 2002-06-07 | Canon Kk | Video data processing technique creates hierarchical segmented movement field |
| WO2002071333A2 (en) * | 2001-03-08 | 2002-09-12 | Université Joseph Fourier | Quantitative analysis, visualization and movement correction in dynamic processes |
| US6809758B1 (en) | 1999-12-29 | 2004-10-26 | Eastman Kodak Company | Automated stabilization method for digital image sequences |
| US20070242900A1 (en) | 2006-04-13 | 2007-10-18 | Mei Chen | Combining multiple exposure images to increase dynamic range |
| US20090168887A1 (en) * | 2009-01-12 | 2009-07-02 | Mediatek Usa Inc. | One step sub-pixel motion esitmation |
| US20100271498A1 (en) | 2009-04-22 | 2010-10-28 | Qualcomm Incorporated | System and method to selectively combine video frame image data |
| US20160035104A1 (en) | 2013-03-18 | 2016-02-04 | Fotonation Limited | A method and apparatus for motion estimation |
| US20180189937A1 (en) * | 2017-01-04 | 2018-07-05 | Samsung Electronics Co., Ltd. | Multiframe image processing using semantic saliency |
| US20190318455A1 (en) * | 2018-04-12 | 2019-10-17 | Nvidia Corporation | Adding greater realism to a computer-generated image by smoothing jagged edges within the image in an efficient manner |
-
2019
- 2019-12-26 US US16/727,751 patent/US11151731B2/en active Active
-
2020
- 2020-07-29 KR KR1020200094683A patent/KR20210018084A/en active Pending
Patent Citations (12)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN1109243A (en) * | 1993-09-09 | 1995-09-27 | 索尼公司 | Motion vector detecting apparatus and method |
| US6173087B1 (en) | 1996-11-13 | 2001-01-09 | Sarnoff Corporation | Multi-view image registration with application to mosaicing and lens distortion correction |
| US6809758B1 (en) | 1999-12-29 | 2004-10-26 | Eastman Kodak Company | Automated stabilization method for digital image sequences |
| WO2002002173A1 (en) * | 2000-06-30 | 2002-01-10 | Cedars Sinai Medical Center | Image guidance method coronary stent deployment |
| FR2817697A1 (en) * | 2000-12-05 | 2002-06-07 | Canon Kk | Video data processing technique creates hierarchical segmented movement field |
| WO2002071333A2 (en) * | 2001-03-08 | 2002-09-12 | Université Joseph Fourier | Quantitative analysis, visualization and movement correction in dynamic processes |
| US20070242900A1 (en) | 2006-04-13 | 2007-10-18 | Mei Chen | Combining multiple exposure images to increase dynamic range |
| US20090168887A1 (en) * | 2009-01-12 | 2009-07-02 | Mediatek Usa Inc. | One step sub-pixel motion esitmation |
| US20100271498A1 (en) | 2009-04-22 | 2010-10-28 | Qualcomm Incorporated | System and method to selectively combine video frame image data |
| US20160035104A1 (en) | 2013-03-18 | 2016-02-04 | Fotonation Limited | A method and apparatus for motion estimation |
| US20180189937A1 (en) * | 2017-01-04 | 2018-07-05 | Samsung Electronics Co., Ltd. | Multiframe image processing using semantic saliency |
| US20190318455A1 (en) * | 2018-04-12 | 2019-10-17 | Nvidia Corporation | Adding greater realism to a computer-generated image by smoothing jagged edges within the image in an efficient manner |
Non-Patent Citations (2)
| Title |
|---|
| International Search Report and Written Opinion of the International Searching Authority dated Nov. 11, 2020 in connection with International Patent Application No. PCT/KR2020/010083, 11 pages. |
| Zhang, Fan et al., "Parallax-tolerant Image Stitching", Department of Computer Science Portland State University, 2014 IEEE Conference on Computer Vision and Pattern Recognitionk pp. 3262-3269 (8 Pages). |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US11373281B1 (en) * | 2021-02-23 | 2022-06-28 | Qualcomm Incorporated | Techniques for anchor frame switching |
| US12518364B2 (en) | 2023-01-26 | 2026-01-06 | Samsung Electronics Co., Ltd. | Machine learning segmentation-based tone mapping in high noise and high dynamic range environments or other environments |
Also Published As
| Publication number | Publication date |
|---|---|
| US20210042941A1 (en) | 2021-02-11 |
| KR20210018084A (en) | 2021-02-17 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US11151731B2 (en) | Apparatus and method for efficient regularized image alignment for multi-frame fusion | |
| US11107205B2 (en) | Techniques for convolutional neural network-based multi-exposure fusion of multiple image frames and for deblurring multiple image frames | |
| US10944914B1 (en) | System and method for generating multi-exposure frames from single input | |
| US11503221B2 (en) | System and method for motion warping using multi-exposure frames | |
| US11503266B2 (en) | Super-resolution depth map generation for multi-camera or other environments | |
| US11430094B2 (en) | Guided multi-exposure image fusion | |
| US11062436B2 (en) | Techniques for combining image frames captured using different exposure settings into blended images | |
| US10554890B1 (en) | Apparatus and method for generating low-light images with improved bokeh using mobile electronic device | |
| US11095829B2 (en) | Apparatus and method for high dynamic range (HDR) image creation of dynamic scenes using graph cut-based labeling | |
| US11200653B2 (en) | Local histogram matching with global regularization and motion exclusion for multi-exposure image fusion | |
| US11556784B2 (en) | Multi-task fusion neural network architecture | |
| US11532091B2 (en) | Guided backpropagation-gradient updating for image processing task using redundant information from image | |
| US11195259B2 (en) | Apparatus and method for dynamic multi-camera rectification using depth camera | |
| US11107191B2 (en) | Apparatus and method for detail enhancement in super-resolution imaging using mobile electronic device | |
| CN114503541B (en) | Apparatus and method for efficient regularized image alignment for multi-frame fusion | |
| US11418766B2 (en) | Apparatus and method for chroma processing for multi-frame fusion | |
| US12541816B2 (en) | Synthetic data generation for machine learning-based post-processing | |
| US12536615B2 (en) | Bad pixel correction in image processing applications or other applications | |
| US12347110B2 (en) | System and method for synthetic data generation using dead leaves images | |
| EP3935601B1 (en) | Local histogram matching with global regularization and motion exclusion for multi-exposure image fusion | |
| US20240185431A1 (en) | System and method for ai segmentation-based registration for multi-frame processing | |
| US20250200728A1 (en) | Machine learning-based multi-frame deblurring | |
| US20250217924A1 (en) | Robust frame registration for multi-frame image processing | |
| US12541826B2 (en) | Texture-preserving halo suppression for imaging systems | |
| US20250225667A1 (en) | Image alignment for multi-frame fusion of tetra or other image data |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| AS | Assignment |
Owner name: SAMSUNG ELECTRONICS CO., LTD., KOREA, REPUBLIC OF Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ZHEN, RUIWEN;GLOTZBACH, JOHN W.;SHEIKH, HAMID R.;REEL/FRAME:051725/0672 Effective date: 20200205 |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |