US20130120543A1 - Method, System and Computer Program Product for Adjusting a Convergence Plane of a Stereoscopic Image - Google Patents
Method, System and Computer Program Product for Adjusting a Convergence Plane of a Stereoscopic Image Download PDFInfo
- Publication number
- US20130120543A1 US20130120543A1 US13/675,797 US201213675797A US2013120543A1 US 20130120543 A1 US20130120543 A1 US 20130120543A1 US 201213675797 A US201213675797 A US 201213675797A US 2013120543 A1 US2013120543 A1 US 2013120543A1
- Authority
- US
- United States
- Prior art keywords
- stereoscopic image
- adjusting
- convergence plane
- response
- predominance
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- H04N13/0018—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/122—Improving the 3D impression of stereoscopic images by modifying image signal contents, e.g. by filtering or adding monoscopic depth cues
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/128—Adjusting depth or disparity
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/144—Processing image signals for flicker reduction
Definitions
- the disclosures herein relate in general to image processing, and in particular to a method, system and computer program product for adjusting a convergence plane of a stereoscopic image.
- a stereoscopic camera's convergence distance is either: (a) infinity (for a parallel configuration of dual imaging sensors); or (b) a fixed distance (for a toe-in configuration of dual imaging sensors).
- the human's natural convergence distance is a distance from the human's eyes to such display screen where the eyes' viewing axes naturally converge.
- the stereoscopic camera's convergence distance varies from the human's natural convergence distance, then such variation (e.g., from image-to-image or scene-to-scene) can strain the human's viewing of such image with 3D effect, thereby causing the human's eventual discomfort (e.g., headaches and/or eye muscle pain).
- Such discomfort is a shortcoming, which discourages the human's viewing of such image with 3D effect.
- First and second views of a stereoscopic image are received.
- a convergence plane of the stereoscopic image is adjusted to improve a depth resolution of at least one foreground feature within the stereoscopic image for display to a human by a display device.
- the convergence plane is adjusted to position at least most of the stereoscopic image as background features for display to the human by the display device.
- FIG. 1 is a block diagram of an information handling system of the illustrative embodiments.
- FIG. 2 is a diagram of an example orientation of dual imaging sensors of a camera of FIG. 1 .
- FIG. 3 is a diagram of viewing axes of a human's left and right eyes.
- FIG. 4 is a diagram of a parallel configuration of dual imaging sensors of the camera of FIG. 1 .
- FIG. 5 is a flowchart of various operations that a computing device of FIG. 1 automatically performs for adjusting a convergence plane of a stereoscopic image.
- FIG. 1 is a block diagram of an information handling system (e.g., one or more computers and/or other electronics devices, such as battery-powered mobile smartphones), indicated generally at 100 , of the illustrative embodiments.
- a scene e.g., including a physical object 102 and its surrounding foreground and background
- a camera 104 which: (a) captures and digitizes images of such views; and (b) outputs a video sequence of such digitized (or “digital”) images to an encoding device 106 .
- a camera 104 which: (a) captures and digitizes images of such views; and (b) outputs a video sequence of such digitized (or “digital”) images to an encoding device 106 .
- FIG. 1 is a block diagram of an information handling system (e.g., one or more computers and/or other electronics devices, such as battery-powered mobile smartphones), indicated generally at 100 , of the illustrative embodiments.
- a scene e.
- the camera 104 is a stereoscopic camera that includes dual imaging sensors, which are spaced apart from one another, namely: (a) a first imaging sensor for capturing, digitizing and outputting (to the encoding device 106 ) a first image of a view for a human's left eye; and (b) a second imaging sensor for capturing, digitizing and outputting (to the encoding device 106 ) a second image of a view for the human's right eye.
- the camera 104 is a light detection and ranging (“LIDAR”) system.
- the encoding device 106 (a) receives the video sequence from the camera 104 ; (b) encodes the video sequence into a binary logic bit stream; and (c) outputs the bit stream to a storage device 108 , which receives and stores the bit stream.
- a decoding device 110 (a) reads the bit stream from the storage device 108 ; (b) in response thereto, decodes the bit stream into the video sequence; and (c) outputs the video sequence to a computing device 112 .
- the computing device 112 receives the video sequence from the decoding device 110 (e.g., automatically, or in response to a command from a display device 114 , such as a command that a user 116 specifies via a touchscreen of the display device 114 ).
- the computing device 112 receives information from the display device 114 , such as: (a) information about the display device 114 (e.g., a type and size of a screen of the display device 114 ); and/or (b) information about the user 116 (e.g., as specified by the user 116 via a touchscreen of the display device 114 ), such as preferences of the user 116 and a viewing distance of the user 116 away from the display device 114 .
- the computing device 112 automatically performs various operations for converting the video sequence's digitized images to adjust their respective convergence planes in accordance with such information, as discussed hereinbelow in connection with FIG. 5 .
- the computing device 112 automatically converts a digitized stereoscopic image to adjust its convergence plane by horizontally shifting its left and/or right views.
- the computing device 112 (a) writes the converted video sequence for storage into the storage device 108 ; and (b) outputs the converted video sequence to the display device 114 for display to the user 116 (e.g., substantially concurrent with such conversion by the computing device 112 in real-time, or after the computing device 112 subsequently reads the converted video sequence from the storage device 108 in response to a command that the user 116 specifies via a touchscreen of the display device 114 ).
- the display device 114 (a) receives the converted video sequence from the computing device 112 (e.g., automatically, or in response to a command that the user 116 specifies via the touchscreen of the display device 114 ); and (b) in response thereto, displays the converted video sequence (e.g., stereoscopic images of the object 102 and its surrounding foreground and background), which are viewable by the user 116 (e.g., with 3D effect).
- the converted video sequence e.g., stereoscopic images of the object 102 and its surrounding foreground and background
- the display device 114 is any suitable display device that includes a screen whose optical components enable viewing by the user 116 , such as a suitable plasma display screen, liquid crystal display (“LCD”) screen, or light emitting diode (“LED”) display screen.
- the display device 114 displays a stereoscopic image with three-dimensional (“3D”) effect for viewing by the user 116 through special glasses that: (a) filter the first image against being seen by the right eye of the user 116 ; and (b) filter the second image against being seen by the left eye of the user 116 .
- the display device 114 displays the stereoscopic image with 3D effect for viewing by the user 116 without relying on special glasses.
- the encoding device 106 performs its operations in response to instructions of computer-readable programs, which are stored on a computer-readable medium 118 (e.g., hard disk drive, nonvolatile flash memory card, and/or other storage device). Also, the computer-readable medium 118 stores a database of information for operations of the encoding device 106 . Similarly, the decoding device 110 and the computing device 112 perform their operations in response to instructions of computer-readable programs, which are stored on a computer-readable medium 120 . Also, the computer-readable medium 120 stores a database of information for operations of the decoding device 110 and the computing device 112 .
- a computer-readable medium 118 e.g., hard disk drive, nonvolatile flash memory card, and/or other storage device.
- the computer-readable medium 118 stores a database of information for operations of the encoding device 106 .
- the decoding device 110 and the computing device 112 perform their operations in response to instructions of computer-readable programs, which are stored on a computer-readable medium 120
- the system 100 includes various electronic circuitry components for performing the system 100 operations, implemented in a suitable combination of software, firmware and hardware, such as one or more digital signal processors (“DSPs”), microprocessors, discrete logic devices, application specific integrated circuits (“ASICs”), and field-programmable gate arrays (“FPGAs”).
- DSPs digital signal processors
- ASICs application specific integrated circuits
- FPGAs field-programmable gate arrays
- a first electronics device includes the camera 104 , the encoding device 106 , and the computer-readable medium 118 , which are housed integrally with one another
- a second electronics device includes the decoding device 110 , the computing device 112 , the display device 114 and the computer-readable medium 120 , which are housed integrally with one another.
- the encoding device 106 outputs the bit stream directly to the decoding device 110 via a network, such as a mobile (e.g., cellular) telephone network, a landline telephone network, and/or a computer network (e.g., Ethernet, Internet or intranet); and (b) accordingly, the decoding device 110 receives and processes the bit stream directly from the encoding device 106 substantially in real-time.
- the storage device 108 either: (a) concurrently receives (in parallel with the decoding device 110 ) and stores the bit stream from the encoding device 106 ; or (b) is absent from the system 100 .
- FIG. 2 is a diagram of an example orientation of the dual imaging sensors 202 and 204 (of the camera 104 ), in which a line between the sensors 202 and 204 is substantially parallel to a line between eyes 206 and 208 of the user 116 .
- the camera 104 captures and digitizes images with a landscape aspect ratio.
- FIG. 3 is a diagram of viewing axes of the left and right eyes of the user 116 .
- a stereoscopic image is displayable by the display device 114 on a screen (which is a convergence plane where viewing axes of the left and right eyes naturally converge to intersect).
- the user 116 experiences the 3D effect by viewing the stereoscopic image on the display device 114 , so that various features (e.g., objects) appear on the screen (e.g., at a point D 1 ), behind the screen (e.g., at a point D 2 ), and/or in front of the screen (e.g., at a point D 3 ).
- various features e.g., objects
- a feature's disparity is a horizontal shift between: (a) such feature's location within the first image; and (b) such feature's corresponding location within the second image.
- the amount of the feature's disparity (e.g., horizontal shift of the feature from P 1 within the first image to P 2 within the second image) is measurable as a number of pixels, so that: (a) positive disparity is represented as a positive number; and (b) negative disparity is represented as a negative number.
- FIG. 4 is a diagram of a parallel configuration of the dual imaging sensors 202 and 204 (of the camera 104 ).
- a parallel configuration of the dual imaging sensors 202 and 204 causes crossed disparity values, which limits a range of depths that can be displayed on the screen of the display device 114 .
- the computing device 112 automatically performs various operations for converting the video sequence's digitized images to adjust their respective convergence planes, so that the computing device 112 thereby uncrosses and shifts the disparity values to fit within a range that is comfortable for the user 116 to view with 3D effect on the screen of the display device 114 (“zone of comfortable viewing”), as discussed hereinbelow in connection with FIG. 5 .
- the zone of comfortable viewing is variable. For example, referring also to FIG. 3 : (a) in some cases,
- FIG. 5 is a flowchart of the various operations that the computing device 112 automatically performs for adjusting a convergence plane of a stereoscopic image (“3D image”).
- the computing device 112 In response to the 3D image, the computing device 112 generates a depth map (“image depth map” or “disparity map”) that assigns respective disparity values to pixels of the 3D image (e.g., in response to estimated or measured depths of such pixels), so that a pixel's disparity value is information about (e.g., indicative of) such pixel's depth and vice versa.
- image depth map or “disparity map”
- the computing device 112 in response to the database of information (e.g., training information) from the computer-readable medium 120 , the computing device 112 : (a) identifies (e.g., detects and classifies) various low level features (e.g., colors, edges, textures, focus/blur, sizes, gradients, and positions) and high level features (e.g., faces, bodies, sky and foliage) within the 3D image, such as by performing a mean shift clustering operation to segment the 3D image into regions; and (b) computes disparities of such features (between the first image and its associated second image).
- various low level features e.g., colors, edges, textures, focus/blur, sizes, gradients, and positions
- high level features e.g., faces, bodies, sky and foliage
- a feature's depth is proportional to a disparity between such feature's corresponding left and right image pixels, so that such disparity is information about such depth and vice versa.
- the computing device 112 determines whether to classify the 3D image as having: (a) a predominance of foreground features (“foreground dominant” image); or (b) a predominance of background features (“background dominant” image). After classifying the 3D image in that manner, the computing device 112 adjusts the 3D image's convergence plane in accordance with a determination (by the computing device 112 ) of the viewer's preferences.
- the computing device 112 in response to classifying the 3D image as being “foreground dominant,” the computing device 112 : (a) identifies at least one dominant foreground feature (e.g., foreground object) within the 3D image; (b) shifts the 3D image to adjust its convergence plane for improving perceptual 3D resolution of such feature's depth, according to a depth resolution function (“DRF”) that models a stereoacuity function; (c) updates maps of the 3D image's disparity values accordingly; and (d) in response thereto, generates the converted 3D image having such adjusted convergence plane.
- DPF depth resolution function
- the computing device 112 In response to classifying the 3D image as being “background dominant,” the computing device 112 : (a) identifies a nearest feature (e.g., object) within the 3D image; (b) shifts the 3D image to adjust its convergence plane for making such feature appear on (or, alternatively, slightly behind) the screen of the display device 114 ; (c) updates maps of the 3D image's disparity values accordingly; and (d) in response thereto, generates the converted 3D image having such adjusted convergence plane.
- a nearest feature e.g., object
- Human depth perception is affected by a combination of monocular cues and binocular cues. Conflicts between such cues may create viewing discomfort or ambiguity in such depth perception.
- the computing device 112 By evaluating a composition of the 3D image, the computing device 112 generates the converted 3D images to reduce such conflicts and thereby helps to enhance an overall 3D viewing experience of the user 116 .
- the focal plane of the user 116 accommodates to obtain better depth resolutions.
- viewing axes of the left and right eyes naturally converge to intersect on the screen of the display device 114 .
- the focal plane of the user 116 is relatively fixed on the screen of the display device 114 (yet the convergence plane may vary), so a high depth resolution is available primarily near the focal plane (according to the modeled stereoacuity function).
- the computing device 112 In response to classifying a 3D image as being “background dominant,” the computing device 112 adjusts the 3D image's convergence plane, so that at least most of the 3D image has a relatively deep depth within the 3D image for appearing behind the screen when the 3D image is displayed by the display device 114 . By comparison, in response to classifying the 3D image as being “foreground dominant,” the computing device 112 adjusts the 3D image's convergence plane, so that at least one dominant foreground feature has a relatively shallow depth within the 3D image for appearing closer to the user 116 when the 3D image is displayed by the display device 114 .
- the computing device 112 For classifying the 3D image as being either “background dominant” or “foreground dominant,” the computing device 112 computes: (a) a skew of a distribution of the 3D image's respective disparity values (“skewness”); and (b) in some cases, relative dominant depth (“RDD”) of the 3D image's respective disparity values.
- skewness a skew of a distribution of the 3D image's respective disparity values
- RDD relative dominant depth
- the computing device 112 computes the 3D image's skewness as:
- d i is a pixel's disparity value
- d is the 3D image's mean disparity. If the 3D image's skewness >1, then the computing device 112 classifies the 3D image as being “foreground dominant.” If the 3D image's skewness ⁇ 1, then the computing device 112 classifies the 3D image as being “background dominant.”
- the computing device 112 computes the 3D image's RDD as:
- R ⁇ ⁇ D ⁇ ⁇ D ( dominant ⁇ ⁇ disparity - minimum ⁇ ⁇ disparity ) ( maximum ⁇ ⁇ disparity - minimum ⁇ ⁇ disparity ) .
- a human's vision system relies upon at least a minimum degree of variation in disparity, which is called a stereo threshold.
- a stereo threshold In one example, lower stereo thresholds were generally obtained at a zero pedestal disparity, and higher stereo thresholds were generally obtained at increasingly crossed or uncrossed pedestal disparity.
- a human's perception of the 3D image may be affected by other content-related factors, such as object contours and composition of the 3D image.
- the stereoacuity function provides the stereo threshold at different disparities.
- a human's stereoacuity is especially sensitive at the focal plane (which is relatively fixed on the screen of the display device 114 ), so the human's vision system achieves a higher depth resolution for features whose disparity is near zero.
- her stereo vision was normal, and her minimal threshold disparity was 24 arcsec at zero disparity.
- the computing device 112 computes the 3D image's DRF by approximating a negative of the stereoacuity function as 1 ⁇ s(d) in terms of pixel disparity d.
- the DRF is a zero mean Gaussian function that: (a) quantifies a human's ability to visually resolve depths of features within the 3D image; and (b) has the following form to weight the pixel disparity d values near zero.
- the computing device 112 determines a shift value i that achieves a maximum opt shift value:
- Hist( 0 ) is a histogram of the various disparity values of the unconverted 3D image without shifting
- Hist(i) is a histogram of the various disparity values of the converted 3D image that has been horizontally shifted by i pixels relative to the unconverted 3D image.
- the computing device 112 horizontally shifts: (a) the left view by i/2 pixels in a left direction (if i is positive) or a right direction (if i is negative); and (b) the right view by i/2 pixels in a right direction (if i is positive) or a left direction (if i is negative).
- the computing device 112 horizontally shifts: (a) only the left view by i pixels in a left direction (if i is positive) or a right direction (if i is negative); or (b) only the right view by i pixels in a right direction (if i is positive) or a left direction (if i is negative). In that manner, the computing device 112 improves the human's perceived depth resolution in viewing of the converted 3D image on the display device 114 .
- the computing device 112 In horizontally shifting the left view and/or the right view, additional image information (e.g., beyond edges of such views) might be unavailable to the computing device 112 .
- the computing device 112 automatically: (a) crops the shifted view(s) to include only available information, while preserving the image's original aspect ratio and the adjusted convergence plane; and (b) optionally, magnifies the left and right views to restore their original dimensions that existed before such cropping.
- a computer program product is an article of manufacture that has: (a) a computer-readable medium; and (b) a computer-readable program that is stored on such medium.
- Such program is processable by an instruction execution apparatus (e.g., system or device) for causing the apparatus to perform various operations discussed hereinabove (e.g., discussed in connection with a block diagram).
- an instruction execution apparatus e.g., system or device
- the apparatus e.g., programmable information handling system
- Such program e.g., software, firmware, and/or microcode
- an object-oriented programming language e.g., C++
- a procedural programming language e.g., C
- any suitable combination thereof e.g., C++
- the computer-readable medium is a computer-readable storage medium.
- the computer-readable medium is a computer-readable signal medium.
- a computer-readable storage medium includes any system, device and/or other non-transitory tangible apparatus (e.g., electronic, magnetic, optical, electromagnetic, infrared, semiconductor, and/or any suitable combination thereof) that is suitable for storing a program, so that such program is processable by an instruction execution apparatus for causing the apparatus to perform various operations discussed hereinabove.
- non-transitory tangible apparatus e.g., electronic, magnetic, optical, electromagnetic, infrared, semiconductor, and/or any suitable combination thereof
- Examples of a computer-readable storage medium include, but are not limited to: an electrical connection having one or more wires; a portable computer diskette; a hard disk; a random access memory (“RAM”); a read-only memory (“ROM”); an erasable programmable read-only memory (“EPROM” or flash memory); an optical fiber; a portable compact disc read-only memory (“CD-ROM”); an optical storage device; a magnetic storage device; and/or any suitable combination thereof.
- a computer-readable signal medium includes any computer-readable medium (other than a computer-readable storage medium) that is suitable for communicating (e.g., propagating or transmitting) a program, so that such program is processable by an instruction execution apparatus for causing the apparatus to perform various operations discussed hereinabove.
- a computer-readable signal medium includes a data signal having computer-readable program code embodied therein (e.g., in baseband or as part of a carrier wave), which is communicated (e.g., electronically, electromagnetically, and/or optically) via wireline, wireless, optical fiber cable, and/or any suitable combination thereof.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
Description
- This application claims priority to U.S. Provisional Patent Application Ser. No. 61/558,579, filed Nov. 11, 2011, entitled A PERCEPTUAL ALGORITHM TO DECIDE THE VERGENCE DISTANCE OF 3D STEREO IMAGES, naming Ming-Jun Chen et al. as inventors, which is hereby fully incorporated herein by reference for all purposes.
- The disclosures herein relate in general to image processing, and in particular to a method, system and computer program product for adjusting a convergence plane of a stereoscopic image.
- In one example, a stereoscopic camera's convergence distance is either: (a) infinity (for a parallel configuration of dual imaging sensors); or (b) a fixed distance (for a toe-in configuration of dual imaging sensors). By comparison, if a human views a stereoscopic image (e.g., within a video sequence of images) with three-dimensional (“3D”) effect on a stereoscopic display screen that receives such image from the stereoscopic camera, then the human's natural convergence distance is a distance from the human's eyes to such display screen where the eyes' viewing axes naturally converge. If the stereoscopic camera's convergence distance varies from the human's natural convergence distance, then such variation (e.g., from image-to-image or scene-to-scene) can strain the human's viewing of such image with 3D effect, thereby causing the human's eventual discomfort (e.g., headaches and/or eye muscle pain). Such discomfort is a shortcoming, which discourages the human's viewing of such image with 3D effect.
- First and second views of a stereoscopic image are received. In response to determining that the stereoscopic image has a predominance of foreground features, a convergence plane of the stereoscopic image is adjusted to improve a depth resolution of at least one foreground feature within the stereoscopic image for display to a human by a display device. In response to determining that the stereoscopic image has a predominance of background features, the convergence plane is adjusted to position at least most of the stereoscopic image as background features for display to the human by the display device.
-
FIG. 1 is a block diagram of an information handling system of the illustrative embodiments. -
FIG. 2 is a diagram of an example orientation of dual imaging sensors of a camera ofFIG. 1 . -
FIG. 3 is a diagram of viewing axes of a human's left and right eyes. -
FIG. 4 is a diagram of a parallel configuration of dual imaging sensors of the camera ofFIG. 1 . -
FIG. 5 is a flowchart of various operations that a computing device ofFIG. 1 automatically performs for adjusting a convergence plane of a stereoscopic image. -
FIG. 1 is a block diagram of an information handling system (e.g., one or more computers and/or other electronics devices, such as battery-powered mobile smartphones), indicated generally at 100, of the illustrative embodiments. In the example ofFIG. 1 , a scene (e.g., including aphysical object 102 and its surrounding foreground and background) is viewed by acamera 104, which: (a) captures and digitizes images of such views; and (b) outputs a video sequence of such digitized (or “digital”) images to anencoding device 106. As shown in the example ofFIG. 1 , thecamera 104 is a stereoscopic camera that includes dual imaging sensors, which are spaced apart from one another, namely: (a) a first imaging sensor for capturing, digitizing and outputting (to the encoding device 106) a first image of a view for a human's left eye; and (b) a second imaging sensor for capturing, digitizing and outputting (to the encoding device 106) a second image of a view for the human's right eye. In another embodiment, thecamera 104 is a light detection and ranging (“LIDAR”) system. - The encoding device 106: (a) receives the video sequence from the
camera 104; (b) encodes the video sequence into a binary logic bit stream; and (c) outputs the bit stream to astorage device 108, which receives and stores the bit stream. A decoding device 110: (a) reads the bit stream from thestorage device 108; (b) in response thereto, decodes the bit stream into the video sequence; and (c) outputs the video sequence to acomputing device 112. Thecomputing device 112 receives the video sequence from the decoding device 110 (e.g., automatically, or in response to a command from adisplay device 114, such as a command that auser 116 specifies via a touchscreen of the display device 114). - Also, the
computing device 112 receives information from thedisplay device 114, such as: (a) information about the display device 114 (e.g., a type and size of a screen of the display device 114); and/or (b) information about the user 116 (e.g., as specified by theuser 116 via a touchscreen of the display device 114), such as preferences of theuser 116 and a viewing distance of theuser 116 away from thedisplay device 114. In response to such information, thecomputing device 112 automatically performs various operations for converting the video sequence's digitized images to adjust their respective convergence planes in accordance with such information, as discussed hereinbelow in connection withFIG. 5 . For example, thecomputing device 112 automatically converts a digitized stereoscopic image to adjust its convergence plane by horizontally shifting its left and/or right views. - Optionally, the computing device 112: (a) writes the converted video sequence for storage into the
storage device 108; and (b) outputs the converted video sequence to thedisplay device 114 for display to the user 116 (e.g., substantially concurrent with such conversion by thecomputing device 112 in real-time, or after thecomputing device 112 subsequently reads the converted video sequence from thestorage device 108 in response to a command that theuser 116 specifies via a touchscreen of the display device 114). The display device 114: (a) receives the converted video sequence from the computing device 112 (e.g., automatically, or in response to a command that theuser 116 specifies via the touchscreen of the display device 114); and (b) in response thereto, displays the converted video sequence (e.g., stereoscopic images of theobject 102 and its surrounding foreground and background), which are viewable by the user 116 (e.g., with 3D effect). - The
display device 114 is any suitable display device that includes a screen whose optical components enable viewing by theuser 116, such as a suitable plasma display screen, liquid crystal display (“LCD”) screen, or light emitting diode (“LED”) display screen. In one example, thedisplay device 114 displays a stereoscopic image with three-dimensional (“3D”) effect for viewing by theuser 116 through special glasses that: (a) filter the first image against being seen by the right eye of theuser 116; and (b) filter the second image against being seen by the left eye of theuser 116. In another example, thedisplay device 114 displays the stereoscopic image with 3D effect for viewing by theuser 116 without relying on special glasses. - The
encoding device 106 performs its operations in response to instructions of computer-readable programs, which are stored on a computer-readable medium 118 (e.g., hard disk drive, nonvolatile flash memory card, and/or other storage device). Also, the computer-readable medium 118 stores a database of information for operations of theencoding device 106. Similarly, thedecoding device 110 and thecomputing device 112 perform their operations in response to instructions of computer-readable programs, which are stored on a computer-readable medium 120. Also, the computer-readable medium 120 stores a database of information for operations of thedecoding device 110 and thecomputing device 112. - The
system 100 includes various electronic circuitry components for performing thesystem 100 operations, implemented in a suitable combination of software, firmware and hardware, such as one or more digital signal processors (“DSPs”), microprocessors, discrete logic devices, application specific integrated circuits (“ASICs”), and field-programmable gate arrays (“FPGAs”). In one embodiment: (a) a first electronics device includes thecamera 104, theencoding device 106, and the computer-readable medium 118, which are housed integrally with one another; and (b) a second electronics device includes thedecoding device 110, thecomputing device 112, thedisplay device 114 and the computer-readable medium 120, which are housed integrally with one another. - In an alternative embodiment: (a) the
encoding device 106 outputs the bit stream directly to thedecoding device 110 via a network, such as a mobile (e.g., cellular) telephone network, a landline telephone network, and/or a computer network (e.g., Ethernet, Internet or intranet); and (b) accordingly, thedecoding device 110 receives and processes the bit stream directly from theencoding device 106 substantially in real-time. In such alternative embodiment, thestorage device 108 either: (a) concurrently receives (in parallel with the decoding device 110) and stores the bit stream from theencoding device 106; or (b) is absent from thesystem 100. -
FIG. 2 is a diagram of an example orientation of thedual imaging sensors 202 and 204 (of the camera 104), in which a line between the 202 and 204 is substantially parallel to a line betweensensors 206 and 208 of theeyes user 116. In this example, while the 202 and 204 have such orientation, thesensors camera 104 captures and digitizes images with a landscape aspect ratio. -
FIG. 3 is a diagram of viewing axes of the left and right eyes of theuser 116. In the example ofFIG. 3 , a stereoscopic image is displayable by thedisplay device 114 on a screen (which is a convergence plane where viewing axes of the left and right eyes naturally converge to intersect). Theuser 116 experiences the 3D effect by viewing the stereoscopic image on thedisplay device 114, so that various features (e.g., objects) appear on the screen (e.g., at a point D1), behind the screen (e.g., at a point D2), and/or in front of the screen (e.g., at a point D3). - Within the stereoscopic image, a feature's disparity is a horizontal shift between: (a) such feature's location within the first image; and (b) such feature's corresponding location within the second image. A limit of such disparity is dependent on the
camera 104. For example, if a feature (within the stereoscopic image) is centered at the point D1 within the first image, and likewise centered at the point D1 within the second image, then: (a) such feature's disparity=D1−D1=0; and (b) theuser 116 will perceive the feature to appear at the point D1 on the screen, which is a natural convergence distance away from the left and right eyes. - By comparison, if the feature is centered at a point P1 within the first image, and centered at a point P2 within the second image, then: (a) such feature's disparity=P2−P1 will be positive; and (b) the
user 116 will perceive the feature to appear at the point D2 behind the screen, which is greater than the natural convergence distance away from the left and right eyes. Conversely, if the feature is centered at the point P2 within the first image, and centered at the point P1 within the second image, then: (a) such feature's disparity=P1−P2 will be negative; and (b) theuser 116 will perceive the feature to appear at the point D3 in front of the screen, which is less than the natural convergence distance away from the left and right eyes. The amount of the feature's disparity (e.g., horizontal shift of the feature from P1 within the first image to P2 within the second image) is measurable as a number of pixels, so that: (a) positive disparity is represented as a positive number; and (b) negative disparity is represented as a negative number. -
FIG. 4 is a diagram of a parallel configuration of thedual imaging sensors 202 and 204 (of the camera 104). As shown inFIG. 4 , a parallel configuration of the 202 and 204 causes crossed disparity values, which limits a range of depths that can be displayed on the screen of thedual imaging sensors display device 114. Accordingly, thecomputing device 112 automatically performs various operations for converting the video sequence's digitized images to adjust their respective convergence planes, so that thecomputing device 112 thereby uncrosses and shifts the disparity values to fit within a range that is comfortable for theuser 116 to view with 3D effect on the screen of the display device 114 (“zone of comfortable viewing”), as discussed hereinbelow in connection withFIG. 5 . The zone of comfortable viewing is variable. For example, referring also toFIG. 3 : (a) in some cases, |α−β|≦1° within the zone of comfortable viewing; and (b) in other cases, |α−β|≦0.5° within the zone of comfortable viewing. -
FIG. 5 is a flowchart of the various operations that thecomputing device 112 automatically performs for adjusting a convergence plane of a stereoscopic image (“3D image”). In response to the 3D image, thecomputing device 112 generates a depth map (“image depth map” or “disparity map”) that assigns respective disparity values to pixels of the 3D image (e.g., in response to estimated or measured depths of such pixels), so that a pixel's disparity value is information about (e.g., indicative of) such pixel's depth and vice versa. Optionally, in response to the database of information (e.g., training information) from the computer-readable medium 120, the computing device 112: (a) identifies (e.g., detects and classifies) various low level features (e.g., colors, edges, textures, focus/blur, sizes, gradients, and positions) and high level features (e.g., faces, bodies, sky and foliage) within the 3D image, such as by performing a mean shift clustering operation to segment the 3D image into regions; and (b) computes disparities of such features (between the first image and its associated second image). For example, in a stereo pair of images (e.g., a stereoscopic image), a feature's depth is proportional to a disparity between such feature's corresponding left and right image pixels, so that such disparity is information about such depth and vice versa. - In response to the 3D image, the
computing device 112 determines whether to classify the 3D image as having: (a) a predominance of foreground features (“foreground dominant” image); or (b) a predominance of background features (“background dominant” image). After classifying the 3D image in that manner, thecomputing device 112 adjusts the 3D image's convergence plane in accordance with a determination (by the computing device 112) of the viewer's preferences. - For example, in response to classifying the 3D image as being “foreground dominant,” the computing device 112: (a) identifies at least one dominant foreground feature (e.g., foreground object) within the 3D image; (b) shifts the 3D image to adjust its convergence plane for improving perceptual 3D resolution of such feature's depth, according to a depth resolution function (“DRF”) that models a stereoacuity function; (c) updates maps of the 3D image's disparity values accordingly; and (d) in response thereto, generates the converted 3D image having such adjusted convergence plane. In response to classifying the 3D image as being “background dominant,” the computing device 112: (a) identifies a nearest feature (e.g., object) within the 3D image; (b) shifts the 3D image to adjust its convergence plane for making such feature appear on (or, alternatively, slightly behind) the screen of the
display device 114; (c) updates maps of the 3D image's disparity values accordingly; and (d) in response thereto, generates the converted 3D image having such adjusted convergence plane. - Human depth perception is affected by a combination of monocular cues and binocular cues. Conflicts between such cues may create viewing discomfort or ambiguity in such depth perception. By evaluating a composition of the 3D image, the
computing device 112 generates the converted 3D images to reduce such conflicts and thereby helps to enhance an overall 3D viewing experience of theuser 116. - While scanning a real-world 3D scene, the focal plane of the
user 116 accommodates to obtain better depth resolutions. By comparison, as discussed hereinabove in connection withFIG. 3 , viewing axes of the left and right eyes naturally converge to intersect on the screen of thedisplay device 114. Accordingly, while viewing 3D images on the screen of thedisplay device 114, the focal plane of theuser 116 is relatively fixed on the screen of the display device 114 (yet the convergence plane may vary), so a high depth resolution is available primarily near the focal plane (according to the modeled stereoacuity function). - In response to classifying a 3D image as being “background dominant,” the
computing device 112 adjusts the 3D image's convergence plane, so that at least most of the 3D image has a relatively deep depth within the 3D image for appearing behind the screen when the 3D image is displayed by thedisplay device 114. By comparison, in response to classifying the 3D image as being “foreground dominant,” thecomputing device 112 adjusts the 3D image's convergence plane, so that at least one dominant foreground feature has a relatively shallow depth within the 3D image for appearing closer to theuser 116 when the 3D image is displayed by thedisplay device 114. For classifying the 3D image as being either “background dominant” or “foreground dominant,” thecomputing device 112 computes: (a) a skew of a distribution of the 3D image's respective disparity values (“skewness”); and (b) in some cases, relative dominant depth (“RDD”) of the 3D image's respective disparity values. - The
computing device 112 computes the 3D image's skewness as: -
- where di is a pixel's disparity value, and
d is the 3D image's mean disparity. If the 3D image's skewness >1, then thecomputing device 112 classifies the 3D image as being “foreground dominant.” If the 3D image's skewness <−1, then thecomputing device 112 classifies the 3D image as being “background dominant.” - If the 3D image's |skewness|<1, then the
computing device 112 computes the 3D image's RDD as: -
- The dominant disparity is a mode of a given disparity set. In one example: (a) ξ=0.25; (b) if |RDD|<ξ, then the
computing device 112 classifies the 3D image as being “foreground dominant”; and (c) if |RDD|>ξ, then thecomputing device 112 classifies the 3D image as being “background dominant.” - For perceiving different depths between features, a human's vision system relies upon at least a minimum degree of variation in disparity, which is called a stereo threshold. In one example, lower stereo thresholds were generally obtained at a zero pedestal disparity, and higher stereo thresholds were generally obtained at increasingly crossed or uncrossed pedestal disparity. Also, a human's perception of the 3D image may be affected by other content-related factors, such as object contours and composition of the 3D image.
- The stereoacuity function provides the stereo threshold at different disparities. In viewing a 3D image on the screen of the
display device 114, a human's stereoacuity is especially sensitive at the focal plane (which is relatively fixed on the screen of the display device 114), so the human's vision system achieves a higher depth resolution for features whose disparity is near zero. In one study of a female's stereoacuity, her stereo vision was normal, and her minimal threshold disparity was 24 arcsec at zero disparity. - For example, if two objects are 50 arcsec apart in depth, and if one of those objects has a zero disparity value, then such female would be expected to see the two objects at different depth planes from one another. By comparison, if one of those objects has a disparity value larger than 20 arcmin, then such female would be expected to see the two objects at approximately the same depth plane as one another. Accordingly, if the
computing device 112 adjusts the 3D image's convergence plane to make both of those objects have disparities near zero (so that such objects appear on or near the screen of the display device 114), then such female would be expected to see the two objects with improved depth resolution. - In response to classifying the 3D image as being “foreground dominant,” the
computing device 112 computes the 3D image's DRF by approximating a negative of the stereoacuity function as 1−s(d) in terms of pixel disparity d. The DRF is a zero mean Gaussian function that: (a) quantifies a human's ability to visually resolve depths of features within the 3D image; and (b) has the following form to weight the pixel disparity d values near zero. -
- In accordance with the following equation, the
computing device 112 determines a shift value i that achieves a maximum opt shift value: -
- where the
computing device 112 computes DRF in response to μ=0 and σ=20 arcmin (which is preselected as a parameter for improving fit to the modeled stereoacuity function, based on a viewing environment, such as screen size, screen resolution, and viewing distance). Accordingly, Hist(0) is a histogram of the various disparity values of the unconverted 3D image without shifting, and Hist(i) is a histogram of the various disparity values of the converted 3D image that has been horizontally shifted by i pixels relative to the unconverted 3D image. - In a first example, the
computing device 112 horizontally shifts: (a) the left view by i/2 pixels in a left direction (if i is positive) or a right direction (if i is negative); and (b) the right view by i/2 pixels in a right direction (if i is positive) or a left direction (if i is negative). In a second example, thecomputing device 112 horizontally shifts: (a) only the left view by i pixels in a left direction (if i is positive) or a right direction (if i is negative); or (b) only the right view by i pixels in a right direction (if i is positive) or a left direction (if i is negative). In that manner, thecomputing device 112 improves the human's perceived depth resolution in viewing of the converted 3D image on thedisplay device 114. - In horizontally shifting the left view and/or the right view, additional image information (e.g., beyond edges of such views) might be unavailable to the
computing device 112. In response to such unavailability, thecomputing device 112 automatically: (a) crops the shifted view(s) to include only available information, while preserving the image's original aspect ratio and the adjusted convergence plane; and (b) optionally, magnifies the left and right views to restore their original dimensions that existed before such cropping. - In the illustrative embodiments, a computer program product is an article of manufacture that has: (a) a computer-readable medium; and (b) a computer-readable program that is stored on such medium. Such program is processable by an instruction execution apparatus (e.g., system or device) for causing the apparatus to perform various operations discussed hereinabove (e.g., discussed in connection with a block diagram). For example, in response to processing (e.g., executing) such program's instructions, the apparatus (e.g., programmable information handling system) performs various operations discussed hereinabove. Accordingly, such operations are computer-implemented.
- Such program (e.g., software, firmware, and/or microcode) is written in one or more programming languages, such as: an object-oriented programming language (e.g., C++); a procedural programming language (e.g., C); and/or any suitable combination thereof. In a first example, the computer-readable medium is a computer-readable storage medium. In a second example, the computer-readable medium is a computer-readable signal medium.
- A computer-readable storage medium includes any system, device and/or other non-transitory tangible apparatus (e.g., electronic, magnetic, optical, electromagnetic, infrared, semiconductor, and/or any suitable combination thereof) that is suitable for storing a program, so that such program is processable by an instruction execution apparatus for causing the apparatus to perform various operations discussed hereinabove. Examples of a computer-readable storage medium include, but are not limited to: an electrical connection having one or more wires; a portable computer diskette; a hard disk; a random access memory (“RAM”); a read-only memory (“ROM”); an erasable programmable read-only memory (“EPROM” or flash memory); an optical fiber; a portable compact disc read-only memory (“CD-ROM”); an optical storage device; a magnetic storage device; and/or any suitable combination thereof.
- A computer-readable signal medium includes any computer-readable medium (other than a computer-readable storage medium) that is suitable for communicating (e.g., propagating or transmitting) a program, so that such program is processable by an instruction execution apparatus for causing the apparatus to perform various operations discussed hereinabove. In one example, a computer-readable signal medium includes a data signal having computer-readable program code embodied therein (e.g., in baseband or as part of a carrier wave), which is communicated (e.g., electronically, electromagnetically, and/or optically) via wireline, wireless, optical fiber cable, and/or any suitable combination thereof.
- Although illustrative embodiments have been shown and described by way of example, a wide range of alternative embodiments is possible within the scope of the foregoing disclosure.
Claims (30)
Priority Applications (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US13/675,797 US9363498B2 (en) | 2011-11-11 | 2012-11-13 | Method, system and computer program product for adjusting a convergence plane of a stereoscopic image |
| US15/155,147 US10447985B2 (en) | 2011-11-11 | 2016-05-16 | Method, system and computer program product for adjusting a convergence plane of a stereoscopic image |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US201161558579P | 2011-11-11 | 2011-11-11 | |
| US13/675,797 US9363498B2 (en) | 2011-11-11 | 2012-11-13 | Method, system and computer program product for adjusting a convergence plane of a stereoscopic image |
Related Child Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US15/155,147 Continuation US10447985B2 (en) | 2011-11-11 | 2016-05-16 | Method, system and computer program product for adjusting a convergence plane of a stereoscopic image |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20130120543A1 true US20130120543A1 (en) | 2013-05-16 |
| US9363498B2 US9363498B2 (en) | 2016-06-07 |
Family
ID=48280250
Family Applications (2)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US13/675,797 Active 2034-02-26 US9363498B2 (en) | 2011-11-11 | 2012-11-13 | Method, system and computer program product for adjusting a convergence plane of a stereoscopic image |
| US15/155,147 Active 2033-05-25 US10447985B2 (en) | 2011-11-11 | 2016-05-16 | Method, system and computer program product for adjusting a convergence plane of a stereoscopic image |
Family Applications After (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US15/155,147 Active 2033-05-25 US10447985B2 (en) | 2011-11-11 | 2016-05-16 | Method, system and computer program product for adjusting a convergence plane of a stereoscopic image |
Country Status (1)
| Country | Link |
|---|---|
| US (2) | US9363498B2 (en) |
Cited By (9)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20130250065A1 (en) * | 2012-03-21 | 2013-09-26 | Ricoh Company, Ltd. | Range-finding system and vehicle mounting the range-finding system |
| CN104238127A (en) * | 2014-09-12 | 2014-12-24 | 京东方科技集团股份有限公司 | Naked-eye three-dimensional display device |
| CN104238126A (en) * | 2014-09-12 | 2014-12-24 | 京东方科技集团股份有限公司 | Naked-eye three-dimensional display device |
| US20150015677A1 (en) * | 2012-03-30 | 2015-01-15 | Fujifilm Corporation | Image processing device and method, and imaging device |
| US20150054926A1 (en) * | 2012-05-09 | 2015-02-26 | Fujifilm Corporation | Image processing device and method, and image capturing device |
| US9224060B1 (en) * | 2013-09-17 | 2015-12-29 | Amazon Technologies, Inc. | Object tracking using depth information |
| US10250864B2 (en) * | 2013-10-30 | 2019-04-02 | Vefxi Corporation | Method and apparatus for generating enhanced 3D-effects for real-time and offline applications |
| CN110766152A (en) * | 2018-07-27 | 2020-02-07 | 富士通株式会社 | Method and apparatus for training deep neural networks |
| CN111862511A (en) * | 2020-08-10 | 2020-10-30 | 湖南海森格诺信息技术有限公司 | Target intrusion detection device and method based on binocular stereo vision |
Citations (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6369831B1 (en) * | 1998-01-22 | 2002-04-09 | Sony Corporation | Picture data generating method and apparatus |
| US6668080B1 (en) * | 1999-09-21 | 2003-12-23 | Microsoft Corporation | Automated layer extraction and pixel assignment from image sequences |
| US20110292045A1 (en) * | 2009-02-05 | 2011-12-01 | Fujifilm Corporation | Three-dimensional image output device and three-dimensional image output method |
Family Cites Families (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US9049423B2 (en) * | 2010-12-01 | 2015-06-02 | Qualcomm Incorporated | Zero disparity plane for feedback-based three-dimensional video |
-
2012
- 2012-11-13 US US13/675,797 patent/US9363498B2/en active Active
-
2016
- 2016-05-16 US US15/155,147 patent/US10447985B2/en active Active
Patent Citations (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6369831B1 (en) * | 1998-01-22 | 2002-04-09 | Sony Corporation | Picture data generating method and apparatus |
| US6668080B1 (en) * | 1999-09-21 | 2003-12-23 | Microsoft Corporation | Automated layer extraction and pixel assignment from image sequences |
| US20110292045A1 (en) * | 2009-02-05 | 2011-12-01 | Fujifilm Corporation | Three-dimensional image output device and three-dimensional image output method |
Cited By (15)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20130250065A1 (en) * | 2012-03-21 | 2013-09-26 | Ricoh Company, Ltd. | Range-finding system and vehicle mounting the range-finding system |
| US20160227188A1 (en) * | 2012-03-21 | 2016-08-04 | Ricoh Company, Ltd. | Calibrating range-finding system using parallax from two different viewpoints and vehicle mounting the range-finding system |
| US9338438B2 (en) * | 2012-03-21 | 2016-05-10 | Ricoh Company, Ltd. | Calibrating range-finding system using parallax from two different viewpoints and vehicle mounting the range-finding system |
| US20150015677A1 (en) * | 2012-03-30 | 2015-01-15 | Fujifilm Corporation | Image processing device and method, and imaging device |
| US9277201B2 (en) * | 2012-03-30 | 2016-03-01 | Fujifilm Corporation | Image processing device and method, and imaging device |
| US20150054926A1 (en) * | 2012-05-09 | 2015-02-26 | Fujifilm Corporation | Image processing device and method, and image capturing device |
| US9288472B2 (en) * | 2012-05-09 | 2016-03-15 | Fujifilm Corporation | Image processing device and method, and image capturing device |
| US9224060B1 (en) * | 2013-09-17 | 2015-12-29 | Amazon Technologies, Inc. | Object tracking using depth information |
| US10250864B2 (en) * | 2013-10-30 | 2019-04-02 | Vefxi Corporation | Method and apparatus for generating enhanced 3D-effects for real-time and offline applications |
| CN104238126A (en) * | 2014-09-12 | 2014-12-24 | 京东方科技集团股份有限公司 | Naked-eye three-dimensional display device |
| CN104238127A (en) * | 2014-09-12 | 2014-12-24 | 京东方科技集团股份有限公司 | Naked-eye three-dimensional display device |
| US9778472B2 (en) | 2014-09-12 | 2017-10-03 | Boe Technology Group Co., Ltd. | Stereoscopic display device |
| US10324303B2 (en) | 2014-09-12 | 2019-06-18 | Boe Technology Group Co., Ltd. | Stereoscopic display device |
| CN110766152A (en) * | 2018-07-27 | 2020-02-07 | 富士通株式会社 | Method and apparatus for training deep neural networks |
| CN111862511A (en) * | 2020-08-10 | 2020-10-30 | 湖南海森格诺信息技术有限公司 | Target intrusion detection device and method based on binocular stereo vision |
Also Published As
| Publication number | Publication date |
|---|---|
| US20160261843A1 (en) | 2016-09-08 |
| US9363498B2 (en) | 2016-06-07 |
| US10447985B2 (en) | 2019-10-15 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US10447985B2 (en) | Method, system and computer program product for adjusting a convergence plane of a stereoscopic image | |
| US12143618B2 (en) | Method, system and computer program product for encoding disparities between views of a stereoscopic image | |
| US11244496B2 (en) | Information processing device and information processing method | |
| CN105930821B (en) | A human eye recognition and tracking method and a device for applying it to naked-eye 3D display | |
| US9355436B2 (en) | Method, system and computer program product for enhancing a depth map | |
| US9451242B2 (en) | Apparatus for adjusting displayed picture, display apparatus and display method | |
| US9277207B2 (en) | Image processing apparatus, image processing method, and program for generating multi-view point image | |
| CN114910052B (en) | Distance measurement method, control method and device based on camera and electronic equipment | |
| US8805059B2 (en) | Method, system and computer program product for segmenting an image | |
| US9838673B2 (en) | Method and apparatus for adjusting viewing area, and device capable of three-dimension displaying video signal | |
| US20180184066A1 (en) | Light field retargeting for multi-panel display | |
| CN101180653A (en) | Method and device for three-dimensional rendering | |
| US20150304625A1 (en) | Image processing device, method, and recording medium | |
| US20100246938A1 (en) | Image Processing Method for Providing Depth Information and Image Processing System Using the Same | |
| US8768073B2 (en) | Method, system and computer program product for coding a region of interest within an image of multiple views | |
| US11726320B2 (en) | Information processing apparatus, information processing method, and program | |
| CN106919246A (en) | The display methods and device of a kind of application interface | |
| US12254131B1 (en) | Gaze-adaptive image reprojection | |
| US20230222727A1 (en) | Method and device for generating three-dimensional image by using plurality of cameras | |
| US8879826B2 (en) | Method, system and computer program product for switching between 2D and 3D coding of a video sequence of images | |
| US20130033490A1 (en) | Method, System and Computer Program Product for Reorienting a Stereoscopic Image | |
| US20130009949A1 (en) | Method, system and computer program product for re-convergence of a stereoscopic image | |
| KR101690256B1 (en) | Method and apparatus for processing image | |
| US20240320853A1 (en) | Eye tracking system and methods of using an eye tracking system | |
| US12081722B2 (en) | Stereo image generation method and electronic apparatus using the same |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: TEXAS INSTRUMENTS INCORPORATED, TEXAS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CHEN, MING-JUN;KWON, DO-KYOUNG;SIGNING DATES FROM 20140609 TO 20140611;REEL/FRAME:034829/0201 |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |