[go: up one dir, main page]

WO2018131803A1 - Procédé et appareil permettant de transmettre un contenu vidéo stéréoscopique - Google Patents

Procédé et appareil permettant de transmettre un contenu vidéo stéréoscopique Download PDF

Info

Publication number
WO2018131803A1
WO2018131803A1 PCT/KR2017/014742 KR2017014742W WO2018131803A1 WO 2018131803 A1 WO2018131803 A1 WO 2018131803A1 KR 2017014742 W KR2017014742 W KR 2017014742W WO 2018131803 A1 WO2018131803 A1 WO 2018131803A1
Authority
WO
WIPO (PCT)
Prior art keywords
packing
regions
region
information
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/KR2017/014742
Other languages
English (en)
Korean (ko)
Inventor
최병두
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Electronics Co Ltd
Original Assignee
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from KR1020170171492A external-priority patent/KR102503342B1/ko
Application filed by Samsung Electronics Co Ltd filed Critical Samsung Electronics Co Ltd
Priority to CN201780082778.3A priority Critical patent/CN110463196B/zh
Priority to US16/477,102 priority patent/US10855968B2/en
Priority to EP17891098.0A priority patent/EP3570540A4/fr
Publication of WO2018131803A1 publication Critical patent/WO2018131803A1/fr
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/816Monomedia components thereof involving special video data, e.g 3D video
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/161Encoding, multiplexing or demultiplexing different image signal components
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/194Transmission of image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/243Image signal generators using stereoscopic image cameras using three or more 2D image sensors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/21805Source of audio or video content, e.g. local disk arrays enabling multiple viewpoints, e.g. using a plurality of cameras

Definitions

  • the present disclosure relates to a method and apparatus for packing data of stereoscopic omni-directional video.
  • the Internet has evolved from a human-centered connection network where humans create and consume information, and an Internet of Things (IoT) network that exchanges and processes information among distributed components such as things.
  • IoT Internet of Things
  • IoE Internet of Everything
  • IoT Internet Technology
  • IoT Internet Technology
  • contents for implementing IoT are also evolving.
  • HD high definition
  • UHD ultrahigh definition television
  • HD ultrahigh definition television
  • HD high dynamic range
  • VR virtual reality
  • the fundamental foundation of a VR system is to monitor the user so that the user can use any kind of controller to provide feedback input to the content display device or processing unit, and that device or unit processes that input and adjusts the content accordingly. This is a system that enables interaction.
  • Basic configurations within the VR ecosystem include, for example, head mounted display (HMD), wireless, mobile VR, TVs, CA automatic virtual environments (CA VE), peripherals and other controllers for providing input to haptics (VR).
  • HMD head mounted display
  • CA VE CA automatic virtual environments
  • VR peripherals and other controllers for providing input to haptics
  • next-generation high efficiency video coding (HEVC) codec which can be specifically designed for 3D, 360-degree content for capturing, encoding and transmitting 360-degree video content, which is performed to construct VR content. I'm facing a challenge.
  • HEVC next-generation high efficiency video coding
  • the present disclosure proposes a method and apparatus for packing data of stereo omni-directional video.
  • the present disclosure also proposes a trapezoid-based region-wise packing method.
  • the present disclosure proposes a packing method of an omnidirectional fisheye image.
  • a method of packing stereoscopic video content according to an aspect of the present disclosure based on stereoscopic image data including a plurality of monoscopic images having a parallax, to the plurality of monoscopic images.
  • the method for transmitting stereoscopic video content based on the data of the stereoscopic image comprising a plurality of omnidirectional images having a parallax, from the plurality of omnidirectional images Generating a first frame comprising a plurality of projected first views; Generating a second frame including a plurality of second views by packing a plurality of first regions included in the plurality of first views based on region-wise packing information; And transmitting data relating to the generated second frame, wherein the plurality of second views includes a plurality of second regions corresponding to the plurality of first regions, and the packing information for each region may include: It includes information about the shape, orientation or transformation of each of the plurality of second regions.
  • An apparatus for transmitting stereoscopic video content comprising: a memory; Transceiver; And at least one processor coupled to the memory and the transceiver, wherein the at least one processor is based on data of the stereoscopic image including a plurality of omnidirectional images having parallax; Generating a first frame including a plurality of first views projected from a plurality of omnidirectional images, and based on region-wise packing information, a plurality of first images included in the plurality of first views Packing the first regions to generate a second frame including a plurality of second views, and transmitting data about the generated second frame, wherein the plurality of second views comprise the plurality of first views; And a plurality of second regions corresponding to regions, wherein the region-specific packing information includes information about a shape, orientation, or transformation of each of the plurality of second regions.
  • FIG. 1 is an exemplary view for explaining the configuration of a computer system that implements a stereo omnidirectional image packing method according to the present invention.
  • FIG. 2 illustrates a left and right stereoscopic 360 format according to the present disclosure
  • FIG. 3 illustrates a top-bottom stereoscopic 360 format.
  • FIG. 4 illustrates image stitching, projection, and packing per region of a single acquisition time instance.
  • FIG. 5 is an exemplary view for explaining a non-area packing method according to the present disclosure.
  • FIG. 6 is an exemplary diagram for explaining a separate and independent packing method according to the present disclosure.
  • FIG. 7 is an exemplary view for explaining a separation and mirroring packing method according to the present disclosure.
  • FIG. 8 is an exemplary diagram for explaining a mixed and independent packing method according to the present disclosure.
  • FIG 9 is an exemplary view for explaining a mixed and pair-wise packing method according to the present disclosure.
  • FIG. 10 is an exemplary view for explaining a packing method for a regular polyhedral projection image according to the present disclosure.
  • FIG. 11 is an exemplary view for explaining a packing method for each region using a triangular patch according to the present disclosure.
  • FIG. 12 is an exemplary view for explaining the layout of the left and right regions used in the non-region-specific packing method according to the present disclosure.
  • FIG. 13 is an exemplary view for explaining the layout of the upper and lower regions used in the non-region-specific packing method according to the present disclosure.
  • FIG. 15 is an exemplary diagram for explaining a region-specific packing method of adjusting and rearranging an area according to latitude in an isotropic projection (ERP) according to the present disclosure.
  • FIG. 16 is an exemplary diagram for explaining region-specific packing for a cube projection for viewport dependent streaming according to the present disclosure.
  • 17 is an exemplary diagram for explaining an embodiment of a method of packing an ERP image according to the present disclosure.
  • FIG. 18 is an exemplary diagram for describing a method of packing an ERP image according to the present disclosure.
  • 19 is an exemplary diagram for explaining a method of converting an isotonic projection according to the present disclosure into a layout similar to a cube.
  • 20 is an exemplary diagram for explaining another embodiment of converting an isotonic projection according to the present disclosure into a layout similar to a cube.
  • 21 is an exemplary diagram for describing a method of converting an ERP image into a cube-like ERP according to the present disclosure.
  • FIG. 22 is an exemplary view for explaining a TSP packing method according to the present disclosure.
  • FIG. 23 is an exemplary view for explaining an embodiment of a TSP packing method according to the present disclosure.
  • FIG. 24 is an exemplary view for explaining another embodiment of a TSP packing method according to the present disclosure.
  • 25 is an illustration of a typical fisheye video comprising two circular images in accordance with the present disclosure.
  • 26A is an exemplary diagram of stereoscopic fisheye video in a vertical stereo format according to the present disclosure.
  • 26B is an illustration of stereoscopic fisheye video in left and right stereo format according to the present disclosure.
  • 27 is an exemplary diagram of stereoscopic fisheye video having a pair-by-pair format for multiview according to the present disclosure.
  • 28 is an exemplary diagram of stereoscopic fisheye video having a group-by-group format for multiview according to the present disclosure.
  • 29 is an exemplary diagram for describing a fisheye camera according to the present disclosure.
  • FIG. 30 shows a displayed FOV for two fisheye images, in a fisheye camera according to the present disclosure.
  • FIG. 31 illustrates an overlapped FOV with a displayed FOV for multiple fisheye images, in a fisheye camera according to the present disclosure.
  • 32 is an exemplary view for explaining the center of a fisheye camera according to the present disclosure.
  • 33 is an exemplary diagram for describing parameters regarding a local field of view according to the present disclosure.
  • a “component surface” includes one or more component surfaces.
  • first and second may be used to describe various components, but the components are not limited by the terms. The terms are used only for the purpose of distinguishing one component from another.
  • first component may be referred to as a second component, and similarly, the second component may also be referred to as a first component.
  • an electronic device may include a communication function.
  • the electronic device may include a smart phone, a tablet personal computer (PC), a mobile phone, a video phone, and an e-book reader (e).
  • -book reader desktop PC, laptop PC, netbook PC, personal digital assistant (PDA), portable Portable multimedia player (PMP, hereinafter referred to as 'PMP'), MP3 player, mobile medical device, camera, wearable device (e.g., head-mounted) Head-mounted device (HMD), for example referred to as 'HMD', electronic clothing, electronic bracelet, electronic necklace, electronic accessory, electronic tattoo, or smart watch ), Etc.
  • the electronic device may be a smart home appliance having a communication function.
  • the smart home appliance includes a television, a digital video disk (DVD) player, an audio, a refrigerator, an air conditioner, a vacuum cleaner, an oven, Microwave oven, washer, dryer, air purifier, set-top box, TV box (e.g. Samsung HomeSyncTM, Apple TVTM, or Google TVTM), gaming console ), An electronic dictionary, a camcorder, an electronic photo frame, and the like.
  • DVD digital video disk
  • an electronic device may include a medical device (eg, magnetic resonance angiography (MRA) device), and magnetic resonance imaging (MRI).
  • MRI magnetic resonance angiography
  • MRI magnetic resonance imaging
  • CT computed tomography
  • EDR event data recorder
  • FDR flight data recorder
  • automotive infotainment device navigation electronic device (e.g. navigation navigation device, gyroscope) ope, or compass), avionics, security devices, industrial or consumer robots, and the like.
  • an electronic device may include furniture, part of a building / structure, an electronic board, an electronic signature receiving device, a projector, and various measurement devices (eg, water) that include communication functionality. And electrical, gas, or electromagnetic wave measuring devices).
  • the electronic device may be a combination of devices as described above.
  • the electronic device according to the preferred embodiments of the present disclosure is not limited to the device as described above.
  • a device for transmitting and receiving VR content may be, for example, an electronic device.
  • the image may be a video, a still image, or the like, and the image content may include various multimedia contents including video, still images, and the like, related audio, subtitles, and the like.
  • the VR content includes image content that provides the image as a 360 degree image, a 3D image, or the like.
  • the media file format may be a media file format according to various media related standards such as an International Organization for Standardization (ISO) -based media file format (ISOBMFF).
  • ISO International Organization for Standardization
  • ISOBMFF International Organization for Standardization
  • projection refers to a process in which a spherical image for representing a 360 degree image or the like is projected onto a planar surface or an image frame according to a result of the processing.
  • Omnidirectional media are, for example, images or videos that can be rendered according to the direction of the user's head movement or when the user uses the HMD or according to the user's viewport. Or related audio.
  • the view port may be referred to as a field of view (FOV), and refers to an area of an image that is displayed to a user at a specific point in time, where the area of the image may be an area of the spherical image.
  • FOV field of view
  • FIG. 1 is an exemplary view for explaining the configuration of a computer system that implements a stereo omnidirectional image packing method according to the present invention.
  • a computer system may include at least one processor 110 and a memory 120.
  • the processor 110 may be a central processing unit (CPU) or a semiconductor device that processes instructions stored in the memory 120.
  • CPU central processing unit
  • semiconductor device that processes instructions stored in the memory 120.
  • the processor 110 may be a controller that controls all operations of the computer system 100.
  • the controller may execute operations in which the computer system 100 operates by reading and executing the program code stored in the memory 120.
  • Computer system 100 may include a user input device 150, a data communication bus 130, a user output device 160, and a storage 140. Each of the above components may be in data communication via the data communication bus 130.
  • the computer system can further include a network interface 170 coupled to the network 180.
  • Memory 120 and storage 140 may include various types of volatile or nonvolatile storage media.
  • the memory 120 may include a ROM 123 and a RAM 126.
  • Storage 140 may include non-volatile memory such as magnetic tape, hard disk drive (HDD), solid state drive (SDD), optical data device, and flash memory.
  • the packing method of the stereo omnidirectional image according to the embodiment of the present invention may be implemented by a computer executable method.
  • computer readable instructions may perform the operating method according to the present invention.
  • the above-described packing method of stereo omnidirectional image according to the present invention may be implemented as computer readable codes on a computer readable recording medium.
  • Computer-readable recording media include all kinds of recording media having data stored thereon that can be decrypted by a computer system. For example, there may be a read only memory (ROM), a random access memory (RAM), a magnetic tape, a magnetic disk, a flash memory, an optical data storage device, and the like.
  • the computer-readable recording medium can also be distributed over computer systems connected by a computer communication network, and stored and executed as code readable in a distributed fashion.
  • patch_shape represents the shape of a patch, that is, a rectangle, an isosceles triangle, a right triangle, and the like.
  • the patch may mean each area included in each view of the packed frame, or may mean each area included in each view of the projected frame.
  • patch_orientation indicates the rotation and flip of a patch shape indicating the orientation of various shapes.
  • patch_transform indicates the rotation and flip of the image data specified by the patch.
  • a region-based packing method for each region is proposed.
  • FIG. 2 illustrates a left and right stereoscopic 360 format according to the present disclosure
  • FIG. 3 illustrates a tom-bottom stereoscopic 360 format.
  • FIG. 4 illustrates image stitching, projection, and packing per region of a single acquisition time instance.
  • the area-specific packing method can flexibly subdivide the projected frame into a plurality of areas. Each region can be resized and relocated to a packed frame.
  • a method of packing by region for both monoscopic 360 video and stereoscopic 360 video will be described.
  • the input images of one time instance are stitched to produce a projected frame representing one view.
  • the input images of one time instance are stitched to produce a projected frame representing two views (one for each eye). Both views are mapped to the same packed frame and encoded by a conventional 2D (2 dimensional) video encoder.
  • each view of the projected frame may be mapped to a packed frame, respectively.
  • the sequence of packed frames of the left view or the right view may be coded independently and when using a multiview video encoder, it may be predicted from another view.
  • the region-by-area packing method of the stereo 360 video format and the stereo 360 video format have been agreed, certain parameters defining the layout of the stereo 360 video format have not been proposed or adopted yet.
  • This disclosure proposes several types of defining the layout of stereoscopic 360 video in packed frames. Each type has its own advantages. For example, according to the fully mixed-independent packing method, the left view and the right view can achieve good performance in terms of coding efficiency, but in tile-based delivery for viewport dependent streaming, it is appropriate to pack the left and right views in pairs. Do. The syntax and meaning for packing by region will be described later.
  • images of concurrent instances (B i ) are mapped to stitched, projected, and packed frames (D).
  • D is a schematic diagram of an image stitching, projection, and packing process for each region.
  • the input images Bi are stitched and projected onto a three-dimensional projection structure such as a sphere or a cube.
  • Image data on the projection structure is further arranged on the two-dimensional projection frame (C).
  • the format of the two-dimensional projection frame is indicated by a projection format indicator defined in coding independent media description code points (CICP) or omnidirectional media application format (OMAF).
  • CICP independent media description code points
  • OMAF omnidirectional media application format
  • Optional per region packing is applied to map the two-dimensional projection frame C into one or more packed frames D.
  • FIG. If no per-field packing is applied, the packed frame will be identical to the projected frame. Otherwise, the regions of the projected frame are mapped to the one or more packed frames D by indicating the location, shape, and size of each area of the one or more packed frames D.
  • the input images are converted into packed frames by a process without an intermediate process.
  • both the left view and the right view may be packed in the same packed frame. Then, when the stereoscopic formats of the left view and the right view are the same, each view of the native layout may be placed in the left or right area. If area-specific packing is applied to each view or both views, for each embodiment various layouts are possible.
  • two parameters are employed. The two parameters are stereo_format and stereo_packing_type.
  • the stereo_format parameter is an indicator that specifies a stereoscopic format, such as side-by-side or top-bottom.
  • stereo_packing_type defines a layout type for packing for each stereoscopic region.
  • the layout type relates to whether positions of respective regions belonging to the left view or the right view are separated, mixed, independent, or correspond to each other.
  • Each stereo_packing_type has advantages in terms of coding efficiency and functionality.
  • the following figures assume the same case as the left-right stereoscopic 360 format.
  • FIG. 5 is an exemplary view for explaining a non-area packing method according to the present disclosure.
  • Non-region-wise packing is possible using native layout rather than per-region packing.
  • stereo_packing_type corresponds to non-region-wise packing
  • each projected frame using the basic layout is placed in the left and right regions without shuffling.
  • the packing method using the default layout is the simplest layout and an efficient way to quickly extract and render each view. Since the projected frame and the packed frame are the same, the data structure of the image data is not changed.
  • FIG. 6 is an exemplary diagram for explaining a separate and independent packing method according to the present disclosure.
  • each projected frame having a basic layout of projection may be placed in the left-right region.
  • each half frame corresponding to each view is internally recognized by region-specific packing.
  • Each view is separated, but the local regions included in each view are sampled again and placed in half packed frames corresponding to the same view.
  • the separate-independent packing layout is effective for fast extraction and coding efficiency.
  • each view will have to be recognized for rendering after being decoded.
  • FIG. 7 is an exemplary view for explaining a separation and mirroring packing method according to the present disclosure.
  • each projected frame having a basic layout of projection may be placed in the left-right region.
  • each half frame corresponding to each view is internally recognized by region-specific packing.
  • each view is separated, but the local areas included in each view are resampled and placed in half packed frames corresponding to the same view.
  • the difference from the separate-independent packing is that the packing method for each area of one view and the packing method for each area of another view are the same. Compared with separate-independent packing, bits can be saved. Since the area-specific packing parameters of one view are the same as the area-specific packing parameters of another view, the area-specific packing parameters of one view do not need to be signaled.
  • FIG. 8 is an exemplary diagram for explaining a mixed and independent packing method according to the present disclosure.
  • stereo_packing_type is a mixed and independent packing method
  • each region of the projected frame of one view is resampled and placed at a particular location of the packed frame.
  • the advantage of the mixed-independent packing method is the coding efficiency. According to the mixed-independent packing method, an optimum layout with full flexibility in terms of compression can be found. However, extracting a view from a frame packed view is complicated, and the view must be recognized for rendering.
  • FIG 9 is an exemplary view for explaining a mixed and pair-wise packing method according to the present disclosure.
  • each region of the projected frame of the left view is resampled and placed at a specific position of the packed frame.
  • the corresponding area (same location, same size) of the projected frame of the right view is then sampled identically to the left view and is located to the right of the projected area of the left view.
  • the right view area can be located at the bottom portion of the packed area of the left view.
  • the main advantage of per-pair packing is that in all the left and right area projected frames. It is located in pairs. Thus, it is suitable for tile based delivery and rendering.
  • the area packed for each pair may be a tile. When specific tiles that are dependent on the current viewport are delivered, the stereoscopic views can always be displayed because each tile includes a left view and a right view. Bits representing the region-specific packing parameters for the right view will be saved as well.
  • FIG. 10 is an exemplary view for explaining a packing method for a regular polyhedral projection image according to the present disclosure.
  • the advantage of the mixed-independent packing method is the coding efficiency. According to the mixed-independent packing method, an optimum layout with full flexibility in terms of compression can be found. However, extracting a view from a frame packed view is complicated, and the view must be recognized for rendering.
  • This disclosure will present multiple layouts of each projection to find the best layout in terms of coding efficiency and memory usage. By observing that the packed projection performs better, several methods for packing to remove projection redundancy can be compared to the native unfolding or unrolling method.
  • FIG. 11 is an exemplary view for explaining a packing method for each region using a triangular patch according to the present disclosure.
  • the present disclosure should determine in advance which projection method OMAF has been adopted.
  • PACK-VE in the scope of pack verification experiments (PACK-VE), a generalized region-based packing method using a plurality of patches is proposed to enable a triangle-based packing method.
  • Some projection methods can be used in OMAF by using the basic projection method or the selective projection method or other extended mechanisms possible by unifrom resource indicators (URIs) and the triangle-based tetrahedrons (octahedrons, icosahedrons) Assume that you can.
  • URIs unifrom resource indicators
  • octahedrons, icosahedrons triangle-based tetrahedrons
  • CMP cube-based projection
  • OHP octahedron based projection
  • ISP icosahedron based projection
  • SSP segmented sphere projection
  • TSP Trunked Square Pyramid
  • each area indicated by a specific patch can be resampled and relocated from the projected frame to the packed frame.
  • the patch is shaped to specify image data to be packed.
  • Three parameters (patch_shape, patch_orientation, patch_transform) are proposed so that regions corresponding to various faces of various three-dimensional geometry (eg, cubes, octahedrons, icosahedrons, etc.) can be specified by various tetrahedra.
  • phatch_shape represents the patch shape (rectangle, isosceles triangle, right triangle, etc.)
  • patch_orientation represents the patch shape rotation and flips representing various shape orientations
  • patch_transform represents the rotation of image data specified by the patch. And flip.
  • FIG. 11 (a) is an exemplary diagram for describing a parameter of a triangular patch of a projected frame, and includes coordinates (proj_region_top_left_x, proj_region_top_left_y), width (proj_region_width), and height (proj_region_height) of the top-left of a region included in the projected frame.
  • Patch type patch_type, patch_shape
  • patch orientation patch_orientation
  • FIG. 11 (b) is an exemplary diagram for describing a parameter of a triangular patch of a packed frame, and includes coordinates (pack_region_top_left_x, pack_region_top_left_y), width (pack_region_width), and height (pack_region_height) of the top-left of a region included in the packed frame.
  • Patch_transform A patch type of 2 means that the patch is an isosceles triangle.
  • a patch transformation of 6 rotates the projected frame area 270 degrees counterclockwise to It means that you have created an area.
  • Table 1 is a syntax illustrating a data structure used to perform a stereoscopic region-specific packing method according to the present disclosure.
  • Table 2 shows setting values of stereo_format for specifying a stereoscopic 360 video format.
  • stereo_format 0x00 Reserved 0x01 Left-right stereoscopic 360 format 0x02 Top-bottom stereoscopic 360 format 0x03-0xFF Reserved
  • Table 3 shows setting values of stereo_packing_type for specifying a region-specific packing type for stereoscopic 360 video.
  • stereo_packing_type 0x00 reserved 0x01 no region-wise packing (native) 0x02 separate and independent packing 0x03 separate and mirroring packing 0x04 mixed and independent packing 0x05 mixed and mirroring packing 0x06-0xFF Reserved
  • stereo_packing_type 1
  • this specifies a projected frame having a basic layout of projections located in the left and right regions (or top and bottom regions) without shuffling.
  • stereo_packing_type is 2
  • each projected frame with a basic layout is located in the left or right area. Then, each half frame corresponding to each view is internally recognized by region-specific packing. Each view is separated, but the local regions included in each view are sampled again and placed in half packed frames corresponding to the same view.
  • the separate-independent packing layout is effective for fast extraction and coding efficiency. However, each view will have to be recognized for rendering after being decoded.
  • each projected frame having a basic layout of projection can be placed in the left-right region. Then, each half frame corresponding to each view is internally recognized by region-specific packing. Thus, each view is separated, but the local areas included in each view are resampled and placed in half packed frames corresponding to the same view. The difference from the separate-independent packing is that the packing method for each area of one view and the packing method for each area of another view are the same.
  • stereo_packing_type 4
  • each area of the projected frame of one view is resampled and placed at a specific location of the packed frame. There is no restriction for recognizing left and right frames projected onto the same packed frame.
  • stereo_packing_type is 5
  • each area of the projected frame of the left view is resampled and placed at a specific position of the packed frame.
  • the corresponding area (same location, same size) of the projected frame of the right view is then sampled identically to the left view and is located to the right of the projected area of the left view.
  • the right view area may be located at the bottom portion of the packed area of the left view.
  • FIG. 12 is an exemplary view for explaining the layout of the left and right regions used in the non-regional packing method according to the present disclosure, in which the projected frames and the stereo_packing_type are no region-wise packing (native), separate and independent packing, and separate.
  • the layout of the left and right regions of the packed frame in the case of and mirroring packing, mixed and independent packing, mixed and mirroring packing
  • FIG. 13 is an exemplary diagram for describing a layout of an upper and lower regions used in a non-regional packing method according to the present disclosure, in which projected frames and stereo_packing_type are no region-wise packing (native) (0x01), separate and independent The layout of the upper and lower regions of the packed frame when packing (0x02), separate and mirroring packing (0x03), mixed and independent packing (0x04), and mixed and mirroring packing (0x05) is shown.
  • width_proj_frame is the width of the projected frame.
  • height_proj_frame means the height of the projected frame.
  • num_of_regions means the number of packed regions specified by the patch.
  • uniform_region_size 1
  • the projected frame is divided into regions of the same size specified by uniform_region_width and uniform_region_height. If uniform_region_size is 0, the i-th region of the projected frame (i is an integer from 0 to num_of_regons-1). It is specified by the size specified by this proj_region_width [i] and proj_region_height [i].
  • uniform_region_width and uniform_region_height specify each region of the projected frame with the same width and height.
  • proj_region_width [i] and proj_region_height [i] specify the i-th region of the projected frame.
  • patch_shape [i] specifies the shape of the i-th region to be rearranged into the packed frame.
  • Table 4 shows the shape of each area of the projected frame according to patch_shape.
  • FIG. 14A shows that patch_shape is 0x01 (rectangle)
  • FIG. 14B shows that patch_shape is 0x02 (isosceles triangle)
  • FIG. 14C shows that patch_shape is 0x03 (right triangle)
  • patch_orientation [i] specifies the shape of the patch that has been rotated and flipped from the original patch shape (i-th area of the projected frame) indicated by patch_shape [i].
  • Table 5 shows the meaning of the rotation or flip according to patch_orientation [i].
  • patch_transform [i] specifies the rotation and flip of the image data specified by patch_orientation [i] and patch_shape [i] to be rearranged into the packed frame.
  • Table 6 shows the meaning of rotation or flip according to patch_transform [i].
  • packed_region_width [i] and packed_region_height [i] specify the width and height of the packed region of the packed frame corresponding to the i th region of the projected frame.
  • packed_region_top_left_x [i] and packed_region_top_left_y [i] specify the horizontal and vertical coordinates of the top-left corner of the packed region of the packed frame corresponding to the i th region of the projected frame.
  • FIG. 15 is an exemplary diagram for explaining a region-specific packing method of adjusting and rearranging an area according to latitude in an isotropic projection (ERP) according to the present disclosure.
  • OMAF incorporates a region-by-region packing method that removes redundant regions, thereby improving the projected coding efficiency.
  • an isotropic projection stitches each parallel of the sphere, transforming the sphere into a planar rectangular region. The range of stitching increases extremely in the polar direction.
  • the coding efficiency of the projected frame may be improved by reducing the region of the polar region.
  • the first and fifth regions corresponding to the high latitude region are sampled at a 1: 3 ratio
  • the middle latitude region (more than 30 degrees and less than 60 degrees, or less than -30 degrees).
  • the second area and the fourth area corresponding to -60 degrees or more) are sampled at a 2: 3 ratio
  • the third area corresponding to the low latitude area is sampled at a 1: 1 ratio
  • the packed frame may be obtained by rearranging the sampled regions as shown in FIG. 15C.
  • FIG. 16 is an exemplary diagram for explaining region-specific packing for a cube projection for viewport dependent streaming according to the present disclosure.
  • FIG. 16 shows an exemplary view of the area-by-area packing for a cube map of a projected frame consisting of a front face and five down sampled faces (left side, right side, back side, top side, bottom side) of 1/5.
  • rectangle-to-trapezoid mapping In order to improve the flexibility of packing by region, we propose a rectangle-to-trapezoid mapping.
  • the rectangle-to-trapezoid mapping enables various and effective area-specific packing methods. If the short edge is 1 pixel, it becomes a triangle.
  • 17 is an exemplary diagram for explaining a method of packing an ERP image according to the present disclosure.
  • ERP creates an extremely stretched pole region.
  • polarity redundancy pixels unnecessarily reduce the coding efficiency of the video.
  • FIG. 17 illustrates a region-specific packing approach that reduces the sampling rate of the pole region of an isquirectangular panorama.
  • the projected frame is first divided into eight rectangular sub-regions, and using line-down downsampling, each region is converted into a triangular shape and rearranged to form a rectangular format.
  • one embodiment of a method of packing an ERP image according to the present disclosure extremely reduces the number of pixels in the polar region, while maintaining the relatively equatorial region. Furthermore, the packaged frame is represented by a rectangular layout without discontinuities between sub-regions, and blank pixels do not contain scene information.
  • FIG. 18 is an exemplary diagram for describing a method of packing an ERP image according to the present disclosure.
  • 19 is an exemplary diagram for explaining a method of converting an isotonic projection according to the present disclosure into a layout similar to a cube.
  • an isotropic projected frame can be converted into a cube-like layout.
  • the top region and the bottom region are each divided into four subregions, each subregion is converted into a triangular region, and the like cube. It is relocated to the layout.
  • 20 is an exemplary diagram for explaining another embodiment of converting an isotonic projection according to the present disclosure into a layout similar to a cube.
  • FIG. 19 is an example of a 4x3 cube map layout
  • FIG. 20 is an example of a 3x2 cube map layout.
  • 21 is an exemplary diagram for describing a method of converting an ERP image into a cube-like ERP according to the present disclosure.
  • FIG. 21 shows an ERP image, a 4x3 cube map layout according to FIG. 20, and a 3x2 cube map layout according to FIG. 20.
  • FIG. 22 is an exemplary view for explaining a TSP packing method according to the present disclosure.
  • the cube map frame can be converted to a TSP.
  • the front may be a square sampled at a 1: 1 ratio
  • the back may be a square sampled at a 1: 9 ratio
  • the right, left, top, and bottom may be sampled at a 2: 9 ratio.
  • FIG. 23 is an exemplary view for explaining an embodiment of a TSP packing method according to the present disclosure.
  • the rectangular area of the packed frame is defined by four parameters.
  • the four parameters are the horizontal and vertical coordinates (pack_reg_left, pack_reg_top) and the width and height (pack_reg_width, pack_reg_height) of the top left vertex.
  • the rectangle side is defined as the shorter side of the trapezoid represented by the offset information (pack_sb_offset) 2320 and the length (pack_sb_length) 2330 indicating the position of the start point 2310. Define the trapezoidal area by setting.
  • FIG. 24 is an exemplary view for explaining another embodiment of a TSP packing method according to the present disclosure.
  • another parameter pack_sb_indicator is defined to indicate which side is a short side. For example, if pack_sb_indicator is 1, the upper side may be shorter, if pack_sb_indicator is 2, the lower side may be shorter, if pack_sb_indicator is 3, the left side may be shorter, and if pack_sb_indicator is 4, the right side may be shorter.
  • Table 7 shows the syntax for implementing the TSP packing method.
  • proj_frame_width specifies the width of the projected frame.
  • proj_frame_height specifies the height of the projected frame.
  • number_of_regions Specifies the number of subregions of the projected frame.
  • proj_reg_top [n] specify the x and y coordinates of the upper left corner of the nth rectangular subarea of the projected frame
  • proj_reg_width [n] specify the nth rectangular subareas of the projected frame Specify the width and height of the area.
  • pack_reg_top [n] specify the x and y coordinates of the upper left corner of the nth rectangular subarea of the packed frame
  • pack_reg_width [n] specify the nth rectangular sub of the packed frame Specify the width and height of the area.
  • pack_sb_offset [n] specifies the distance from the upper left vertex of the nth rectangular sub-region of the projected frame to the start of the shorter side.
  • pack_sb_length [n] specifies the length of the shorter side of the nth rectangular subregion of the projected frame.
  • pack_sb_indicators [n] specifies the location with the shorter side of the nth trapezoidal subregion of the packed frame that corresponds to the nth rectangular subregion of the projected frame. If pack_sb_indicators [n] is greater than zero, the nth rectangular subregion of the projected frame is trapezoidal, and if pack_sb_indicators [n] is zero, it is rectangular. Table 8 shows the positions of the shorter sides according to pack_sb_indicators [n].
  • pack_sb_indicators [n] 0 no shorter base (rectangular region)
  • proj_reg_rotation [n] specifies the clockwise rotation of the image data corresponding to the nth sub-region of the projected frame.
  • Table 9 shows rotation angles according to proj_reg_rotation [n].
  • the circular images taken by the fisheye cameras are directly encoded and transmitted.
  • the decoded image / video is rendered directly according to the viewport intended by the user. This method is useful for low latency live streaming or high quality 360 video delivery because images taken without intermediate projection methods, such as isotropic or cube map projection, are rendered directly.
  • 25 is an illustration of a typical fisheye video comprising two circular images in accordance with the present disclosure.
  • 26A is an exemplary diagram of stereoscopic fisheye video in a vertical stereo format according to the present disclosure.
  • 26B is an illustration of stereoscopic fisheye video in left and right stereo format according to the present disclosure.
  • 27 is an exemplary diagram of stereoscopic fisheye video having a pair-by-pair format for multiview according to the present disclosure.
  • 28 is an exemplary diagram of stereoscopic fisheye video having a group-by-group format for multiview according to the present disclosure.
  • the image frame may comprise omnidirectional fisheye video.
  • the decoded omnidirectional fisheye video is stitched and rendered according to the user's intended viewport using the signaled fisheye video parameters.
  • the fisheye video parameters include lens distortion correction (LDC) parameters with a local field of view (FOV), lens shading compensation parameters with red-green-blue gains. At least one of a displayed field of view information and a camera extrinsic parameter.
  • Table 10 shows syntax for stereoscopic fisheye video for multiview.
  • 29 is an exemplary diagram for describing a fisheye camera according to the present disclosure.
  • the meaning of each term is as follows.
  • num_circular_images specifies the number of circular images in the coded picture of each sample. num_circular_images can be 2 or any other nonzero integer.
  • image_center_x is a fixed point 16,16 value indicating the horizontal coordinate of the center of the circular image in the encoded picture of each sample to which the present syntax is applied in the luma samples.
  • image_center_y is a fixed point 16,16 value indicating the vertical coordinate of the center of the circular image in the encoded picture of each sample to which the present syntax is applied in the luma samples.
  • full_radius is a fixed point 16,16 value that indicates the radius from the center of the circular image to the edge of the full round image in luma samples.
  • frame_radius is a fixed point 16,16 value that indicates the radius from the center of the circular image to the edge of the nearest image boundary in luma samples.
  • the circular fisheye image can be cropped by the camera frame, and frame_radius is the radius of the circle indicating the pixels that are not available.
  • scene_radius is a fixed point 16,16 value that indicates the radius from the center of the circular image to the edge of the region of the nearest image in the luma samples.
  • the image area is an area free of obstructions from the camera body itself, and for stitching, there is no lens distortion too large.
  • FIG. 30 shows a displayed FOV for two fisheye images, in a fisheye camera according to the present disclosure.
  • image_rotation is a fixed point 16.16 that indicates the amount of rotation of the circular image in degrees.
  • Different video camera manufacturers use different coordinate systems or different layouts for each photographed individual fisheye image.
  • the image can range from -90 degrees to +90 degrees or from -180 degrees to +180 degrees.
  • image_flip indicates whether the image is flipped or how flipped. Thus, the reverse flip operation needs to be applied. If image_flip is 0, the image was not flipped. If image_flip is 1, the image is flipped vertically. If image_flip is 2, the image is flipped horizontally. If image_flip is 3, the image is flipped horizontally and flipped vertically
  • image_scale_axis_angle, image_scale_x, and image_scale_y are fixed point 16.16 values that indicate along which axis the image is scaled and how scaled. By indicating the value of image_scale_axis_angle in angle units, the axis is defined by a single angle. An angle of zero (image_scale_axis_angle) means that the horizontal vector is completely horizontal and the vertical vector is completely vertical.
  • image_scale_x and image_scale_y indicate the scaling ratios of the directions parallel and perpendicular to the axis, respectively.
  • field_of_view is a fixed point 16.16 value indicating the FOV of the fisheye lens in angle units.
  • the typical value (field_of_view) of the hemispherical fisheye lens is 180 degrees.
  • num_angle_for_displaying_fov indicates the number of angles. If num_angle_for_displaying_fov is 12, the fisheye image is divided into 12 sectors. The angle of the angular sector is 30 degrees. The value of the FOV superimposed with the displayed FOV is defined clockwise.
  • displayed_fov indicates the rendered and displayed FOV and the corresponding image area of each fisheye camera image.
  • overlapped_fov indicates overlapped regions in terms of FOV between multiple circular images.
  • scene_radius represents the relationship between the fisheye lens and the camera body.
  • the values may vary depending on the characteristics of the lens and the content.
  • the stitching quality having displayed_fov values is 170 degrees for the left camera and the quality for the right camera is better than the default value (180 degrees) of 190 degrees, the values of the displayed display_fov may be updated.
  • FIG. 31 illustrates an overlapped FOV with a displayed FOV for multiple fisheye images, in a fisheye camera according to the present disclosure.
  • a single displayed_fov value may not account for the exact region of each fisheye image.
  • displayed_fov (dark portion) varies depending on the direction.
  • num_angle_for_displaying_fov is introduced, and displayed_fov and overlapped_fov are defined in the clockwise direction.
  • 32 is an exemplary view for explaining the center of a fisheye camera according to the present disclosure.
  • num_polynomial_coefficients is an integer specifying the number of coefficients present in the polynomial.
  • List of coefficients of polynomial polynomial_coefficient_K is a fixed point 16.16 value representing the coefficients of the polynomial describing the transformation of the fisheye space into an undistorted plane image.
  • num_local_fov_region indicates the number of local fitting regions having different field of view (FOV).
  • Start_radius, end_radius, start_angle, and end_angle indicate an area for local fitting / warping that changes the actual FOV for locally displaying.
  • radius_delta indicates a delta value for indicating a different FOV for each radius.
  • angle_delta indicates a delta value for indicating a different FOV for each angle.
  • local_fov_weight indicates a weight value for the FOV of the position specified by start_radius, end_radius, start_angle, end_angle, the angle index i and the radius index j.
  • 33 is an exemplary diagram for describing parameters regarding a local field of view according to the present disclosure.
  • num_polynomial_coefficeients_lsc may be an order of polynomial approximation of the lens shading curve.
  • polynomial_coefficient_K_lsc_R may be a polynomial coefficient approximating the lens shading curve for the red color component in the fixed point 16.16 format.
  • polynomial_coefficient_K_lsc_G may be a polynomial coefficient that approximates the lens shading curve for the green color component in the fixed point 16.16 format.
  • polynomial_coefficient_K_lsc_B may be a polynomial coefficient that approximates the lens shading curve for the blue color component in the fixed point 16.16 format.
  • num_deadzones is an integer indicating the number of dead zones in the coded picture of each sample applied by this syntax.
  • deadzone_left_horizontal_offset, deadzone_top_vertical_offset, deadzone_width, and deadzone_height are integer values indicating the position and size of the dead zone rectangular area. You can't use pixels in the dead zone.
  • deadzone_left_horizontal_offset and deadzone_top_vertical_offset indicate, in luma samples, the horizontal and vertical coordinates of the upper left corner of the dead zone in the encoded picture, respectively.
  • deadzone_width and deadzone_height indicate the width and height of the dead zone in luma samples, respectively.
  • all the pixels in the dead zone are set to the same pixel value (eg all black).
  • a method for transmitting stereoscopic video content comprising: a plurality of projections from the plurality of omnidirectional images based on data of a stereoscopic image including a plurality of omnidirectional images having parallax; Generating a first frame comprising first views; Generating a second frame including a plurality of second views by packing a plurality of first regions included in the plurality of first views based on region-wise packing information; And transmitting data relating to the generated second frame, wherein the plurality of second views includes a plurality of second regions corresponding to the plurality of first regions, and the packing information for each region may include: It includes information about the shape, orientation or transformation of each of the plurality of second regions.
  • the packing information for each region may further include information indicating whether the stereoscopic video has a left and right stereoscopic 360 format or a vertical stereoscopic 360 format.
  • the packing information for each area may be stereoscopic indicating one of non-application of packing by area, packing by separate-independent area, packing by separate-mirroring area, packing by mixed-independent area, and packing by mixed-pair area. It may further include a packing type.
  • the information on the shape of each of the plurality of second regions indicates one of the plurality of shapes as the shape of each of the plurality of second regions, and the plurality of shapes may include a trapezoid.
  • the method for transmitting stereoscopic video content according to the present disclosure further includes generating an omnidirectional image of one of the plurality of omnidirectional images based on images acquired by the plurality of fisheye lenses,
  • the information about the one omnidirectional image may include: information indicating the number of divided regions for dividing an image acquired by each of the plurality of fisheye lenses according to a specific angle with respect to a center; Information indicating an area corresponding to a field of view (FOV) in each of the divided areas; And information indicating an area overlapping an image acquired by another fisheye lens in each of the divided areas.
  • FOV field of view
  • each of the plurality of first views may be a spherical projection image, an equirectangular projection image (ERP image), or a tetrahedral projection image
  • the regular polyhedral projection image may be a tetrahedral projection image, a cube projection image, an octahedron projection image, It may be a dodecahedron projection image or a dodecahedron projection image.
  • the packing information for each area may further include location information and size information of the plurality of first areas and location information and size information of the plurality of second areas.
  • the position information and the size information of each of the plurality of first regions may include the position information of the corresponding second region among the plurality of second regions, and It may be the same as the size information.
  • the stereoscopic packing type indicates packing for each separation-independent area
  • the plurality of second views may be separated and packed independently.
  • the plurality of second views may be separated and packed in the same manner.
  • the plurality of second views may be mixed with each other, and the plurality of second views may be independently packed.
  • the plurality of second views may be mixed with each other, paired, and packed.
  • the plurality of first views may be a cube projection images including a front surface, a rear surface, a left surface, a right surface, an upper surface, and a lower surface
  • the plurality of second regions may be the front surface, the rear surface, the left surface, and the right surface.
  • Each of the areas corresponding to the left side, the right side, the top side, and the bottom side of the plurality of second regions may have a trapezoidal shape.
  • the size of the region corresponding to the front surface of the plurality of second regions may be larger than the size of the region corresponding to the rear surface.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Security & Cryptography (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)

Abstract

Un procédé de transmission de contenu vidéo stéréoscopique selon la présente invention comprend les étapes consistant : à générer, sur la base de données d'une vidéo stéréoscopique qui comprend une pluralité de vidéos omnidirectionnelles ayant une parallaxe, une première trame comprenant une pluralité de premières vues projetées à partir de la pluralité de vidéos omnidirectionnelles ; à générer une seconde trame comprenant une pluralité de secondes vues par emballage, sur la base d'informations d'emballage par région, une pluralité de premières régions comprises dans la pluralité de premières vues ; et à transmettre des données sur la seconde trame générée, la pluralité de secondes vues comprenant une pluralité de secondes régions correspondant à la pluralité de premières régions, et les informations d'emballage par région comprenant des informations sur la forme, l'orientation ou la transformation pour chacune de la pluralité de secondes régions.
PCT/KR2017/014742 2017-01-10 2017-12-14 Procédé et appareil permettant de transmettre un contenu vidéo stéréoscopique Ceased WO2018131803A1 (fr)

Priority Applications (3)

Application Number Priority Date Filing Date Title
CN201780082778.3A CN110463196B (zh) 2017-01-10 2017-12-14 用于传输立体视频内容的方法和装置
US16/477,102 US10855968B2 (en) 2017-01-10 2017-12-14 Method and apparatus for transmitting stereoscopic video content
EP17891098.0A EP3570540A4 (fr) 2017-01-10 2017-12-14 Procédé et appareil permettant de transmettre un contenu vidéo stéréoscopique

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US201762444411P 2017-01-10 2017-01-10
US62/444,411 2017-01-10
KR10-2017-0171492 2017-12-13
KR1020170171492A KR102503342B1 (ko) 2017-01-10 2017-12-13 스테레오 스코픽 비디오 콘텐츠를 전송하기 위한 방법 및 장치

Publications (1)

Publication Number Publication Date
WO2018131803A1 true WO2018131803A1 (fr) 2018-07-19

Family

ID=62839605

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2017/014742 Ceased WO2018131803A1 (fr) 2017-01-10 2017-12-14 Procédé et appareil permettant de transmettre un contenu vidéo stéréoscopique

Country Status (1)

Country Link
WO (1) WO2018131803A1 (fr)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113748681A (zh) * 2019-01-09 2021-12-03 腾讯美国有限责任公司 动态点云分区打包方法和装置
CN113841416A (zh) * 2019-05-31 2021-12-24 倬咏技术拓展有限公司 互动沉浸洞穴网络

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140078249A1 (en) * 2012-09-20 2014-03-20 Qualcomm Incorporated Indication of frame-packed stereoscopic 3d video data for video coding
US20140132598A1 (en) * 2007-01-04 2014-05-15 Hajime Narukawa Method of mapping image information from one face onto another continous face of different geometry
US20160021354A1 (en) * 2014-07-16 2016-01-21 Arris Enterprises, Inc. Adaptive stereo scaling format switch for 3d video encoding
US20160337670A1 (en) * 2014-02-10 2016-11-17 Lg Electronics Inc. Method and apparatus for transmitting/receiving broadcast signal for 3-dimensional (3d) broadcast service

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140132598A1 (en) * 2007-01-04 2014-05-15 Hajime Narukawa Method of mapping image information from one face onto another continous face of different geometry
US20140078249A1 (en) * 2012-09-20 2014-03-20 Qualcomm Incorporated Indication of frame-packed stereoscopic 3d video data for video coding
US20160337670A1 (en) * 2014-02-10 2016-11-17 Lg Electronics Inc. Method and apparatus for transmitting/receiving broadcast signal for 3-dimensional (3d) broadcast service
US20160021354A1 (en) * 2014-07-16 2016-01-21 Arris Enterprises, Inc. Adaptive stereo scaling format switch for 3d video encoding

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
LIN, HUNG-CHIH ET AL.: "AHG8: An Efficient Compact Layout for Octahedron Format", JVET-D0142 (VERSION 2), JOINT VIDEO EXPLORATION TEAM (JVET) OF ITU-T SG 16 WP 3 AND ISO/IEC JTC 1/SC 29/WG 11 4TH MEETING, 16 October 2016 (2016-10-16), Chengdu, CN, pages 1 - 5, XP055506578 *
See also references of EP3570540A4 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113748681A (zh) * 2019-01-09 2021-12-03 腾讯美国有限责任公司 动态点云分区打包方法和装置
EP3759648A4 (fr) * 2019-01-09 2022-01-12 Tencent America LLC Procédé et appareil de regroupement de partition en nuage de points dynamique
US11825128B2 (en) 2019-01-09 2023-11-21 Tencent America LLC Method and apparatus for dynamic point cloud partition packing
CN113748681B (zh) * 2019-01-09 2024-04-26 腾讯美国有限责任公司 动态点云分区打包方法、装置和电子设备
US12316883B2 (en) 2019-01-09 2025-05-27 Tencent America LLC Techniques and apparatus for automatic ROI chunking for content-aware point cloud compression using HEVC tiles
CN113841416A (zh) * 2019-05-31 2021-12-24 倬咏技术拓展有限公司 互动沉浸洞穴网络

Similar Documents

Publication Publication Date Title
WO2018048288A1 (fr) Procédé et appareil de transmission et de réception de contenu de réalité virtuelle
WO2018182192A1 (fr) Procédé et appareil d'affichage d'image sur la base d'informations de mouvement d'utilisateur
WO2018070810A1 (fr) Procédé et appareil de traitement d'image de réalité virtuelle
WO2018048286A1 (fr) Procédé et appareil pour transmettre et reproduire un contenu dans un système de réalité virtuelle
WO2017142355A1 (fr) Procédé de transmission et de réception de métadonnées d'image omnidirectionnelle
WO2017217763A1 (fr) Dispositif et procédé de traitement d'image
KR102503342B1 (ko) 스테레오 스코픽 비디오 콘텐츠를 전송하기 위한 방법 및 장치
WO2018217057A1 (fr) Procédé de traitement de vidéo à 360 degrés et appareil associé
WO2018074850A1 (fr) Appareil de traitement d'images et procédé de traitement d'images associé
WO2020145668A1 (fr) Procédé de traitement et de transmission de contenu tridimensionnel
WO2016048014A1 (fr) Assemblage d'images pour video en trois dimensions
WO2018044073A1 (fr) Procédé de diffusion en continu d'image et dispositif électronique pour prendre en charge celui-ci
WO2019098728A1 (fr) Procédé et appareil de traitement d'images tridimensionnelles
WO2018097682A1 (fr) Appareil de traitement d'image et procédé de traitement d'image associé
WO2018043905A1 (fr) Procédés de transmission et de réception de vidéo à 360 degrés, dispositifs de transmission et de réception de vidéo à 360 degrés
WO2019198943A1 (fr) Appareil d'affichage vestimentaire, et procédé d'affichage d'images tridimensionnelles sur celui-ci
WO2017026705A1 (fr) Dispositif électronique pour générer une image tridimensionnelle sur 360 degrés, et procédé associé
WO2018131803A1 (fr) Procédé et appareil permettant de transmettre un contenu vidéo stéréoscopique
EP3494706A1 (fr) Procédé de diffusion en continu d'image et dispositif électronique pour prendre en charge celui-ci
WO2019125017A1 (fr) Appareil de mappage d'une image à un polyèdre en fonction de l'emplacement d'une région d'intérêt d'une image, et son procédé de traitement
WO2018221946A1 (fr) Procédé et dispositif de traitement de signal vidéo
WO2019083119A1 (fr) Procédé et dispositif de décodage d'image utilisant des paramètres de rotation dans un système de codage d'image pour une vidéo à 360 degrés
WO2019045393A1 (fr) Procédé et dispositif de traitement de signal vidéo
WO2024058369A1 (fr) Appareil de projection et son procédé de fonctionnement
WO2018174542A1 (fr) Procédé et dispositif de traitement de signal vidéo

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17891098

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 2017891098

Country of ref document: EP

Effective date: 20190812

WWG Wipo information: grant in national office

Ref document number: 201917031821

Country of ref document: IN