[go: up one dir, main page]

WO2018219202A1 - Procédé de présentation et d'encapsulation d'une image vidéo, et dispositif de présentation et d'une encapsulation d'image vidéo - Google Patents

Procédé de présentation et d'encapsulation d'une image vidéo, et dispositif de présentation et d'une encapsulation d'image vidéo Download PDF

Info

Publication number
WO2018219202A1
WO2018219202A1 PCT/CN2018/088197 CN2018088197W WO2018219202A1 WO 2018219202 A1 WO2018219202 A1 WO 2018219202A1 CN 2018088197 W CN2018088197 W CN 2018088197W WO 2018219202 A1 WO2018219202 A1 WO 2018219202A1
Authority
WO
WIPO (PCT)
Prior art keywords
video image
image
video
information
spherical
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/CN2018/088197
Other languages
English (en)
Chinese (zh)
Inventor
邸佩云
谢清鹏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huawei Technologies Co Ltd
Original Assignee
Huawei Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huawei Technologies Co Ltd filed Critical Huawei Technologies Co Ltd
Publication of WO2018219202A1 publication Critical patent/WO2018219202A1/fr
Priority to US16/689,517 priority Critical patent/US20200092531A1/en
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/816Monomedia components thereof involving special video data, e.g 3D video
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/172Processing image signals image signals comprising non-image signal components, e.g. headers or format information
    • H04N13/178Metadata, e.g. disparity information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/161Encoding, multiplexing or demultiplexing different image signal components
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/266Channel or content management, e.g. generation and management of keys and entitlement messages in a conditional access system, merging a VOD unicast channel into a multicast channel
    • H04N21/2662Controlling the complexity of the video stream, e.g. by scaling the resolution or bitrate of the video stream based on the client capabilities
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/85406Content authoring involving a specific file format, e.g. MP4 format

Definitions

  • the present application relates to the field of processing of video images, and more particularly to presentation of video images, packaging methods, and presentation of video images, packaging devices.
  • VR Virtual Reality
  • the rise of Virtual Reality (VR) has brought new visual experiences to people, and it has also brought new technical challenges.
  • the VR video image is typically divided into a plurality of independent video images, and each video image is then encoded to obtain a code stream of different video images. Since different video images may contain different image information, how to present a video image is a problem that needs to be solved.
  • the present application provides a video image presentation method, a packaging method, and a video image presentation and packaging device to improve display effects.
  • a method for presenting a video image comprising: acquiring a code stream of a first video image; parsing the code stream, determining the first video image and the first of the first video image Information, the first information is used to indicate whether the first video image is presented as a continuous area; and the first video image is presented according to the first information.
  • the first video image may be part of the original complete video image, or the first video image is a sub-video image obtained by dividing the original complete video image, and the sub-video image may also be directly It is called a sub image.
  • the video image when a video image is a continuous area in the finally displayed image, the video image can be directly presented.
  • the video image is not a continuous area in the finally displayed image, the video image may be spliced with other video images before being displayed.
  • the presenting the first video image according to the first information comprising: when the first information indicates the first video image as a The first video image is presented when the continuous area is presented.
  • the first video image is presented as a continuous region, the first video image is ultimately mapped onto the spherical surface for continuous image content.
  • the first video image When it is determined that the first video image can be presented as a continuous area, the first video image is presented, and continuous image content can be displayed, and the display effect is better.
  • the method includes: when the first information indicates that the first video image is not presented as a continuous area, the first video image and the second video image are spliced according to a positional relationship at the time of presentation.
  • first video image cannot be rendered as a continuous area
  • first video image is mapped to a spherical display
  • discontinuous image content may appear on the spherical surface
  • the second video image adjacent to the first video image content and the first video image need to be spliced according to the positional relationship at the time of presentation, and then displayed to ensure that the display is Continuous images for improved display.
  • a method for presenting a video image comprising: acquiring a code stream of a first video image; parsing the code stream, determining the first video image and a second of the first video image Information, the second information is used to indicate an image type of the first video image, and the image type of the first video image includes a spherical image, a two-dimensional planar image that has not undergone the first operation processing, and is processed through the first operation. a second two-dimensional planar image, wherein the first operation is at least one of segmentation, sampling, flipping, rotating, mirroring, and stitching; and the first video image is presented according to the second information.
  • the first video image may be part of the original complete video image, or the first video image is a sub-video image obtained by dividing the original complete video image, and the sub-video image may also be directly It is called a sub image.
  • the video image and the image type of the video image can be obtained from the code stream of a certain video image, and the subsequent operations can be initialized according to the image type of the video image, thereby reducing the time when the video image is presented. Delay, improve display efficiency.
  • the image type of the video image can be acquired while parsing a certain video image, and the operation type of the video image can be determined according to the image type of the video image earlier, and then these operations can be processed. Initialization is performed without having to parse the code stream of the full video as in the prior art, and these operations can be started, which can reduce the delay of presenting the video image and improve the display efficiency.
  • the presenting the first video image according to the second information including: when the second information indicates that the first video image is a spherical surface In the case of an image, the first video image is presented in a spherical display.
  • the presenting the first video image according to the second information including: when the second information indicates that the first video image is When the two-dimensional planar image that has not been processed by the first operation is described, the first video image is mapped to a spherical image; and the spherical image is presented in a spherical display manner.
  • the first video image When the first video image is a first type of two-dimensional planar image, the first video image needs to be mapped to a spherical image before being displayed on the spherical surface; otherwise, if the image type of the first video image is not known, the image is directly presented. A display error may occur during the first video image. Therefore, the image type of the first video image can be determined by the second information, thereby correctly displaying the first video image.
  • the presenting the first video image according to the second information including: when the second information indicates that the first video image is In the case of the two-dimensional planar image processed by the first operation, performing a second operation on the first video image to obtain a first video image after the second operation processing, where the second operation is the first An inverse operation of the operation; mapping the first video image after the second operation processing to a spherical image; and presenting the spherical image in a spherical display manner.
  • the second video image is first subjected to the second operation, and then the second video image after the second operation is mapped to the spherical image, and then the spherical image is Display, otherwise, a display error may also occur if the first video image is directly mapped to a spherical image and the spherical image is presented in a spherical display. Therefore, the image type of the first video image can be determined by the second information, thereby correctly displaying the first video image.
  • a third aspect provides a method for encapsulating a video image, the method comprising: determining first information of the first video image, where the first information is used to indicate whether the first video image is the first a continuous area in the image to be encoded corresponding to the video image; encoding the first video image and the first information to obtain a code stream of the first video image; encapsulating the code stream to obtain the first The image track of a video image.
  • the first video image may be part of the original complete video image, or the first video image is a sub-video image obtained by dividing the original complete video image, and the sub-video image may also be directly It is called a sub image.
  • the video image when the video image is a continuous area in the final displayed image, the video image can be directly presented; and when the video image is not a continuous area in the final displayed image, the video image can be compared with other videos. The image is spliced and then displayed.
  • a method for encoding a video image comprising: determining second information of a first video image, the second information being used to indicate an image type of the first video image, the image type
  • the method includes a spherical image, a two-dimensional planar image that has not undergone the first operation processing, and a two-dimensional planar image that has undergone the first operation processing, wherein the first operation is at least one of segmentation, sampling, flipping, rotating, mirroring, and splicing.
  • Encoding the first video image and the second information to obtain a code stream of the first video image; encapsulating the code stream to obtain an image track of the first video image.
  • the first video image may be part of the original complete video image, or the first video image is a sub-video image obtained by dividing the original complete video image, and the sub-video image may also be directly It is called a sub image.
  • the video image and the image type of the video image can be obtained from the code stream of a certain video image when the image is presented, thereby being able to
  • the image type of the video image is initialized in advance for subsequent operations, which can reduce the delay of presenting the video image and improve display efficiency.
  • the video rendering device can acquire the image type of the video image while parsing a certain video image, and can determine, according to the image type of the video image, which operation processing is to be performed on the video image, and then These operations can be initialized first, without having to parse the code stream of the full video as in the prior art, and these operations can be started, which can reduce the delay of presenting the video image and improve the display efficiency.
  • a presentation apparatus for a video image comprising means for performing the method of the first aspect or various implementations thereof.
  • a video image presentation apparatus comprising means for performing the method of the second aspect or various implementations thereof.
  • a package apparatus for a video image comprising means for performing the method of the third aspect or various implementations thereof.
  • a package apparatus for a video image comprising means for performing the method of the fourth aspect or various implementations thereof.
  • a ninth aspect a video image presenting apparatus is provided, the apparatus comprising: a storage medium, and a central processing unit, wherein the storage medium stores a computer executable program, the central processor is connected to the storage medium, and The computer executable program is executed to implement the method of the first aspect or various implementations thereof.
  • a video image presenting apparatus comprising: a storage medium, and a central processing unit, wherein the storage medium stores a computer executable program, the central processor is connected to the storage medium, and The computer executable program is executed to implement the method of the second aspect or various implementations thereof.
  • a video image encapsulating apparatus comprising: a storage medium, and a central processing unit, wherein the storage medium stores a computer executable program, and the central processor is connected to the storage medium, The computer executable program is executed to implement the method of the third aspect or various implementations thereof.
  • a video image encapsulating apparatus comprising: a storage medium, and a central processing unit, wherein the storage medium stores a computer executable program, and the central processing unit is connected to the storage medium, And executing the computer executable program to implement the method of the fourth aspect or various implementations thereof.
  • the storage medium may be a nonvolatile storage medium.
  • a thirteenth aspect a computer readable medium storing program code for device execution, the program code comprising instructions for performing the method of the first aspect or various implementations thereof .
  • a computer readable medium storing program code for device execution, the program code comprising instructions for performing the method of the second aspect or various implementations thereof .
  • a computer readable medium storing program code for device execution, the program code comprising instructions for performing the method of the third aspect or various implementations thereof .
  • a computer readable medium storing program code for device execution, the program code comprising instructions for performing the method of the fourth aspect or various implementations thereof .
  • FIG. 1 is a schematic flowchart of a method for presenting a video image according to an embodiment of the present application.
  • FIG. 2 is a schematic diagram of a spherical image and a two-dimensional planar image.
  • Figure 3 is a schematic illustration of a video image.
  • FIG. 4 is a schematic diagram of the position of a video image in a two-dimensional planar image.
  • Figure 5 is a schematic illustration of the video image at a spherical position.
  • FIG. 6 is a schematic flowchart of a method for presenting a video image according to an embodiment of the present application.
  • FIG. 7 is a schematic flowchart of a method for encapsulating a video image according to an embodiment of the present application.
  • FIG. 8 is a schematic flowchart of a method for encapsulating a video image according to an embodiment of the present application.
  • FIG. 9 is a schematic flow chart of generating a code stream of a sub-image.
  • FIG. 10 is a schematic flow chart of analyzing a code stream of a sub-image.
  • FIG. 11 is a schematic block diagram of a video image presentation apparatus according to an embodiment of the present application.
  • FIG. 12 is a schematic block diagram of a video image presentation apparatus according to an embodiment of the present application.
  • FIG. 13 is a schematic block diagram of a video image encapsulation apparatus according to an embodiment of the present application.
  • FIG. 14 is a schematic block diagram of a video image encapsulation apparatus according to an embodiment of the present application.
  • FIG. 15 is a schematic block diagram of a codec apparatus according to an embodiment of the present application.
  • FIG. 16 is a schematic diagram of a codec apparatus according to an embodiment of the present application.
  • FIG. 17 is a schematic block diagram of a video codec system according to an embodiment of the present application.
  • FIG. 1 is a schematic flowchart of a method for presenting a video image according to an embodiment of the present application.
  • the method 100 of Figure 1 includes:
  • the first video image may be part of an original complete video image (which may also be referred to as an original video image, an original image, or an original image), or the first video image is a complete video.
  • a sub-video image obtained after the image is divided, and the sub-video image may also be directly referred to as a sub-image.
  • the first video image is divided into an original video image, which is a sub-image of the original image.
  • the original image may be a spherical image as shown in FIG. 2, which may be an image having a 360 degree angle of view.
  • the original image may also be a first type of two-dimensional planar image shown in FIG. 2.
  • the first type of two-dimensional planar image is obtained by mapping a spherical image to a plane, and the first type of two-dimensional planar image may be a longitude and longitude image. It may also be a planar image obtained by expanding the six faces of the hexahedron after the spherical image is mapped to the hexahedron.
  • the original image may also be a second type of two-dimensional image shown in FIG. 2, and the second type of two-dimensional image performs certain operations on the first type of two-dimensional image (for example, segmentation, sampling, flipping, and rotating).
  • the original image when the original image is a second type of two-dimensional image, the original image may be divided into nine sub-images (four dashed lines divide the two-dimensional planar image into nine regions, one for each region) Sub-image), the resulting sub-images A, B, C, D, E, F, G, H, and I.
  • the first video image may be any one of the nine sub-images.
  • the video image is conveniently encoded by dividing the original image into a plurality of sub-images.
  • the first information may be used to indicate whether the first video image is presented as a continuous area.
  • the code stream of the first video image may be a code stream generated by the encoding end when encoding the first video image, and the first video image may be acquired not only by parsing the code stream, but also the first video image may be acquired. a message.
  • the first information of the sub-image A is specifically used to indicate that the sub-image A is presented as a continuous area when presented, because the sub-image A Contains an image of a continuous region of the middle region of the first type of two-dimensional image.
  • the first information of the first video image also indicates that the first video image can be presented as a continuous region.
  • the first information of the sub-image is specifically used to indicate that the sub-image is not a continuous region in the final displayed image, because the sub-image G includes An image of a middle portion of the first type of two-dimensional image and a two-part region of the top portion, and the images of the two partial regions are not adjacent, and therefore, when the sub-image is the sub-image G, in the final displayed image A discontinuous area.
  • the first information of the sub-image is also a continuous region indicating that the sub-image is not the final displayed image.
  • a video rendering device which may also be a decoding end device, a decoder, or a device having a decoding function.
  • the video image when a video image is a continuous area in the finally displayed image, the video image can be directly presented.
  • the video image is not a continuous area in the finally displayed image, the video image may be spliced with other video images before being displayed.
  • the first video image can be presented as a continuous area
  • the image content of the first video image can be directly displayed.
  • the image content of the first video image is presented again, which can ensure that the displayed image content is continuous, and a certain display effect can be ensured.
  • the first video image when the first video image can be presented as a continuous area, the first video image is finally mapped onto the spherical surface to display continuous image content; and when the first video image is not presented as a continuous area, if still If the first video image is directly mapped to the spherical display, then there may be discontinuous image content displayed on the sphere, which may affect the visual experience.
  • the first video image is directly presented to the spherical display, there may be a case where the image content is discontinuous on the spherical surface (for example, two completely unrelated image contents may be displayed).
  • the second video image may be acquired according to the code stream of the second video image, wherein the second video image is a video image adjacent to at least a portion (image content) of the first video image at the time of presentation Then, the first video image and the second video image are spliced according to the positional relationship at the time of presentation and then presented.
  • the positional relationship of the first video image and the second video image at the time of presentation may be directly obtained by parsing from the code stream of the entire video, or may be separately acquired according to the code streams of the first video image and the second video image.
  • the position information of the first video image and the second video image is determined.
  • the second video image adjacent to the first video image content and the first video image need to be spliced according to the positional relationship at the time of presentation, and then displayed to ensure that the display is Continuous images for improved display.
  • the positions of the sub-image G on the first-type two-dimensional planar image and the spherical image are as shown in FIGS. 4 and 5, respectively.
  • the positions of the sub-image G on the first-type two-dimensional plane image in FIG. 4 are respectively on the left side of the top and the lower left corner of the middle (shaded portion area in FIG. 4).
  • the position of the sub-image G on the spherical image in Fig. 5 is the shaded area shown by 1 and 2. It can be seen that the sub-image G is two discontinuous regions on the first type of two-dimensional plane image and the spherical image. Therefore, if the sub-image is directly presented, then two discontinuous image contents will be displayed. The display is not good.
  • the first information may be implemented in various manners.
  • the first information may be described in a new syntax extended in the Track Group Type Box of the first video image.
  • the syntax in the SubPicture Composition Box may be used to describe First information.
  • the value of content_continuity may be used to indicate whether the first video image is presented as a continuous area, and the specific syntax is as follows:
  • SubPictureCompositionBox extends TrackGroupTypeBox(‘spco’)
  • the first video image is not presented as a continuous area.
  • content_continuity indicates whether the first video image is presented as a continuous region.
  • the content_continuity may also take other values to respectively indicate whether the first video image is presented as a continuous region. .
  • FIG. 6 is a schematic flowchart of a method for decoding a video image according to an embodiment of the present application.
  • the method 600 of Figure 6 includes:
  • the first video image may be part of an original complete video image (which may also be referred to as an original video image, an original image, or an original image), or the first video image is a complete video.
  • a sub-video image obtained after the image is divided, and the sub-video image may also be directly referred to as a sub-image.
  • the original video image When the first video image is a sub-image obtained by dividing the original video image, the original video image may be a spherical image, a first type of two-dimensional planar image, or a second type of planar two-dimensional image as shown in FIG. 2.
  • the spherical image may be an image having a 360 degree angle of view
  • the first type of two-dimensional planar image may be a planar image obtained by mapping the spherical image to a plane.
  • the first type of two-dimensional planar image may be a warp and a weft image, or may be A planar image obtained by mapping a spherical image to a hexahedron and then expanding the six faces of the hexahedron.
  • the second type of two-dimensional image may be a planar image obtained after performing certain operations (for example, segmentation, sampling, flipping, rotating, mirroring, splicing) on the first type of two-dimensional image.
  • certain operations for example, segmentation, sampling, flipping, rotating, mirroring, splicing
  • FIG. 2 The top and bottom regions of the first type of two-dimensional image are compressed and spliced together, and arranged below the middle region to obtain a second type of two-dimensional planar image.
  • the original video image of the first video image when the original video image of the first video image is a second type of two-dimensional image, the original video image may be divided into nine sub-images, and the first video image may be the nine sub-images. Any of the sub-images in .
  • Parsing a code stream determining a first video image and second information of the first video image, where the second information is used to indicate an image type of the first video image, and the image type of the first video image includes a spherical image, A two-dimensional planar image processed by the operation and a two-dimensional planar image processed by the first operation.
  • the first operation may be at least one of segmentation, sampling, flipping, rotating, mirroring, and splicing.
  • the image type of the first video image is the same as the image type of the original video image of the first video image, for example, if the original video image is a first type of two-dimensional planar image, the first obtained by dividing the original video image The image type of the video image is also the first type of two-dimensional planar image.
  • the above two-dimensional planar image lesson that has not been processed by the first operation may be the first type of two-dimensional planar image in FIG. 2, and the two-dimensional planar image is obtained by directly mapping the spherical image to the plane, and is mapped to the plane. The first operation was not performed afterwards.
  • the two-dimensional planar image processed by the first operation may be the second type of two-dimensional planar image in FIG. 2, and the two-dimensional planar image is directly mapped to the plane by the spherical image to obtain the first type of two-dimensional planar image, and then Then, the planar image obtained by dividing, sampling, flipping, splicing, etc. of the first type of two-dimensional planar image is performed.
  • first operation may be referred to as packing
  • second operation may be referred to as reverse packing
  • the video image and the image type of the video image can be obtained from the code stream of a certain video image, and the subsequent operations can be initialized according to the image type of the video image, thereby reducing the time when the video image is presented. Delay, improve display efficiency.
  • the image type of the video image can be acquired while parsing a certain video image, and the operation type of the video image can be determined according to the image type of the video image earlier, and then these operations can be processed. Initialization is performed without having to parse the code stream of the full video as in the prior art, and these operations can be started, which can reduce the delay of presenting the video image and improve the display efficiency.
  • a video rendering device which may also be a decoding end device, a decoder, or a device having a decoding function.
  • the process of presenting the first video image is also different, and specifically includes the following three cases:
  • the first video image is a spherical image
  • the first video image When the first video image is a spherical image, the first video image may be directly presented on the spherical surface, that is, the first video image may be presented in a spherical display manner. Specifically, when the first video image is a spherical image, the first video image is a part (or all) of the original video image (the original video image is also a spherical image), and is directly based on the first video image on the spherical surface. The position information may be displayed by presenting the first video image to a corresponding position on the spherical surface.
  • the first video image is a two-dimensional planar image that has not been processed by the first operation
  • the first video image may be a first type of two-dimensional planar image as shown in FIG. 2.
  • the first video image is first mapped to a spherical image, and then the spherical image is presented in a spherical display.
  • the first video image When the first video image is a first type of two-dimensional planar image, the first video image needs to be mapped to a spherical image before being displayed on the spherical surface; otherwise, if the image type of the first video image is not known, the image is directly presented. A display error may occur during the first video image. Therefore, the image type of the first video image can be determined by the second information, thereby correctly displaying the first video image.
  • the first video image is a two-dimensional planar image processed by the first operation
  • the first video image may be a second type of two-dimensional planar image as shown in FIG. 2. If the first video sub-image is to be presented, the first video image is subjected to a second operation process to obtain a second operation-processed first video image, wherein the second operation is an inverse operation of the first operation. (Or called reverse operation, reverse operation), next, the first video image after the second operation is mapped to the spherical surface to obtain a spherical image, and then the spherical image is presented in a spherical display manner.
  • a second operation process to obtain a second operation-processed first video image
  • reverse operation reverse operation
  • the second video image is first subjected to the second operation, and then the second video image after the second operation is mapped to the spherical image, and then the spherical image is Display, otherwise, a display error may also occur if the first video image is directly mapped to a spherical image and the spherical image is presented in a spherical display. Therefore, the image type of the first video image can be determined by the second information, thereby correctly displaying the first video image.
  • the second operation is also flipping, and the second image can restore the video image to the state before the first operation, that is, the second operation is the restoration of the first operation. Operation, by the second operation, the image after the first operation processing can be restored to the state before the first operation processing.
  • the image of the top region and the bottom region may be enlarged, and the images of the enlarged top region and the bottom region may be respectively moved to the intermediate region.
  • the first type of two-dimensional planar image shown in Figure 2 is finally obtained.
  • the foregoing second information may be implemented in various manners.
  • the second information may be described in a new syntax extended in the Track Group Type Box of the first video image.
  • the syntax in the SubPicture Composition Box may be used to describe Second message.
  • the value of the fullpictureType may be used to indicate the image type of the video image, and the specific syntax is as follows:
  • SubPictureCompositionBox extends TrackGroupTypeBox(‘spco’)
  • the first video image is a spherical image
  • the first video image is a two-dimensional planar image, and the first video image does not undergo a first operation
  • the first video image is a two-dimensional planar image, and the first video image undergoes a first operation.
  • fullpictureType uses different values to represent the image type of the first video image.
  • fullpictureTyp may also take other values to represent the image type of the first video image.
  • the second information may further include two sub-information, a first sub-information and a second sub-information, where the first sub-information is used to indicate whether the first video image is a spherical image or a two-dimensional planar image, when the first sub- When the information indicates that the video image is a two-dimensional planar image, the second sub-information indicates whether the video image has undergone the first operation.
  • the second information when the first video image is a spherical image, the second information includes only the first sub-information, and when the first video image is a two-dimensional planar image, the second information includes the first sub-information.
  • the two sub-information wherein the first sub-information indicates that the first video image is a two-dimensional planar image, and the second sub-information indicates whether the first video image has undergone the first operation.
  • the value of the fullpictureType may also be used to indicate the image type of the first video image, and the specific syntax is as follows:
  • SubPictureCompositionBox extends TrackGroupTypeBox(‘spco’)
  • the first video image is a spherical image
  • the first video image is a two-dimensional planar image.
  • fullpictureType uses different values to represent the image type of the first video image.
  • fullpictureTyp may also take other values to represent the image type of the first video image.
  • the second sub-information in the second information may also be represented by a statement similar to the first sub-information.
  • the value of the packing may be used to indicate the image type of the first video image.
  • the specific syntax is as follows:
  • SubPictureCompositionBox extends TrackGroupTypeBox(‘spco’)
  • packing uses different values to represent the image type of the first video image (whether the video image passes the first operation). In fact, packing can also take other values to represent the first video image. The type of image.
  • the method 100 or the method 600 further includes: determining third information of the first video image according to the code stream of the first video image, where the third information is used to indicate whether the first video image is a full image; The three information presents the first video image.
  • the full image image herein may be a complete image to be displayed, and the first video image may be either all of the complete image to be displayed or only a part of the complete image to be displayed.
  • the decoding end or the device that presents the video may determine that the first video image includes the entire image instead of the partial image after parsing the third information. It is not necessary to render the image content at any position in the entire image by means of other video images; and when the third information indicates that the first video image is the entire image, the decoding end or the device that presents the video after parsing the third information.
  • the position information of the first video image and the resolution information need to be parsed to determine the position of the first video image in the entire image, and then the first video image is presented.
  • the value of the fullpicture may also be used to indicate whether the first video image is a full image, and the specific syntax is as follows:
  • SubPictureCompositionBox extends TrackGroupTypeBox(‘spco’)
  • the first video image is a full image
  • the first video image is a partial image of the full image.
  • the fullpicture uses different values to indicate whether the first video image is a full image.
  • the fullpicture may also take other values to indicate whether the first video image is a full image.
  • At least one of the first information, the second information, and the third information may be obtained by parsing the code stream of the first video sub-image, and may be based on the three types when parsing and parsing the first video image.
  • One or more of the information to present the first video image may be obtained by parsing the code stream of the first video sub-image, and may be based on the three types when parsing and parsing the first video image.
  • the scheme of presenting the first video image according to one or more of the first information, the second information, and the third information is within the protection scope of the present application.
  • the parsing the first video image according to the first information and the second information comprises: determining, according to the first information, whether the first video image is presented as a continuous region; determining an image type of the first video image according to the second information And presenting the image content of the first video image according to whether the first video image is presented as one continuous region and the image type of the first video image.
  • the image content of the first video image is presented, including: presenting the first video image as a continuous area And in the case where the first video image is a spherical image, the first video image (direct) is presented in a spherical display.
  • the image content of the first video image is presented, including: presenting the first video image as a continuous area And the first video image is a two-dimensional planar image that has not undergone the first operation processing, and the first video image is mapped to a spherical image; the spherical image is presented in a spherical display manner.
  • the image content of the first video image is presented, including: presenting the first video image as a continuous area And the first video image is a two-dimensional planar image processed by the first operation, performing a second operation on the first video image to obtain a first video image after the second operation processing, where the second operation is The inverse operation or the reverse operation of the first operation; mapping the first video image after the second operation processing to a spherical image; and presenting the spherical image in a spherical display manner.
  • the location information of the first video image in the entire video image may also be used.
  • the location information of the first video image can be implemented in various manners.
  • the location information of the first video image can be described in a new syntax extended in the Track Group Type Box of the first video image.
  • SubPicture Composition can be used.
  • the grammar in the Box describes the location information of the first video image.
  • the syntax for describing the location information of the first video image is as follows:
  • SubPictureCompositionBox extends TrackGroupTypeBox(‘spco’)
  • track_x represents the horizontal position of the upper left corner of the first video image in the entire video image (or referred to as the original video image), taking a natural number, the range [0, composition_width-1];
  • Track_y represents the vertical position of the upper left corner of the first video image in the entire video image, taking a natural number, the range [0, composition_height-1];
  • Track_width indicates the width of the first video image, which is an integer, the range [1, composition_width–track_x];
  • Track_height indicates the height of the first video image, which is an integer, the range [1, composition_height–track_y].
  • Composition_width represents the width of the first video image
  • Composition_height represents the height of the first video image.
  • FIG. 7 and FIG. 8 The method for presenting the video image of the embodiment of the present application is described in detail with reference to FIG. 1 to FIG. 6 , and the method for encapsulating the video image of the embodiment of the present application is described below with reference to FIG. 7 and FIG. 8 .
  • the packaging methods of the video images shown in FIGS. 7 and 8 correspond to the methods 100 and 600 of the above, respectively.
  • the repeated description is appropriately omitted below.
  • FIG. 7 is a schematic flowchart of a method for encapsulating a video image according to an embodiment of the present application.
  • the method 700 of Figure 7 includes:
  • the first video image may be part of the original complete video image, or the first video image is a sub-video image obtained by dividing the original complete video image, and the sub-video image may also be directly referred to as a sub-video image. image.
  • the video image when the video image is a continuous area in the final displayed image, the video image can be directly presented; and when the video image is not a continuous area in the final displayed image, the video image can be compared with other videos. The image is spliced and then displayed.
  • FIG. 8 is a schematic flowchart of a method for encapsulating a video image according to an embodiment of the present application.
  • the method 800 of Figure 8 includes:
  • the image type includes a spherical image, a two-dimensional planar image that has not been processed by the first operation, and a two-dimensional planar image that is processed by the first operation, wherein the first operation is segmentation, sampling, flipping, rotating, mirroring, and splicing. At least one type;
  • the first video image may be part of the original complete video image, or the first video image is a sub-video image obtained by dividing the original complete video image, and the sub-video image may also be directly referred to as a sub-video image. image.
  • the video image and the image type of the video image can be obtained from the code stream of a certain video image when the image is presented, thereby being able to
  • the image type of the video image is initialized in advance for subsequent operations, which can reduce the delay of presenting the video image and improve display efficiency.
  • the video rendering device can acquire the image type of the video image while parsing a certain video image, and can determine, according to the image type of the video image, which operation processing is to be performed on the video image, and then These operations can be initialized first, without having to parse the code stream of the full video as in the prior art, and these operations can be started, which can reduce the delay of presenting the video image and improve the display efficiency.
  • FIG. 9 is a schematic flow chart of generating a code stream of a sub-image.
  • the sub-image dividing module divides the input entire image into a plurality of sub-images, and determines metadata of each sub-image, and then outputs the sub-images;
  • the encoder encodes the input sub-images to generate a video.
  • the code stream encapsulation module encapsulates the input video bare code stream and metadata into the sub-picture code stream.
  • the video bare code stream data is a code stream conforming to the ITU-T H.264 or ITU-T H.265 specifications;
  • the metadata of the sub-image may include the first information, the second information, and the third information in the foregoing.
  • At least one of the metadata may be obtained from the sub-image partitioning module or from the divided preset conditions.
  • FIG. 10 is a schematic flow chart of analyzing a code stream of a sub-image.
  • the code stream decapsulation module obtains the code stream data of the sub image, and parses the metadata of the video and the video bare code stream data.
  • the image information of the sub-image can be obtained from the metadata of the video, and then the sub-image is presented according to the image information of the sub-image and the sub-image parsed in the video bare stream data of the sub-image.
  • the video image presentation method and the packaging method of the embodiments of the present application are described above with reference to FIG. 1 to FIG. 10 .
  • the video image presentation device and the packaging device of the embodiments of the present application are described below with reference to FIG. 11 to FIG.
  • the presentation device in FIG. 11 to FIG. 14 can implement the decoding method of the video image in FIG. 1 to FIG. 10
  • the packaging device can implement the encoding method of the video image in FIG. 1 to FIG. 10 . description.
  • FIG. 11 is a schematic block diagram of a video image presentation apparatus according to an embodiment of the present application.
  • the device 1100 includes:
  • the obtaining module 1110 is configured to acquire a code stream of the first video image.
  • a parsing module 1120 configured to parse the code stream, and determine first information of the first video image and the first video image, where the first information is used to indicate whether the first video image is a continuous area Present
  • the presentation module 1130 is configured to present the first video image according to the first information.
  • the presentation module 1130 is specifically configured to: when the first information indicates that the first video image is presented as one continuous area, present the first video image.
  • the presentation module 1130 is specifically configured to: when the first information indicates the first video image When not presented as a continuous area, the first video image and the second video image are stitched together according to the positional relationship at the time of presentation.
  • FIG. 12 is a schematic block diagram of a video image presentation apparatus according to an embodiment of the present application.
  • the device 1200 includes:
  • the obtaining module 1210 is configured to acquire a code stream of the first video image.
  • a parsing module 1220 configured to parse the code stream, determine the first video image and second information of the first video image, where the second information is used to indicate an image type of the first video image,
  • the first image type includes a spherical image, a two-dimensional planar image that has not undergone the first operation processing, and a two-dimensional planar image that has undergone the first operation processing, wherein the first operation is segmentation, sampling, flipping, rotating, mirroring At least one of splicing;
  • the presentation module 1230 is configured to present the first video image according to the second information.
  • the presentation module 1230 is specifically configured to: when the second information indicates that the first video image is a spherical image, present the first video image in a spherical display manner.
  • the presentation module 1230 is specifically configured to: when the second information indicates that the first video image is the two-dimensional planar image that has not undergone the first operation processing, The sub-image is mapped to a spherical image; the spherical image is presented in a spherical display.
  • the presentation module 1230 is specifically configured to: when the second information indicates that the first video image is the two-dimensional planar image after the first operation processing, Performing a second operation on the first video image to obtain a first video image after the second operation processing, where the second operation is an inverse operation of the first operation; and the first video processed by the second operation
  • the image is mapped to a spherical image; the spherical image is presented in a spherical display.
  • FIG. 13 is a schematic block diagram of a video image encapsulation apparatus according to an embodiment of the present application.
  • the device 1300 includes:
  • a determining module 1310 configured to determine first information of the first video image, where the first information is used to indicate whether the first video image is a continuous region in an image to be encoded corresponding to the first video image ;
  • the encoding module 1320 is configured to encode the first video image and the first information to obtain a code stream of the first video image
  • the encapsulating module 1330 is configured to encapsulate the code stream to obtain an image track of the first video image.
  • FIG. 14 is a schematic block diagram of a video image encapsulation apparatus according to an embodiment of the present application.
  • the device 1400 includes:
  • a determining module 1410 configured to determine second information of the first video image, where the second information is used to indicate an image type of the first video image, the image type includes a spherical image, and the second image is not processed by the first operation a planar image and a two-dimensional planar image processed by the first operation, wherein the first operation is at least one of segmentation, sampling, flipping, rotating, mirroring, and splicing;
  • the encoding module 1420 is configured to encode the first video image and the second information to obtain a code stream of the first video image
  • the encapsulating module 1430 is configured to encapsulate the code stream to obtain an image track of the first video image.
  • the presentation method and the encapsulation method of the video image in the present application may be performed by a system composed of a codec device or a codec device.
  • the presentation device and the encapsulation device of the video image above may also be specifically a codec device or Codec system.
  • a codec system composed of a codec device and a codec device will be described in detail below with reference to FIGS. 15 to 17. It should be understood that the codec device and the codec system in FIGS. 15 to 17 are capable of performing the above-described video image presentation method and video image encapsulation method.
  • FIG. 15 and 16 illustrate a codec device 50 of an embodiment of the present application, which may be a mobile terminal or user equipment of a wireless communication system. It should be understood that embodiments of the present application can be implemented in any electronic device or device that may require encoding and/or decoding of video images.
  • the codec device 50 may include a housing 30 for incorporating and protecting the device, a display 32 (which may specifically be a liquid crystal display), a keypad 34.
  • Codec device 50 may include a microphone 36 or any suitable audio input, which may be a digital or analog signal input.
  • the codec device 50 may also include an audio output device, which in the embodiment of the present application may be any of the following: an earphone 38, a speaker, or an analog audio or digital audio output connection.
  • Codec device 50 may also include battery 40, and in other embodiments of the present application, the device may be powered by any suitable mobile energy device, such as a solar cell, fuel cell, or clock mechanism generator.
  • the device may also include an infrared port 42 for short-range line of sight communication with other devices.
  • codec device 50 may also include any suitable short range communication solution, such as a Bluetooth wireless connection or a USB/FireWire wired connection.
  • Codec device 50 may include a controller 56 or processor for controlling codec device 50.
  • the controller 56 can be coupled to a memory 58, which in the embodiments of the present application can store data in the form of data and audio in the form of images, and/or can also store instructions for execution on the controller 56.
  • the controller 56 can also be coupled to a codec 54 suitable for implementing encoding and decoding of audio and/or video data or assisted encoding and decoding by the controller 56.
  • the codec device 50 may also include a card reader 48 and a smart card 46 for providing user information and for providing authentication information for authenticating and authorizing users on the network, such as a Universal Integrated Circuit Card (UICC) and a UICC. Reader.
  • a card reader 48 for providing user information and for providing authentication information for authenticating and authorizing users on the network, such as a Universal Integrated Circuit Card (UICC) and a UICC. Reader.
  • UICC Universal Integrated Circuit Card
  • UICC Universal Integrated Circuit Card
  • the codec device 50 may also include a radio interface circuit 52 coupled to the controller and adapted to generate, for example, a wireless communication signal for communicating with a cellular communication network, a wireless communication system, or a wireless local area network.
  • the codec device 50 may also include an antenna 44 coupled to the radio interface circuit 52 for transmitting radio frequency signals generated at the radio interface circuit 52 to other device(s) and for receiving radio frequency signals from other device(s) .
  • codec device 50 includes a camera capable of recording or detecting a single frame, and codec 54 or controller receives these single frames and processes them. In some embodiments of the present application, codec device 50 may receive video image data to be processed from another device prior to transmission and/or storage. In some embodiments of the present application, codec device 50 may receive images for encoding/decoding over a wireless or wired connection.
  • FIG. 17 is a schematic block diagram of a video codec system 10 according to an embodiment of the present application.
  • the video codec system 10 includes a source device 12 and a destination device 14.
  • Source device 12 produces encoded video data.
  • source device 12 may be referred to as a video encoding device or a video encoding device.
  • Destination device 14 can decode the encoded video data produced by source device 12.
  • destination device 14 may be referred to as a video decoding device or a video decoding device.
  • Source device 12 and destination device 14 may be examples of video codec devices or video codec devices.
  • Source device 12 and destination device 14 may include desktop computers, mobile computing devices, notebook (eg, laptop) computers, tablet computers, set top boxes, smart phones, and the like, televisions, cameras, display devices, digital media players, Video game console, onboard computer, or other similar device.
  • Channel 16 may include one or more media and/or devices capable of moving encoded video data from source device 12 to destination device 14.
  • channel 16 may include one or more communication media that enable source device 12 to transmit encoded video data directly to destination device 14 in real time.
  • source device 12 may modulate the encoded video data in accordance with a communication standard (eg, a wireless communication protocol) and may transmit the modulated video data to destination device 14.
  • the one or more communication media may include wireless and/or wired communication media, such as a radio frequency (RF) spectrum or one or more physical transmission lines.
  • RF radio frequency
  • the one or more communication media may form part of a packet-based network (eg, a local area network, a wide area network, or a global network (eg, the Internet)).
  • the one or more communication media may include routers, switches, base stations, or other devices that facilitate communication from source device 12 to destination device 14.
  • channel 16 can include a storage medium that stores encoded video data generated by source device 12.
  • destination device 14 can access the storage medium via disk access or card access.
  • the storage medium may include a variety of locally accessible data storage media, such as Blu-ray Disc, DVD, CD-ROM, flash memory, or other suitable digital storage medium for storing encoded video data.
  • channel 16 can include a file server or another intermediate storage device that stores encoded video data generated by source device 12.
  • destination device 14 may access the encoded video data stored at a file server or other intermediate storage device via streaming or download.
  • the file server may be a server type capable of storing encoded video data and transmitting the encoded video data to the destination device 14.
  • the file server can include a web server (eg, for a website), a file transfer protocol (FTP) server, a network attached storage (NAS) device, and a local disk drive.
  • FTP file transfer protocol
  • NAS network attached storage
  • Destination device 14 can access the encoded video data via a standard data connection (e.g., an internet connection).
  • a standard data connection e.g., an internet connection.
  • An instance type of a data connection includes a wireless channel (eg, a Wi-Fi connection), a wired connection (eg, DSL, cable modem, etc.), or both, suitable for accessing encoded video data stored on a file server. combination.
  • the transmission of the encoded video data from the file server may be streaming, downloading, or a combination of both.
  • the codec method of the present application is not limited to a wireless application scenario.
  • the codec method may be applied to video codec supporting multiple multimedia applications such as: aerial television broadcasting, cable television transmission, satellite television transmission, Streaming video transmission (e.g., via the Internet), encoding of video data stored on a data storage medium, decoding of video data stored on a data storage medium, or other application.
  • video codec system 10 may be configured to support one-way or two-way video transmission to support applications such as video streaming, video playback, video broadcasting, and/or video telephony.
  • source device 12 includes video source 18, video encoder 20, and output interface 22.
  • output interface 22 can include a modulator/demodulator (modem) and/or a transmitter.
  • Video source 18 may include a video capture device (eg, a video camera), a video archive containing previously captured video data, a video input interface to receive video data from a video content provider, and/or a computer for generating video data.
  • Video encoder 20 may encode video data from video source 18.
  • source device 12 transmits the encoded video data directly to destination device 14 via output interface 22.
  • the encoded video data may also be stored on a storage medium or file server for later access by the destination device 14 for decoding and/or playback.
  • destination device 14 includes an input interface 28, a video decoder 30, and a display device 32.
  • input interface 28 includes a receiver and/or a modem.
  • Input interface 28 can receive the encoded video data via channel 16.
  • Display device 32 may be integral with destination device 14 or may be external to destination device 14. In general, display device 32 displays the decoded video data.
  • Display device 32 may include a variety of display devices such as liquid crystal displays (LCDs), plasma displays, organic light emitting diode (OLED) displays, or other types of display devices.
  • LCDs liquid crystal displays
  • OLED organic light emitting diode
  • Video encoder 20 and video decoder 30 may operate in accordance with a video compression standard (eg, the High Efficiency Video Codec H.265 standard) and may conform to the HEVC Test Model (HM).
  • a video compression standard eg, the High Efficiency Video Codec H.265 standard
  • HM HEVC Test Model
  • a textual description of the H.265 standard is published on April 29, 2015, ITU-T.265(V3) (04/2015), available for download from http://handle.itu.int/11.1002/1000/12455 The entire contents of the document are incorporated herein by reference.
  • the disclosed systems, devices, and methods may be implemented in other manners.
  • the device embodiments described above are merely illustrative.
  • the division of the unit is only a logical function division.
  • there may be another division manner for example, multiple units or components may be combined or Can be integrated into another system, or some features can be ignored or not executed.
  • the mutual coupling or direct coupling or communication connection shown or discussed may be an indirect coupling or communication connection through some interface, device or unit, and may be in an electrical, mechanical or other form.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of the embodiment.
  • each functional unit in each embodiment of the present application may be integrated into one processing unit, or each unit may exist physically separately, or two or more units may be integrated into one unit.
  • the functions may be stored in a computer readable storage medium if implemented in the form of a software functional unit and sold or used as a standalone product.
  • the technical solution of the present application which is essential or contributes to the prior art, or a part of the technical solution, may be embodied in the form of a software product, which is stored in a storage medium, including
  • the instructions are used to cause a computer device (which may be a personal computer, server, or network device, etc.) to perform all or part of the steps of the methods described in various embodiments of the present application.
  • the foregoing storage medium includes: a U disk, a mobile hard disk, a read-only memory (ROM), a random access memory (RAM), a magnetic disk, or an optical disk, and the like, which can store program codes. .

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Library & Information Science (AREA)
  • Databases & Information Systems (AREA)
  • Computer Security & Cryptography (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

L'invention concerne un procédé de présentation et d'encapsulation d'une image vidéo, et un dispositif de présentation et d'encapsulation d'une image vidéo. Le procédé de présentation d'une image vidéo consiste : à acquérir un flux de codes d'une première image vidéo; à analyser le flux de codes, et à déterminer la première image vidéo et des premières informations concernant la première image vidéo, les premières informations servant à indiquer si la première image vidéo est présentée sous la forme d'une région consécutive; et à présenter, selon la première image vidéo et les premières informations, la première image vidéo. La présente invention peut mieux présenter la première image vidéo en fonction des premières informations concernant la première image vidéo.
PCT/CN2018/088197 2017-05-27 2018-05-24 Procédé de présentation et d'encapsulation d'une image vidéo, et dispositif de présentation et d'une encapsulation d'image vidéo Ceased WO2018219202A1 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/689,517 US20200092531A1 (en) 2017-05-27 2019-11-20 Video image presentation and encapsulation method and video image presentation and encapsulation apparatus

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201710387835.0 2017-05-27
CN201710387835.0A CN108965917B (zh) 2017-05-27 2017-05-27 视频图像的呈现、封装方法和视频图像的呈现、封装装置

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US16/689,517 Continuation US20200092531A1 (en) 2017-05-27 2019-11-20 Video image presentation and encapsulation method and video image presentation and encapsulation apparatus

Publications (1)

Publication Number Publication Date
WO2018219202A1 true WO2018219202A1 (fr) 2018-12-06

Family

ID=64455214

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2018/088197 Ceased WO2018219202A1 (fr) 2017-05-27 2018-05-24 Procédé de présentation et d'encapsulation d'une image vidéo, et dispositif de présentation et d'une encapsulation d'image vidéo

Country Status (3)

Country Link
US (1) US20200092531A1 (fr)
CN (1) CN108965917B (fr)
WO (1) WO2018219202A1 (fr)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR102598082B1 (ko) * 2016-10-28 2023-11-03 삼성전자주식회사 영상 표시 장치, 모바일 장치 및 그 동작방법
WO2020027349A1 (fr) * 2018-07-31 2020-02-06 엘지전자 주식회사 Procédé pour traitement vidéo 360° basé sur de multiples points de vue et appareil associé
CN113489791B (zh) * 2021-07-07 2024-05-14 佳都科技集团股份有限公司 图像上传方法、图像处理方法及相关装置

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101247513A (zh) * 2007-12-25 2008-08-20 谢维信 利用单摄像机实时生成360°无缝全景视频图像的方法
US20130162641A1 (en) * 2010-09-14 2013-06-27 Thomson Licensing Method of presenting three-dimensional content with disparity adjustments
CN104735464A (zh) * 2015-03-31 2015-06-24 华为技术有限公司 一种全景视频交互传输方法、服务器和客户端
CN105791882A (zh) * 2016-03-22 2016-07-20 腾讯科技(深圳)有限公司 视频编码方法及装置
CN105869113A (zh) * 2016-03-25 2016-08-17 华为技术有限公司 全景图像的生成方法和装置

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101661265B (zh) * 2009-09-29 2011-01-05 哈尔滨师范大学 数字信息立体显示的多通道全息记录方法
US9407904B2 (en) * 2013-05-01 2016-08-02 Legend3D, Inc. Method for creating 3D virtual reality from 2D images
US10204658B2 (en) * 2014-07-14 2019-02-12 Sony Interactive Entertainment Inc. System and method for use in playing back panorama video content
US9858706B2 (en) * 2015-09-22 2018-01-02 Facebook, Inc. Systems and methods for content streaming
CN106341673A (zh) * 2016-08-15 2017-01-18 李文松 一种新型2d/3d全景vr视频的存储方法
CN106162207B (zh) * 2016-08-25 2019-02-12 北京字节跳动科技有限公司 一种全景视频并行编码方法和装置
CN106358033B (zh) * 2016-08-25 2018-06-19 北京字节跳动科技有限公司 一种全景视频关键帧编码方法和装置

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101247513A (zh) * 2007-12-25 2008-08-20 谢维信 利用单摄像机实时生成360°无缝全景视频图像的方法
US20130162641A1 (en) * 2010-09-14 2013-06-27 Thomson Licensing Method of presenting three-dimensional content with disparity adjustments
CN104735464A (zh) * 2015-03-31 2015-06-24 华为技术有限公司 一种全景视频交互传输方法、服务器和客户端
CN105791882A (zh) * 2016-03-22 2016-07-20 腾讯科技(深圳)有限公司 视频编码方法及装置
CN105869113A (zh) * 2016-03-25 2016-08-17 华为技术有限公司 全景图像的生成方法和装置

Also Published As

Publication number Publication date
CN108965917A (zh) 2018-12-07
CN108965917B (zh) 2021-07-20
US20200092531A1 (en) 2020-03-19

Similar Documents

Publication Publication Date Title
US11758187B2 (en) Methods, devices and stream for encoding and decoding volumetric video
US11025955B2 (en) Methods, devices and stream for encoding and decoding volumetric video
US11095920B2 (en) Method and apparatus for encoding a point cloud representing three-dimensional objects
CN112425177B (zh) 用于体积视频传输的方法和装置
CN107454468B (zh) 对沉浸式视频进行格式化的方法、装置和流
US11388442B2 (en) Point cloud encoding method, point cloud decoding method, encoder, and decoder
CN110419224B (zh) 消费视频内容的方法、电子设备和服务器
CN108271044B (zh) 一种信息的处理方法及装置
CN111264058A (zh) 用于对三自由度和体积兼容视频流进行编码和解码的方法、设备
US10735826B2 (en) Free dimension format and codec
CN109218755B (zh) 一种媒体数据的处理方法和装置
CN108235031B (zh) 一种运动矢量解码方法及解码器
WO2018141116A1 (fr) Procédé et dispositif de codage et de décodage
WO2020149990A1 (fr) Procédés et appareil de traitement multi-encodeur de contenu à haute résolution
US20200092531A1 (en) Video image presentation and encapsulation method and video image presentation and encapsulation apparatus
JP2019514313A (ja) レガシー及び没入型レンダリングデバイスのために没入型ビデオをフォーマットする方法、装置、及びストリーム
CN112771878B (zh) 处理媒体数据的方法、客户端和服务器
WO2020015517A1 (fr) Procédé de codage de nuage de points, procédé de décodage de nuage de points et décodeur
CN114080799B (zh) 处理体数据

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18810212

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18810212

Country of ref document: EP

Kind code of ref document: A1