CN1568015A - Multiple visual-angle video coding-decoding prediction compensation method and apparatus - Google Patents
Multiple visual-angle video coding-decoding prediction compensation method and apparatus Download PDFInfo
- Publication number
- CN1568015A CN1568015A CN 03148031 CN03148031A CN1568015A CN 1568015 A CN1568015 A CN 1568015A CN 03148031 CN03148031 CN 03148031 CN 03148031 A CN03148031 A CN 03148031A CN 1568015 A CN1568015 A CN 1568015A
- Authority
- CN
- China
- Prior art keywords
- motion vector
- global motion
- view
- image
- prediction
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 90
- 239000013598 vector Substances 0.000 claims abstract description 141
- 230000008569 process Effects 0.000 claims abstract description 20
- 238000013139 quantization Methods 0.000 claims abstract description 11
- 230000009466 transformation Effects 0.000 claims abstract description 9
- 101100283413 Saccharomyces cerevisiae (strain ATCC 204508 / S288c) GMC1 gene Proteins 0.000 claims description 10
- 238000012545 processing Methods 0.000 claims description 6
- 238000004364 calculation method Methods 0.000 claims 2
- 230000001131 transforming effect Effects 0.000 claims 1
- 230000000875 corresponding effect Effects 0.000 description 18
- 238000005516 engineering process Methods 0.000 description 8
- 238000010586 diagram Methods 0.000 description 5
- 230000000007 visual effect Effects 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 2
- 230000006835 compression Effects 0.000 description 2
- 238000007906 compression Methods 0.000 description 2
- 238000006073 displacement reaction Methods 0.000 description 2
- 238000012804 iterative process Methods 0.000 description 2
- 238000005457 optimization Methods 0.000 description 2
- 230000008520 organization Effects 0.000 description 2
- 238000010187 selection method Methods 0.000 description 2
- 238000013519 translation Methods 0.000 description 2
- PXFBZOLANLWPMH-UHFFFAOYSA-N 16-Epiaffinine Natural products C1C(C2=CC=CC=C2N2)=C2C(=O)CC2C(=CC)CN(C)C1C2CO PXFBZOLANLWPMH-UHFFFAOYSA-N 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000002457 bidirectional effect Effects 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 238000003786 synthesis reaction Methods 0.000 description 1
Images
Landscapes
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
本发明公开了一种多视角视频编解码预测补偿方法,编码为:在两个视间,使用其一为参考视,另一视作目标视,经过变形后对参考视进行全局运动预测,获得全局运动预测参数;然后进行每个宏块全局运动预测补偿和运动向量预测补偿;根据选定的预测补偿,获得宏块残差图像;对残差图像进行变换、量化、熵编码,最终输出宏块编码码流;所述解码首先获得目标视中当前帧对参考视的全局运动预测参数,进入每个宏块解码过程;进行熵解码、反量化及反变换,获得宏块采用的预测补偿方式、残差图像等信息。本发明在两个相关视序列间采用全局运动预测补偿和运动向量预测补偿方法,可更充分利用多视角视频中各个视之间的相关性,可有效对多视角视频进行编解码。
The invention discloses a multi-view video encoding and decoding prediction and compensation method. The encoding is as follows: between two views, use one as a reference view and the other as a target view, perform global motion prediction on the reference view after deformation, and obtain Global motion prediction parameters; then perform global motion prediction compensation and motion vector prediction compensation for each macroblock; obtain the residual image of the macroblock according to the selected prediction compensation; transform, quantize, and entropy encode the residual image, and finally output the macro Block coding code stream; the decoding first obtains the global motion prediction parameters of the current frame in the target view to the reference view, and enters the decoding process of each macroblock; performs entropy decoding, inverse quantization and inverse transformation, and obtains the prediction compensation method adopted by the macroblock , residual image and other information. The present invention adopts global motion prediction and compensation and motion vector prediction and compensation methods between two related view sequences, can make more full use of the correlation between views in multi-view video, and can effectively encode and decode multi-view videos.
Description
技术领域technical field
本发明涉及数字图像处理技术,特别是涉及数字图像的编解码技术,具体地说,涉及一种多视角视频编解码的方法。The present invention relates to a digital image processing technology, in particular to a digital image coding and decoding technology, in particular to a multi-view video coding and decoding method.
背景技术Background technique
随着网络及多媒体技术的快速发展,越来越多的图像和视频以数字媒体的形式出现和传输,而高效的视频编解码技术是实现数字媒体存储与传输的关键。目前,一种新的数字媒体形式,即三维视频,已经在多媒体应用系统中广泛采用。所谓的三维视频是指时间上同步,空间上相关的若干视频序列的集合。与传统的二维视频相比,三维视频可以提供更丰富的视觉信息,因此可以为用户提供更高质量的主观视觉享受,并可以应用到视频会议,数字娱乐等应用中。多视角视频(Multiview Video)是一种常见的三维视频。所谓的多视角视频是指通过多个摄像机在不同的视角同步采集的一组视频序列。通过某种合成技术,多视角视频可以为用户提供具有三维视感觉的场景。此外,多视角视频的一个关键特征是场景中的交互性,用户还能够在场景中选择他自己的视点。With the rapid development of network and multimedia technology, more and more images and videos appear and transmit in the form of digital media, and efficient video codec technology is the key to realize digital media storage and transmission. At present, a new digital media form, that is, three-dimensional video, has been widely used in multimedia application systems. The so-called 3D video refers to a collection of several video sequences that are synchronized in time and correlated in space. Compared with traditional two-dimensional video, three-dimensional video can provide richer visual information, so it can provide users with higher-quality subjective visual enjoyment, and can be applied to applications such as video conferencing and digital entertainment. Multiview Video is a common 3D video. The so-called multi-view video refers to a group of video sequences that are synchronously collected by multiple cameras at different viewing angles. Through some synthesis technology, multi-view video can provide users with scenes with three-dimensional visual sense. Furthermore, a key feature of multi-view video is the interactivity in the scene, where the user is also able to choose his own viewpoint.
相对于传统的二维视频,多视角视频的数据量更大,实现其功能的一个代价便是表示场景的视频数据随着视点(即相机数目)的增加以几何速度增长,因此,对高效的多视角视频数据的处理方法就显得非常重要。为了实现多视角视频的存储与传输,就需要对多视角视频进行高效地编解码。目前,先进的视频编解码技术通常以标准的形式存在,典型的视频压缩标准有国际标准化组织(ISO)下设的运动图像专家组(Moving Picture Expert Group,简称MPEG)推出的MPEG系列国际标准MPEG-1,MPEG-2,MPEG-4等,以及国际电信联盟(ITU)提出的H.26x系列视频压缩推荐等。目前,MPEG组织正在进行三维音视频(3DAV)的标准化工作,但目前还没有形成具体的标准。Compared with the traditional two-dimensional video, the multi-view video has a larger amount of data, and one of the costs of realizing its function is that the video data representing the scene grows at a geometric speed with the increase of the viewpoint (that is, the number of cameras). Therefore, for efficient The processing method of multi-view video data is very important. In order to realize the storage and transmission of multi-view video, it is necessary to efficiently encode and decode the multi-view video. At present, advanced video coding and decoding technologies usually exist in the form of standards. Typical video compression standards include the MPEG series of international standards MPEG developed by the Moving Picture Expert Group (MPEG) under the International Organization for Standardization (ISO). -1, MPEG-2, MPEG-4, etc., and the H.26x series of video compression recommendations proposed by the International Telecommunication Union (ITU). At present, the MPEG organization is working on the standardization of three-dimensional audio and video (3DAV), but no specific standard has been formed yet.
对多视角视频数据进行编解码的一个最简单的方法是利用现有的编解码标准,例如MPEG-1,MPEG-2,MPEG-4,JVT或者H.26x等,将每个视角的视频数据看成普通的视频序列分别进行编解码。但是,这种方法没有利用各个视角间的相关性,编码效率并不是很高。虽然MPEG标准中的一些技术可以用来在某种程度上利用各个视角间的相关性来进行编解码,但是这些技术提出的初衷并不是专门针对多视角视频编解码,没有充分考虑多视角视频的特性,因此编码效率没有达到最优。也有一些编解码系统利用了多视角视频在同一时刻不同视角的图像之间的空间相关性信息,但是,由于它们主要是针对单视(2D)视频的编码而提出的,它们没有考虑多视角视频各个视之间的特殊关系,例如,相邻视之间对应图像的场景内容大部分都相同,而且场景的运动趋势完全相同,只不过角度有所变化而已,这在许多实例中可以看到(除了快速运动的实例外),一个视中前后帧间的相似性要比不同视点间的相似性要高,于是,运动补偿帧间预测很可能替代视通道间的位差补偿(各个视间的运动预测补偿)预测,而视通道间的位差补偿仅仅在某些地方比分别编码每个视占优势,其编码效率并不十分理想。One of the simplest methods for encoding and decoding multi-view video data is to use existing codec standards, such as MPEG-1, MPEG-2, MPEG-4, JVT or H.26x, etc., to convert the video data of each view It is regarded as an ordinary video sequence and is coded and decoded separately. However, this method does not take advantage of the correlation between views, and the coding efficiency is not very high. Although some technologies in the MPEG standard can be used to some extent to use the correlation between various views for encoding and decoding, the original intention of these technologies is not specifically for multi-view video encoding and decoding, and does not fully consider the multi-view video. characteristics, so the coding efficiency is not optimal. There are also some codec systems that utilize the spatial correlation information between images of different perspectives at the same time in multi-view video, but because they are mainly proposed for the coding of single-view (2D) video, they do not consider multi-view video The special relationship between views, for example, the scene content of the corresponding images between adjacent views is mostly the same, and the motion trend of the scene is exactly the same, but the angle changes, which can be seen in many instances ( Except for fast motion instances), the similarity between frames before and after a view is higher than the similarity between different views, so motion compensated inter prediction is likely to replace disparity compensation between view channels (the motion prediction compensation) prediction, and the disparity compensation between viewing channels is only in some places superior to encoding each viewing separately, and its coding efficiency is not very ideal.
发明内容Contents of the invention
本发明要解决的技术问题是提出一种新型的基于全局运动预测补偿和运动向量预测补偿的多视角视频编解码预测补偿方法及装置,以提高多视角视频编解码的编码效率。The technical problem to be solved by the present invention is to propose a novel multi-view video coding and decoding prediction compensation method and device based on global motion prediction compensation and motion vector prediction compensation, so as to improve the coding efficiency of multi-view video coding and decoding.
本发明所述多视角视频编解码预测补偿的方法,在两个视之间使用全局运动预测补偿来提取各个视之间对应图像的空间相关性,进而提高编码效率,然后在实现全局运动预测补偿的基础上,以一个视(参考视)的前后帧之间的运动向量来预测另一个视(目标视)前后帧之间的运动向量。具体的说:The method for predictive compensation of multi-view video encoding and decoding in the present invention uses global motion prediction compensation between two views to extract the spatial correlation of corresponding images between each view, thereby improving coding efficiency, and then realizes global motion prediction compensation On the basis of , the motion vector between the frames before and after another view (target view) is predicted by the motion vector between the frames before and after one view (reference view). Specifically:
在编码时,使用其中的一个视作为参考帧,另一个相邻视作为目标视,目标视图像参考变形(平移、旋转、缩放等)后的参考视图像进行全局运动预测补偿,获得全局运动预测参数,然后再利用这些参数计算每个宏块的全局运动预测代价(一般使用原始宏块和对编码后的宏块进行解码得到的重构宏块各个象素之间的绝对差来衡量),获得该宏块的全局运动向量。运动向量预测补偿利用获得的全局运动向量和已编码的参考视前后帧之间的运动向量,预测获得目标视前后帧之间的运动向量,并计算该方法的运动预测代价。根据运动预测代价,选择当前宏块采用的编码预测补偿方法,然后根据选定的预测补偿方法或者进行宏块的全局运动补偿(若是全局运动预测补偿方法被选中)或者进行宏块的局部运动补偿(若运动向量预测补偿方法被选中),进而获得宏块残差图像。最后,对残差图像进行变换、量化、熵编码,最终输出编码码流。When encoding, use one of the views as the reference frame, and the other adjacent view as the target view, and the target view image refers to the reference view image after deformation (translation, rotation, scaling, etc.) for global motion prediction compensation to obtain global motion prediction parameters, and then use these parameters to calculate the global motion prediction cost of each macroblock (generally measured by the absolute difference between each pixel of the original macroblock and the reconstructed macroblock obtained by decoding the coded macroblock), Get the global motion vector for this macroblock. Motion vector prediction and compensation utilizes the obtained global motion vector and the coded motion vector between frames before and after the reference view to predict the motion vector between the frames before and after the target view, and calculates the motion prediction cost of this method. According to the motion prediction cost, select the encoding prediction compensation method used by the current macroblock, and then perform global motion compensation of the macroblock (if the global motion prediction compensation method is selected) or local motion compensation of the macroblock according to the selected prediction compensation method (if the motion vector prediction and compensation method is selected), and then obtain the residual image of the macroblock. Finally, transform, quantize, and entropy code the residual image, and finally output the coded stream.
解码时,首先解码获得目标视中当前帧对参考视的全局运动预测参数,然后利用这些参数计算获得当前解码宏块的全局运动向量,并保存,以备后用。再解码获得宏块使用的预测补偿方法及残差图像。根据不同的预测补偿方法,按照编码时的运动补偿方法重构该宏块的图像,并与解码的残差图像合并形成最终的宏块解码图像;对每个宏块使用上述方法进行解码,即可得到该帧的解码图像。When decoding, first obtain the global motion prediction parameters of the current frame in the target view to the reference view, and then use these parameters to calculate the global motion vector of the currently decoded macroblock and save it for later use. Re-decoding to obtain the prediction compensation method used by the macroblock and the residual image. According to different prediction and compensation methods, the image of the macroblock is reconstructed according to the motion compensation method during encoding, and combined with the decoded residual image to form the final macroblock decoded image; each macroblock is decoded using the above method, that is A decoded image of the frame is available.
本发明还提出了一种多视角视频编码预测补偿装置,包括参考视图像编码单元和目标视图像编码单元,参考视图像编码单元经过局部运动预测补偿进行编码,最终输出编码后的参考视码流;目标视图像编码单元经过全局运动预测补偿和运动向量预测补偿进行编码,最后输出编码后的目标视码流。The present invention also proposes a multi-view video coding prediction and compensation device, including a reference view image encoding unit and a target view image encoding unit. The reference view image encoding unit performs encoding through local motion prediction and compensation, and finally outputs the encoded reference view code stream ; The target-view image coding unit encodes through global motion prediction compensation and motion vector prediction compensation, and finally outputs the coded target-view code stream.
本发明利用在两个相关视序列之间采用的全局运动预测补偿和运动向量预测补偿方法,较之现有的编码方法,可以更充分利用多视角视频中各个视之间的相关性,因此可以有效对多视角视频进行编解码。The present invention utilizes the global motion prediction and compensation and motion vector prediction and compensation methods used between two related view sequences. Compared with the existing coding methods, it can make full use of the correlation between each view in multi-view video, so it can Effectively encode and decode multi-view video.
全局运动预测补偿作用在同一时刻通过多个摄像机采集的多幅图像上,其目的是提取这些图像的空间相关性,充分考虑了多视角视频相邻视之间的特性,特别是场景内容大部分相同,运动趋势相同,所不同的地方是或者角度不同,或者整个场景之间存在相对位移等特性,这些特性使得全局运动预测补偿较之直接的局部运动预测补偿可以获得更准确的运动预测,而且它不需要编码运动向量,减少编码运动向量信息的比特,进而提高编码效率。The global motion prediction compensation acts on multiple images collected by multiple cameras at the same time. Its purpose is to extract the spatial correlation of these images, fully considering the characteristics between adjacent views of multi-view videos, especially most of the scene content. The same, the same motion trend, the difference is that either the angle is different, or there are characteristics such as relative displacement between the entire scene, these characteristics make the global motion prediction compensation can obtain more accurate motion prediction than the direct local motion prediction compensation, and It does not need to code the motion vector, and reduces the bits for coding the motion vector information, thereby improving the coding efficiency.
运动向量预测补偿方式是在实现全局运动预测补偿方式的基础上进行的,其目的是利用已经编码的某个摄像机采集的序列的运动信息,预测其它摄像机采集的序列中的运动信息,这种预测可以采用相同的方法分别在编码器和解码器中实现,因此不需要编码运动向量,减少编码运动矢量信息的比特,进而提高编码效率。The motion vector prediction and compensation method is based on the realization of the global motion prediction and compensation method. Its purpose is to use the motion information of the sequence captured by a certain camera that has been encoded to predict the motion information in the sequence captured by other cameras. The same method can be used to implement in the encoder and the decoder respectively, so there is no need to encode the motion vector, and the bits for encoding the motion vector information are reduced, thereby improving the encoding efficiency.
附图说明Description of drawings
图1是运动向量关系示意图;Fig. 1 is a schematic diagram of motion vector relationship;
图2是本发明提出的运动向量中心点预测流程图;Fig. 2 is the motion vector central point prediction flow chart that the present invention proposes;
图3是通过迭代匹配获取参考视到目标视的全局运动向量的流程图;Fig. 3 is a flow chart of obtaining the global motion vector from the reference view to the target view through iterative matching;
图4是多视角视频编码预测补偿装置之一示意图;Fig. 4 is a schematic diagram of a multi-view video coding prediction compensation device;
图5是多视角视频编码预测补偿装置之二示意图;Fig. 5 is a schematic diagram of a second multi-view video coding prediction compensation device;
图6是基于MPEG-4时间可扩展编码方法的预测原理示意图。Fig. 6 is a schematic diagram of the prediction principle based on the MPEG-4 time-scalable coding method.
具体实施方式Detailed ways
在多视角视频序列中,相邻视角的两个视频序列的对应帧存在着相当大的相关性,即相邻视角的两个视频序列的对应帧场景内容大部分相同,运动趋势相同,所不同的地方是角度不同,或者整个场景之间存在相对位移,这也隐含一个视(目标视)的前后帧之间的运动向量可以通过另一个视(参考视)前后帧之间的运动向量预测来获得。它们经过简单的变形(平移、旋转、仿射变换等)后,大部分图像非常相似。如果能利用这些特性,在两个对应帧之间采用全局运动预测补偿方式,及在目标视的前后帧之间采用运动向量预测补偿方式,就可以提高多视角视频的编码效率。包括如下步骤:In multi-view video sequences, there is a considerable correlation between the corresponding frames of two video sequences of adjacent viewing angles, that is, the scene content of the corresponding frames of two video sequences of adjacent viewing angles is mostly the same, and the motion trend is the same. The place where the angle is different, or there is a relative displacement between the entire scene, which also implies that the motion vector between the frames before and after a view (target view) can be predicted by the motion vector between the frames before and after another view (reference view) to get. After they undergo simple deformations (translation, rotation, affine transformation, etc.), most of the images are very similar. If these characteristics can be used, the global motion prediction and compensation method is adopted between two corresponding frames, and the motion vector prediction and compensation method is adopted between the frames before and after the target view, the coding efficiency of multi-view video can be improved. Including the following steps:
编码步骤如下:The encoding steps are as follows:
首先使用其中的一个视作为参考视,另一个相邻视(以下称为目标视)经过变形后对参考视进行全局运动预测,获得全局运动预测参数。然后进入每个宏块全局运动预测补偿和运动向量预测补偿的编码过程,其步骤如下:First, one of the views is used as a reference view, and the other adjacent view (hereinafter referred to as the target view) is deformed to perform global motion prediction on the reference view to obtain global motion prediction parameters. Then enter the encoding process of each macroblock global motion prediction compensation and motion vector prediction compensation, the steps are as follows:
步骤1、全局运动预测补偿编码过程:
步骤1.1、利用获得的全局运动预测参数计算宏块的全局运动预测代价(一般使用原始宏块与对编码后的宏块进行解码得到的重构宏块各个象素之间的绝对差和来衡量);Step 1.1, use the obtained global motion prediction parameters to calculate the global motion prediction cost of the macroblock (generally measured by the sum of absolute differences between the pixels of the original macroblock and the reconstructed macroblock obtained by decoding the encoded macroblock );
步骤1.2、计算获得该宏块的全局运动向量,将其保存,已备后用;Step 1.2, calculating and obtaining the global motion vector of the macroblock, saving it for future use;
步骤2、运动向量预测补偿编码过程:Step 2, motion vector prediction compensation encoding process:
步骤2.1、利用获得的该宏块的全局运动向量和前一帧已保存的全局运动向量及已编码的参考视前后帧之间的运动向量预测获得目标视前后帧之间的运动向量;Step 2.1, using the obtained global motion vector of the macroblock, the saved global motion vector of the previous frame and the coded motion vector prediction between the frames before and after the reference view to obtain the motion vector between the frames before and after the target view;
步骤2.2、利用预测得到的运动向量计算宏块的局部运动预测代价;Step 2.2, using the predicted motion vector to calculate the local motion prediction cost of the macroblock;
步骤3、选择计算获得的全局运动预测代价和局部运动预测代价中较小的预测补偿作为宏块的预测补偿;Step 3, selecting the smaller prediction compensation among the calculated global motion prediction cost and the local motion prediction cost as the prediction compensation of the macroblock;
步骤4、根据选定的预测补偿或者进行宏块的全局运动补偿(若是全局运动预测补偿方法被选中)或者进行宏块的局部运动补偿(若运动向量预测补偿方法被选中),进而与原始宏块做差获得宏块残差图像;Step 4, according to the selected prediction compensation or carry out the global motion compensation of macroblock (if the global motion prediction compensation method is selected) or carry out the local motion compensation of macroblock (if the motion vector prediction compensation method is selected), and then with the original macro Block difference to obtain macroblock residual image;
步骤5、对残差图像进行变换、量化、熵编码,最终输出宏块编码码流。Step 5: Perform transformation, quantization, and entropy coding on the residual image, and finally output a coded stream of macroblocks.
解码步骤如下:The decoding steps are as follows:
首先解码获得目标视中当前帧对参考视的全局运动预测参数,然后进入每个宏块解码过程,其步骤如下:First decode to obtain the global motion prediction parameters of the current frame in the target view to the reference view, and then enter the decoding process of each macroblock, the steps are as follows:
步骤1、进行熵解码、反量化及反变换,获得宏块采用的预测补偿方式、残差图像等信息;
步骤2、利用获得的全局运动预测参数计算获得当前解码宏块的全局运动向量,并保存,以备后用;Step 2, using the obtained global motion prediction parameters to calculate and obtain the global motion vector of the currently decoded macroblock, and save it for later use;
步骤3、若宏块采用的是全局运动预测补偿,则按如下步骤进行解码:Step 3, if the macroblock adopts global motion prediction compensation, then decode according to the following steps:
步骤3.1、将参考视中对应的参考帧按照获得的全局运动预测参数进行变形;Step 3.1, deforming the corresponding reference frame in the reference view according to the obtained global motion prediction parameters;
步骤3.2、利用获得的全局运动向量参考变形后的参考帧对当前解码宏块进行全局运动补偿,获得宏块补偿图像;Step 3.2, using the obtained global motion vector to refer to the deformed reference frame to perform global motion compensation on the currently decoded macroblock to obtain a macroblock compensated image;
步骤3.3、宏块补偿图像与解码的残差图像合并得到解码宏块图像。In step 3.3, the macroblock compensation image is combined with the decoded residual image to obtain a decoded macroblock image.
步骤4、若宏块采用的是运动预测补偿,则按如下步骤进行解码:Step 4, if the macroblock adopts motion prediction compensation, then decode according to the following steps:
步骤4.1、利用已保存的解码后的全局运动向量和已解码的参考视前后帧间的运动向量,计算获得目标视前后帧间的运动向量;Step 4.1, using the saved decoded global motion vector and the decoded motion vector between frames before and after the reference view, to calculate and obtain the motion vector between frames before and after the target view;
步骤4.2、利用计算获得的运动向量参考目标视的前一解码帧对当前解码宏块进行运动补偿,获得宏块补偿图像;Step 4.2, using the calculated motion vector to refer to the previous decoded frame of the target view to perform motion compensation on the currently decoded macroblock to obtain a macroblock compensated image;
步骤4.3、宏块补偿图像与解码的残差图像合并得到解码宏块图像。In step 4.3, the macroblock compensation image is combined with the decoded residual image to obtain a decoded macroblock image.
图1是运动向量关系示意图,图1中目标视和参考视是多视角视频序列中的两个视,一个被作为参考视,另一个作为目标视。它表示的是某一编码时刻序列参考视和目标视前后帧之间的关系(第i-1帧表示前一帧,第i帧表示当前帧),描述了参考视和目标视之间的全局运动向量(在图1中使用GMC0和GMC1来标明)和同一个视中的运动向量(在图1中使用MV0和MV1来标明)之间的关系,其中:Figure 1 is a schematic diagram of the relationship between motion vectors. In Figure 1, the target view and the reference view are two views in a multi-view video sequence, one is used as the reference view, and the other is used as the target view. It represents the relationship between the reference view and the frame before and after the target view at a certain encoding time (the i-1th frame represents the previous frame, and the i-th frame represents the current frame), and describes the global relationship between the reference view and the target view. The relationship between motion vectors (indicated by GMC0 and GMC1 in Figure 1) and motion vectors in the same view (indicated by MV0 and MV1 in Figure 1), where:
上面的两个网格表示某一时刻目标视的前后两帧图像,下面的两个网格表示某一时刻参考视的前后两帧图像,网格中的每个小方块表示一个宏块(图像象素块),目标视网格中的灰色方块表示当前编码宏块。The upper two grids represent the two frames of images before and after the target view at a certain moment, the two lower grids represent the two frames of images before and after the reference view at a certain moment, and each small square in the grid represents a macroblock (image pixel block), the gray square in the target view grid represents the current coded macroblock.
GMC0表示前一个已编码目标帧到前一个已编码参考帧之间的全局运动向量;GMC0 represents the global motion vector between the previous coded target frame and the previous coded reference frame;
GMC1表示当前编码目标帧到对应的已编码参考帧之间的全局运动向量;GMC1 represents the global motion vector between the current encoding target frame and the corresponding encoded reference frame;
MV0表示参考视前后帧间的运动向量;MV0 refers to the motion vector between frames before and after the reference view;
MV1表示目标视前后帧间的运动向量,它也是本发明中需要进行运动向量预测获得的运动向量。MV1 represents the motion vector between frames before and after the target view, and it is also the motion vector that needs to be obtained by motion vector prediction in the present invention.
图2是本发明提出的运动向量中心点预测流程图。它描述的是目标视序列通过参考视的前后帧运动向量及已获得的全局运动向量,来预测目标视对应前后帧运动向量的一种预测编码的做法。Fig. 2 is a flow chart of motion vector central point prediction proposed by the present invention. It describes a method of predictive coding in which the target view sequence predicts the motion vectors of the previous and subsequent frames corresponding to the target view by referring to the previous and subsequent frame motion vectors of the view and the obtained global motion vector.
具体做法是,当编码目标视的某个宏块(在图1中以灰色方块来表示)时,完成如下步骤:Specifically, when encoding a certain macroblock of the target view (indicated by a gray square in Figure 1), complete the following steps:
步骤1、获得全局运动向量GMC0和GMC1(解释见图1说明,在全局运动预测补偿过程中已经获得,见编码实施步骤1.2)及局部运动向量MV0(在编码参考视时产生);
步骤2、计算获得当前宏块的中心点(CurXcent,CurYcent),(图1中用CurCentPos来表示,即灰色方块中的小黑点);Step 2, calculate and obtain the center point (CurXcent, CurYcent) of current macroblock, (represent by CurCentPos in Fig. 1, promptly the small black dot in the gray square);
步骤3、通过GMC1运动向量和位置(CurXcent,CurYcent)计算获得点(CurXcent,CurYcent)在参考视图像中的位置(CurRefX,CurRefY),(图1中用CurPosRef来表示);Step 3, calculate and obtain the position (CurRefX, CurRefY) of point (CurXcent, CurYcent) in the reference view image by GMC1 motion vector and position (CurXcent, CurYcent), (represented by CurPosRef among Fig. 1);
步骤4、利用MV0,计算获得点(CurRefX,CurRefY)在其前一帧参考视图像中对应的位置(PreRefX,PreRefY),(图1中用PrePosRef来表示);Step 4, use MV0 to calculate the corresponding position (PreRefX, PreRefY) of the obtained point (CurRefX, CurRefY) in its previous frame reference view image (represented by PrePosRef in Figure 1);
步骤5、利用GMC0,通过迭代匹配获得参考视到目标视的全局运动向量(GMVX,GMVY),(即将图1中的GMC0反向),详细过程将在图3中描述;Step 5. Using GMC0, obtain the global motion vector (GMVX, GMVY) from the reference view to the target view through iterative matching, (reverse the GMC0 in Fig. 1), and the detailed process will be described in Fig. 3;
步骤6、将获得的全局运动向量(GMVX,GMVY)反向,进而利用其计算位置(PreRefX,PreRefY)在目标视的前一帧图像中的对应位置(PreXcent,PreYcent),(图1中用PrePosCur来表示);Step 6, reverse the obtained global motion vector (GMVX, GMVY), and then use it to calculate the corresponding position (PreXcent, PreYcent) of the previous frame image of the target view (PreRefX, PreRefY), (use in Fig. 1 PrePosCur to represent);
步骤7、计算获得运动向量MV1,MV1X=PreXcent-CurXcent,MV1Y=PreXcent-CurYcent;(图1中粗虚线表示的向量)。Step 7. Calculate and obtain the motion vector MV1, MV1X=PreXcent-CurXcent, MV1Y=PreXcent-CurYcent; (the vector represented by the thick dotted line in Fig. 1 ).
值得指出的是,如果为了追求更高的预测精度,除了本发明提出的中心点预测方式外,也可以采用多点预测,然后求其平均值的方式,但这仍然属于本发明的处理思路。It is worth pointing out that, in order to pursue higher prediction accuracy, in addition to the central point prediction method proposed by the present invention, multi-point prediction and then calculating the average value can also be used, but this still belongs to the processing idea of the present invention.
图3是迭代匹配获取参考视到目标视的全局运动向量的方法。它描述了图2中加粗功能框的实现。它的输入是经过计算得到的在参考视中的对应位置点(PreRefX,PreRefY)、目标视到参考视的全局运动向量GMC0,最终输出获得对应的参考视到目标视的全局运动向量(GMVX,GMVY)。其中,Fig. 3 is a method of iterative matching to obtain the global motion vector from the reference view to the target view. It describes the implementation of the bold functional box in Figure 2. Its input is the calculated corresponding position point (PreRefX, PreRefY) in the reference view, the global motion vector GMC0 from the target view to the reference view, and finally outputs the corresponding global motion vector from the reference view to the target view (GMVX, GMVY). in,
MBSIZE表示宏块的大小,可以根据不同的环境来取不同的值;MBSIZE represents the size of the macroblock, which can take different values according to different environments;
N控制最大的迭代次数,可根据实际需要设置。它用来保证当通过迭代不能获得完全匹配时,控制迭代结束。它的大小控制迭代搜索的范围;N controls the maximum number of iterations, which can be set according to actual needs. It is used to ensure that when an exact match cannot be obtained through iteration, the control iteration ends. Its size controls the scope of the iterative search;
Minerr初始化时,为MAXERROR,MAXERROR可以根据实际情况取任何最大阈值。Minerr记录迭代过程中最小的迭代误差,最终采用Minerr(迭代误差)最小的目标视宏块的全局运动向量来表示参考视到目标视的全局运动向量(GMVX,GMVY)。When Minerr is initialized, it is MAXERROR, and MAXERROR can take any maximum threshold according to the actual situation. Minerr records the smallest iteration error in the iterative process, and finally uses the global motion vector of the target view macroblock with the smallest Minerr (iteration error) to represent the global motion vector (GMVX, GMVY) from the reference view to the target view.
具体地,Minerr初始化,令迭代误差Minerr=MAXERROR;Specifically, Minerr is initialized, making iterative error Minerr=MAXERROR;
计算获得点(PreRefX,PreRefY)所在的宏块位置(PreMBX,PreMBY),通过GMC0和宏块位置(PreMBX,PreMBY)获得全局运动向量(GMVX,GMVY);Calculate the macroblock position (PreMBX, PreMBY) where the obtained point (PreRefX, PreRefY) is located, and obtain the global motion vector (GMVX, GMVY) through GMC0 and the macroblock position (PreMBX, PreMBY);
用点(PreRefX,PreRefY)减去全局运动向量(GMVX,GMVY),获得新的位置(PreCurX,PreCurY);再计算获得点(PreCurX,PreCurY)所在的宏块位置(PreCurMBX,PreCurMBY)。Subtract the global motion vector (GMVX, GMVY) from the point (PreRefX, PreRefY) to obtain a new position (PreCurX, PreCurY); then calculate the macroblock position (PreCurMBX, PreCurMBY) where the obtained point (PreCurX, PreCurY) is located.
迭代过程是这样的:将迭代后获得的宏块位置(PreCurMBX,PreCurMBY)与迭代前的宏块位置(PreMBX,PreMBY)比较,若相等,则表示完全获得,迭代误差Minerr为0,直接退出迭代过程,如果不相等,若在迭代前的宏块位置的左侧,则获得宏块左边X方向位置和迭代后的对应位置点的X方向位置的绝对差Xerr,否则获得宏块右边X方向位置和迭代后的对应位置点的X方向位置的绝对差Xerr,与此类似,若在上方,则获得宏块上边Y方向位置和迭代后的对应位置点的Y方向位置的绝对差Yerr,否则获得宏块下边Y方向位置和迭代后的对应位置点的Y方向位置的绝对差Yerr,最后Xerr和Yerr的和与Minerr比较,取其最小值给Minerr。The iterative process is as follows: compare the macroblock position (PreCurMBX, PreCurMBY) obtained after iteration with the macroblock position (PreMBX, PreMBY) before iteration, if they are equal, it means that it is completely obtained, the iteration error Minerr is 0, and exit the iteration directly Process, if not equal, if it is on the left side of the macroblock position before iteration, then obtain the absolute difference Xerr between the X-direction position on the left side of the macroblock and the X-direction position of the corresponding position point after iteration, otherwise obtain the X-direction position on the right side of the macroblock and the absolute difference Xerr of the X-direction position of the corresponding position point after iteration, similar to this, if it is above, then obtain the absolute difference Yerr of the Y-direction position on the upper side of the macroblock and the Y-direction position of the corresponding position point after iteration, otherwise obtain The absolute difference Yerr between the position in the Y direction of the lower side of the macroblock and the position in the Y direction of the corresponding point after iteration, and finally the sum of Xerr and Yerr is compared with Minerr, and the minimum value is given to Minerr.
图4作为本发明的具体实施例,表示的是多视角视频编码预测补偿装置1。该装置的输入是多视角视频中任意两个视的原始视频流(相邻视效果更佳),其中一个视称为参考视,另一个视称为目标视。参考视图像经过局部运动预测补偿方法进行编码,最终输出编码后的参考视码流;目标视图像经过全局运动预测补偿和运动向量预测补偿进行编码,最后输出编码后的目标视码流。其中,FIG. 4 shows a multi-view video coding
参考视图像和目标视图像分别是输入的参考视原始视频图像和目标视原始视频图像;The reference view image and the target view image are the input reference view original video image and the target view original video image respectively;
参考视重构图像缓冲区和目标视重构图像缓冲区分别存储前一帧编码后重构的参考视图像和目标视图像。The reference view reconstructed image buffer and the target view reconstructed image buffer respectively store the reconstructed reference view image and target view image after encoding of the previous frame.
它的工作顺序是首先对一帧参考视图像进行编码,然后再对一帧目标视图像进行编码。下面分别是参考视图像编码和目标视图像编码的过程:Its working sequence is to first encode a frame of reference view image, and then encode a frame of target view image. The following are the processes of encoding the reference view image and encoding the target view image respectively:
1、参考视图像编码单元1. Reference view image coding unit
它实现一帧参考视图像的预测补偿编码。它采用局部运动估计,并由此获得运动向量,将其进行熵编码。在运动向量缓冲区中保存该运动向量以备在编码目标视图像时使用。运动向量与前一帧重构参考视图像合作完成参考视运动补偿,获得补偿后的参考图像,即图4中参考视预测图像,该图像在与原始参考视图像做差,求得残差图像。残差图像经过变换、量化和熵编码,最终形成参考视码流。经过量化后的残差图像再经过反量化和反变换,并与参考视预测图像累加,进而得到参考视重构图像,并将其放入参考视重构缓冲区,以备下一帧参考视图像和目标视图像编码时使用。It implements prediction compensation coding of a frame of reference view. It uses local motion estimation, and thus obtains motion vectors, which are entropy coded. Save the motion vector in the motion vector buffer for use when encoding the target view image. The motion vector cooperates with the reconstructed reference view image in the previous frame to complete the reference view motion compensation, and obtain the compensated reference image, that is, the reference view prediction image in Figure 4, which is compared with the original reference view image to obtain the residual image . The residual image is transformed, quantized and entropy coded to finally form a reference video stream. The quantized residual image is dequantized and inversely transformed, and accumulated with the reference view prediction image to obtain the reference view reconstruction image, which is put into the reference view reconstruction buffer for the next frame of reference view Used when encoding images and target views.
2、目标视图像编码单元2. Target view image coding unit
它实现一帧目标视图像的预测补偿编码。It realizes the predictive compensation encoding of a frame of target image.
在输入的目标视原始图像和参考视重构图像之间进行全局运动估计,获得全局运动参数,并由此获得全局运动预测代价和全局运动向量,并将全局运动向量保存在全局运动向量缓冲区中,以备后用;Perform global motion estimation between the input original image of the target view and the reconstructed image of the reference view, obtain the global motion parameters, and thus obtain the global motion prediction cost and the global motion vector, and save the global motion vector in the global motion vector buffer in, for later use;
利用前一帧目标视图像编码时获得的全局运动向量及当前获得的全局运动向量和对应的参考视图像编码时保存的运动向量进行目标视运动向量预测,得到目标视图像参考前一帧重构目标视的局部运动向量;Use the global motion vector obtained during the encoding of the target view image in the previous frame, the currently obtained global motion vector and the motion vector saved during the encoding of the corresponding reference view image to perform target view motion vector prediction, and obtain the target view image with reference to the previous frame reconstruction The local motion vector of the target view;
计算目标视图像的局部运动预测代价,并通过选择控制器选择它和全局运动预测代价中较小的预测补偿方法进行下面的工作;Calculate the local motion prediction cost of the target view image, and select the smaller prediction compensation method in it and the global motion prediction cost by the selection controller to perform the following work;
根据选定的预测补偿方法或者进行全局运动补偿(若是全局运动预测补偿方法被选中)或者进行局部运动补偿(若运动向量预测补偿方法被选中),获得补偿后的目标图像,即图4中目标视预测图像,该图像在与原始目标视图像做差,求得残差图像;According to the selected prediction compensation method, either perform global motion compensation (if the global motion prediction compensation method is selected) or perform local motion compensation (if the motion vector prediction compensation method is selected), and obtain the compensated target image, that is, the target in Figure 4 Viewing the predicted image, the image is compared with the original target viewing image to obtain the residual image;
残差图像经过变换、量化和熵编码,最终形成目标视码流。经过量化后的残差图像再经过反量化和反变换,并与目标视预测图像累加,进而得到目标视重构图像,并将其放入目标视重构缓冲区,以备下一帧目标视图像编码时使用。The residual image is transformed, quantized and entropy coded to finally form the target video stream. The quantized residual image is dequantized and inversely transformed, and accumulated with the target-view prediction image to obtain the target-view reconstruction image, which is put into the target-view reconstruction buffer for the next frame of target-view reconstruction. Used when encoding images.
图5是本发明的另一个实施例,表示的是多视角视频编码预测补偿装置2。该装置2与装置1不同之处在于在编码目标视图像时,全局运动预测补偿和运动向量预测补偿作为两种预测模式与其它的预测模式共同完成编码工作。即,目标视图像除了使用全局运动预测补偿和运动向量预测补偿进行编码外,还可以引入当前已经被使用的其它预测补偿方法(例如,直接预测补偿方法),它们有助于多视角视频编码效率的提高。该装置2的输入是多视角视频中任意两个视的原始视频流(相邻视效果更佳),其中一个视称为参考视,另一个视称为目标视。FIG. 5 is another embodiment of the present invention, showing a multi-view video coding prediction compensation device 2 . The difference between the device 2 and the
图5的上半部分(上面矩形框包含的部分)是参考视图像的编码过程,下半部分(下面矩形框包含的部分)是目标视图像的编码过程;虚粗线框中包含的是全局运动预测补偿模块和运动向量预测模块,它们作为两种模式加入到整个装置中;The upper part of Figure 5 (the part enclosed by the upper rectangle) is the encoding process of the reference view image, and the lower half (the part enclosed by the lower rectangle box) is the encoding process of the target view image; Motion prediction compensation module and motion vector prediction module, which are added to the whole device as two modes;
参考视图像和目标视图像分别是输入的参考视原始视频图像和目标视原始视频图像;The reference view image and the target view image are the input reference view original video image and the target view original video image respectively;
参考帧缓冲区和目标帧缓冲区分别存储前一帧编码后重构的参考视图像和目标视图像。The reference frame buffer and the target frame buffer respectively store the reconstructed reference view image and target view image after encoding of the previous frame.
它的工作顺序是首先对一帧参考视图像进行编码,然后再对一帧目标视图像进行编码。下面分别是参考视图像编码和目标视图像编码的过程:Its working sequence is to first encode a frame of reference view image, and then encode a frame of target view image. The following are the processes of encoding the reference view image and encoding the target view image respectively:
1、参考视图像编码单元1. Reference view image coding unit
该过程与装置1中的相同,详细叙述参见图4中参考视图像编码说明。This process is the same as that in
2、目标视图像编码单元2. Target view image coding unit
它实现一帧目标视图像的预测补偿编码。It realizes the predictive compensation encoding of a frame of target image.
在输入的目标视原始图像和参考视之间进行全局运动估计,获得全局运动参数,并由此获得全局运动预测代价和全局运动向量,并将全局运动向量保存,以备后用。在全局运动估计模块中完成;The global motion estimation is performed between the input original image of the target view and the reference view to obtain the global motion parameters, and thus obtain the global motion prediction cost and the global motion vector, and save the global motion vector for later use. done in the global motion estimation module;
利用前一帧目标视图像编码时获得的全局运动向量及当前获得的全局运动向量和对应的参考视图像编码时保存的运动向量进行目标视运动向量预测,得到目标视图像参考前一帧重构目标视的局部运动向量,并计算目标视图像的局部运动预测代价。在运动向量预测模块中完成;Use the global motion vector obtained during the encoding of the target view image in the previous frame, the currently obtained global motion vector and the motion vector saved during the encoding of the corresponding reference view image to perform target view motion vector prediction, and obtain the target view image with reference to the previous frame reconstruction The local motion vector of the target view, and calculate the local motion prediction cost of the target view image. done in the motion vector prediction module;
进行其它的预测补偿方法,并由此获得运动向量,将其进行熵编码。计算该种预测补偿方法的预测代价。在目标帧运动估计模块中完成;Perform other prediction and compensation methods, and thus obtain motion vectors, which are entropy coded. Calculate the prediction cost of this prediction compensation method. Done in the target frame motion estimation module;
按照预测代价最小的原则选择合适的预测补偿方法。根据选定的预测补偿方法或者进行全局运动补偿(若是全局运动预测补偿方法被选中,在全局运动补偿模块中完成),或者进行局部运动补偿(若运动向量预测补偿方法被选中,在目标帧运动补偿模块中完成),或者进行其它方式的补偿(在目标帧运动补偿模块中完成),获得补偿后的目标视图像,该图像与原始目标视图像做差,求得残差图像;According to the principle of minimum prediction cost, choose the appropriate prediction compensation method. According to the selected prediction compensation method, either perform global motion compensation (if the global motion prediction compensation method is selected, it will be completed in the global motion compensation module), or perform local motion compensation (if the motion vector prediction compensation method is selected, it will be completed in the target frame motion Completing in the compensation module), or performing compensation in other ways (finishing in the target frame motion compensation module) to obtain a compensated target view image, which is compared with the original target view image to obtain a residual image;
残差图像经过变换、量化和熵编码,最终形成目标视码流。经过量化后的残差图像再经过反量化和反变换,并与补偿后的目标视图像累加,进而得到目标视重构图像,并将其放入目标帧缓冲区,以备下一帧目标视图像编码时使用。The residual image is transformed, quantized and entropy coded to finally form the target video stream. The quantized residual image is dequantized and inversely transformed, and accumulated with the compensated target view image to obtain the target view reconstructed image, which is put into the target frame buffer for the next frame of target view. Used when encoding images.
图6描述了基于MPEG-4时间可扩展编码方法的预测原理:Figure 6 describes the prediction principle based on the MPEG-4 time scalable coding method:
本发明的在两个相关视序列之间采用全局运动估计预测和运动向量预测方法较之现有编码方法可以更充分利用多视角视频中各个视之间的相关性,因此可以有效对多视角视频进行编解码。它们可以与MPEG-4编码标准结合使用,采用如下步骤实现全局运动估计预测和运动向量预测处理(以立体视序列为例,它只有两个视,分别称为左视和右视,因此是最简单的多视角视频序列)。Compared with the existing coding method, the method of the present invention adopting global motion estimation prediction and motion vector prediction between two related view sequences can make full use of the correlation between each view in the multi-view video, so it can effectively analyze the multi-view video Codec. They can be used in combination with the MPEG-4 coding standard, and the following steps are used to realize global motion estimation prediction and motion vector prediction processing (take the stereoscopic view sequence as an example, it has only two views, which are called left view and right view respectively, so it is the most simple multi-view video sequence).
一、编码端处理:1. Encoding end processing:
图6中,左视图像作为参考视序列,右视图像作为当前视序列,GMC0表示前一帧左右图像间的全局运动估计预测,GMC1表示当前帧左右图像间的全局运动估计预测,MV0表示参考视(即左视图像)前后帧之间的运动预测向量,MV1表示当前视(即右视图像)前后帧之间的运动预测向量,它需要通过使用GMC0、GMC1和MV0进行运动向量预测获得。I表示帧内编码,P表示预测编码,B表示双向预测编码。为了与MPEG-4标准兼容,本发明采用增加宏块预测模式的方式嵌入全局运动估计预测和运动向量预测。因此,B图像除了原有的预测模式外,又增加两种预测模式,分别是全局运动预测(使用GME来表示)和运动向量预测(使用MVP来表示)。其具体的实现过程如下:In Figure 6, the left-view image is used as the reference view sequence, and the right-view image is used as the current view sequence. GMC0 represents the global motion estimation and prediction between the left and right images of the previous frame, GMC1 represents the global motion estimation and prediction between the left and right images of the current frame, and MV0 represents the reference MV1 represents the motion prediction vector between frames before and after the current view (that is, the left-view image), and MV1 needs to be obtained by using GMC0, GMC1 and MV0 for motion vector prediction. I represents intra-frame coding, P represents predictive coding, and B represents bidirectional predictive coding. In order to be compatible with the MPEG-4 standard, the present invention embeds global motion estimation prediction and motion vector prediction by adding a macroblock prediction mode. Therefore, in addition to the original prediction mode, two prediction modes are added to the B picture, which are global motion prediction (represented by GME) and motion vector prediction (represented by MVP). Its specific implementation process is as follows:
1.参考视图像编码:这一步与MPEG-4原有步骤相同,并由此获得前后帧之间的运动向量MV0。1. Reference video encoding: This step is the same as the original step of MPEG-4, and thus obtains the motion vector MV0 between the preceding and following frames.
2.当前视图像不仅可以参考参考视图像,还可以参考当前视的前一帧图像,因此这里除了第一帧外使用MPEG-4中的B帧图像来编码实现。这里局部运动估计预测方式与MPEG-4原有步骤相同。2. The image of the current view can not only refer to the image of the reference view, but also refer to the image of the previous frame of the current view. Therefore, except for the first frame, the B frame image in MPEG-4 is used for encoding. Here, the local motion estimation prediction method is the same as the original steps of MPEG-4.
3.当前视图像与参考视图像间进行全局运动估计和补偿,并获得全局运动参数GMC1和宏块的全局运动向量。3. Perform global motion estimation and compensation between the current view image and the reference view image, and obtain the global motion parameter GMC1 and the global motion vector of the macroblock.
4.利用已经编码的MV0、GMC0和GMC1进行当前视前后帧之间的运动向量预测。采用中心点预测的方法,其预测方法见附图2。4. Use the encoded MV0, GMC0 and GMC1 to perform motion vector prediction between the frames before and after the current view. The method of central point prediction is adopted, and its prediction method is shown in Figure 2.
5.模式选择方法:5. Mode selection method:
1)在MPEG-4中的B帧原有的模式之间选出最优的,采用MPEG-4原有的选择策略;1) select the best among the original modes of the B frame in MPEG-4, adopt the original selection strategy of MPEG-4;
2)在选择出的B帧最优的原有模式、GME和MVP模式中通过率失真优化(RDO-rate-distortion optimization)策略选择最优预测模式。选择方法是选择MSE+λRate最小的模式,其中,MSE是宏块原始图像与宏块重构图像的均方差,Rate是该宏块编码后的码位。λ是拉格朗日算子,这里λ=(0.85×2QP/3)1/2,QP是量化系数。2) Select the optimal prediction mode through the rate-distortion optimization (RDO-rate-distortion optimization) strategy in the selected original mode, GME and MVP mode with the best B frame. The selection method is to select the mode with the smallest MSE+λRate, where MSE is the mean square error between the original image of the macroblock and the reconstructed image of the macroblock, and Rate is the encoded code bit of the macroblock. λ is a Lagrangian operator, where λ=(0.85×2QP/3)1/2, and QP is a quantization coefficient.
6.进行运动补偿。GME模式采用全局运动估计补偿,其它的模式均采用MPEG-4原有的补偿方法。6. Perform motion compensation. The GME mode adopts the global motion estimation compensation, and other modes adopt the original compensation method of MPEG-4.
7.形成宏块码流。采用GME和MVP模式的宏块无需编码运动向量,其它的采用MPEG-4原有的方法。7. Form a macroblock code stream. Macroblocks using GME and MVP modes do not need to encode motion vectors, and others use the original method of MPEG-4.
二、解码端处理:2. Decoder processing:
1.解码参考视图像,采用MPEG-4原有的解码方法。1. To decode the reference video image, the original decoding method of MPEG-4 is adopted.
2.解码当前视图像,采用如下步骤:2. To decode the current view image, the following steps are adopted:
1)解码当前帧的全局运动估计参数;1) decoding the global motion estimation parameters of the current frame;
2)若当前宏块是MPEG-4原有预测模式,则按照MPEG-4原有解码方法解码,否则若是GME类型,则进行全局运动补偿重构该宏块,否则若是MVP类型,则按照编码步骤4的方法来预测获得运动向量,然后在利用MPEG-4原有的补偿方法进行运动补偿。2) If the current macroblock is in the original prediction mode of MPEG-4, then decode it according to the original decoding method of MPEG-4, otherwise, if it is the GME type, perform global motion compensation to reconstruct the macroblock, otherwise, if it is the MVP type, then follow the encoding method The method in step 4 is used to predict and obtain the motion vector, and then perform motion compensation using the original compensation method of MPEG-4.
3)解码残差图像,并与补偿的图像合并产生解码图像。3) The residual image is decoded and combined with the compensated image to generate a decoded image.
以上实施例仅用以说明而非限制本发明的技术方案,本领域的普通技术人员应当理解:可以对本发明进行修改或者等同替换,而不脱离本发明的精神和范围的任何修改或局部替换,其均应涵盖在本发明的权利要求范围当中。The above embodiments are only used to illustrate rather than limit the technical solution of the present invention. Those skilled in the art should understand that: the present invention can be modified or equivalently replaced without departing from any modification or partial replacement of the spirit and scope of the present invention. All of them should be included in the scope of the claims of the present invention.
Claims (14)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN 03148031 CN1258925C (en) | 2003-06-27 | 2003-06-27 | Multi-view video coding and decoding prediction compensation method and device |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN 03148031 CN1258925C (en) | 2003-06-27 | 2003-06-27 | Multi-view video coding and decoding prediction compensation method and device |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN1568015A true CN1568015A (en) | 2005-01-19 |
| CN1258925C CN1258925C (en) | 2006-06-07 |
Family
ID=34472161
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN 03148031 Expired - Fee Related CN1258925C (en) | 2003-06-27 | 2003-06-27 | Multi-view video coding and decoding prediction compensation method and device |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN1258925C (en) |
Cited By (22)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN101690220A (en) * | 2007-04-25 | 2010-03-31 | Lg电子株式会社 | Method and apparatus for decoding/encoding video signal |
| CN101222639B (en) * | 2007-01-09 | 2010-04-21 | 华为技术有限公司 | Inter-view prediction method, encoder and decoder in multi-view video technology |
| CN101175210B (en) * | 2006-10-30 | 2010-08-11 | 中国科学院计算技术研究所 | Entropy decoding method and device used for decoding video estimation residual error coefficient |
| CN101248669B (en) * | 2005-09-21 | 2011-01-12 | 三星电子株式会社 | Apparatus and method for encoding and decoding multi-view video |
| CN102036078A (en) * | 2011-01-21 | 2011-04-27 | 哈尔滨商业大学 | Motion estimation method based on relevance among viewing angles for multiple visual angle video coding and decoding system |
| CN101411181B (en) * | 2006-04-06 | 2011-07-06 | 高通股份有限公司 | Electronic video image stabilization device and method |
| CN101529918B (en) * | 2006-10-30 | 2011-08-03 | 日本电信电话株式会社 | Predictive reference information generation method, moving picture encoding and decoding method and device thereof |
| CN102236798A (en) * | 2011-08-01 | 2011-11-09 | 清华大学 | Image matching method and device |
| CN102257817A (en) * | 2008-12-19 | 2011-11-23 | 汤姆森特许公司 | Video coding based on global movement compensation |
| CN101366285B (en) * | 2006-01-09 | 2012-04-18 | 汤姆逊许可证公司 | Method and apparatus for multi-view video coding |
| CN101743750B (en) * | 2007-05-14 | 2012-06-27 | 三星电子株式会社 | Method and apparatus for encoding and decoding multi-view image |
| CN102572418A (en) * | 2010-12-28 | 2012-07-11 | 深圳市融创天下科技股份有限公司 | Motion prediction method and motion prediction system of interframe encoding |
| CN101390396B (en) * | 2006-03-09 | 2012-07-18 | 三星电子株式会社 | Method and apparatus for encoding and decoding multi-view video to provide uniform picture quality |
| CN102868879A (en) * | 2011-07-05 | 2013-01-09 | 北京大学 | Method and system for converting video frame rate |
| CN104380744A (en) * | 2012-06-19 | 2015-02-25 | Lg电子株式会社 | Method and device for processing video signals |
| CN104871541A (en) * | 2012-11-07 | 2015-08-26 | Lg电子株式会社 | Method and apparatus for processing video signals |
| CN105075256A (en) * | 2013-03-27 | 2015-11-18 | 国立研究开发法人情报通信研究机构 | Method and device for encoding multiple input images, and storage medium for storing programs |
| CN105103546A (en) * | 2013-03-27 | 2015-11-25 | 国立研究开发法人情报通信研究机构 | Method and device for encoding multiple input images, and storage medium for storing programs |
| CN109076234A (en) * | 2016-05-24 | 2018-12-21 | 华为技术有限公司 | Image prediction method and relevant device |
| CN110024404A (en) * | 2016-09-30 | 2019-07-16 | 交互数字Vc控股公司 | Method, apparatus and stream for encoding a globally rotated motion compensated image |
| CN113271464A (en) * | 2021-05-11 | 2021-08-17 | 北京奇艺世纪科技有限公司 | Video encoding method, decoding method and related devices |
| CN114697678A (en) * | 2020-12-30 | 2022-07-01 | 中科寒武纪科技股份有限公司 | Image coding method, device, storage medium and device |
-
2003
- 2003-06-27 CN CN 03148031 patent/CN1258925C/en not_active Expired - Fee Related
Cited By (40)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN101248669B (en) * | 2005-09-21 | 2011-01-12 | 三星电子株式会社 | Apparatus and method for encoding and decoding multi-view video |
| US9521429B2 (en) | 2006-01-09 | 2016-12-13 | Thomson Licensing | Methods and apparatus for multi-view video coding |
| CN102685489A (en) * | 2006-01-09 | 2012-09-19 | 汤姆逊许可证公司 | Methods and apparatus for multi-view video coding |
| US10194171B2 (en) | 2006-01-09 | 2019-01-29 | Thomson Licensing | Methods and apparatuses for multi-view video coding |
| US9143782B2 (en) | 2006-01-09 | 2015-09-22 | Thomson Licensing | Methods and apparatus for multi-view video coding |
| US9525888B2 (en) | 2006-01-09 | 2016-12-20 | Thomson Licensing | Methods and apparatus for multi-view video coding |
| US8842729B2 (en) | 2006-01-09 | 2014-09-23 | Thomson Licensing | Methods and apparatuses for multi-view video coding |
| CN102685489B (en) * | 2006-01-09 | 2015-04-29 | 汤姆逊许可证公司 | Methods and apparatus for multi-view video coding |
| CN101366285B (en) * | 2006-01-09 | 2012-04-18 | 汤姆逊许可证公司 | Method and apparatus for multi-view video coding |
| CN101390396B (en) * | 2006-03-09 | 2012-07-18 | 三星电子株式会社 | Method and apparatus for encoding and decoding multi-view video to provide uniform picture quality |
| US8270482B2 (en) | 2006-03-09 | 2012-09-18 | Samsung Electronics Co., Ltd. | Method and apparatus for encoding and decoding multi-view video to provide uniform picture quality |
| CN101411181B (en) * | 2006-04-06 | 2011-07-06 | 高通股份有限公司 | Electronic video image stabilization device and method |
| CN101529918B (en) * | 2006-10-30 | 2011-08-03 | 日本电信电话株式会社 | Predictive reference information generation method, moving picture encoding and decoding method and device thereof |
| CN101175210B (en) * | 2006-10-30 | 2010-08-11 | 中国科学院计算技术研究所 | Entropy decoding method and device used for decoding video estimation residual error coefficient |
| CN101222639B (en) * | 2007-01-09 | 2010-04-21 | 华为技术有限公司 | Inter-view prediction method, encoder and decoder in multi-view video technology |
| CN101690220A (en) * | 2007-04-25 | 2010-03-31 | Lg电子株式会社 | Method and apparatus for decoding/encoding video signal |
| CN101690220B (en) * | 2007-04-25 | 2013-09-25 | Lg电子株式会社 | Method and apparatus for decoding/encoding video signal |
| CN101743750B (en) * | 2007-05-14 | 2012-06-27 | 三星电子株式会社 | Method and apparatus for encoding and decoding multi-view image |
| CN102257817B (en) * | 2008-12-19 | 2014-10-08 | 汤姆森特许公司 | Video coding based on global movement compensation |
| CN102257817A (en) * | 2008-12-19 | 2011-11-23 | 汤姆森特许公司 | Video coding based on global movement compensation |
| CN102572418B (en) * | 2010-12-28 | 2014-09-03 | 深圳市云宙多媒体技术有限公司 | Motion prediction method and motion prediction system of interframe encoding |
| CN102572418A (en) * | 2010-12-28 | 2012-07-11 | 深圳市融创天下科技股份有限公司 | Motion prediction method and motion prediction system of interframe encoding |
| CN102036078B (en) * | 2011-01-21 | 2012-07-25 | 哈尔滨商业大学 | Motion estimation method based on relevance among viewing angles for multiple visual angle video coding and decoding system |
| CN102036078A (en) * | 2011-01-21 | 2011-04-27 | 哈尔滨商业大学 | Motion estimation method based on relevance among viewing angles for multiple visual angle video coding and decoding system |
| CN102868879A (en) * | 2011-07-05 | 2013-01-09 | 北京大学 | Method and system for converting video frame rate |
| CN102868879B (en) * | 2011-07-05 | 2015-04-29 | 北京大学 | Method and system for converting video frame rate |
| CN102236798A (en) * | 2011-08-01 | 2011-11-09 | 清华大学 | Image matching method and device |
| CN104380744B (en) * | 2012-06-19 | 2018-05-04 | Lg电子株式会社 | The method and apparatus for handling vision signal |
| CN104380744A (en) * | 2012-06-19 | 2015-02-25 | Lg电子株式会社 | Method and device for processing video signals |
| US10021417B2 (en) | 2012-06-19 | 2018-07-10 | Lg Electronics Inc. | Method and device for processing video signal |
| CN104871541A (en) * | 2012-11-07 | 2015-08-26 | Lg电子株式会社 | Method and apparatus for processing video signals |
| CN105103546A (en) * | 2013-03-27 | 2015-11-25 | 国立研究开发法人情报通信研究机构 | Method and device for encoding multiple input images, and storage medium for storing programs |
| CN105075256A (en) * | 2013-03-27 | 2015-11-18 | 国立研究开发法人情报通信研究机构 | Method and device for encoding multiple input images, and storage medium for storing programs |
| CN109076234A (en) * | 2016-05-24 | 2018-12-21 | 华为技术有限公司 | Image prediction method and relevant device |
| CN110024404A (en) * | 2016-09-30 | 2019-07-16 | 交互数字Vc控股公司 | Method, apparatus and stream for encoding a globally rotated motion compensated image |
| CN110024404B (en) * | 2016-09-30 | 2023-06-20 | 交互数字Vc控股公司 | Method, apparatus and stream for encoding global rotary motion compensated images |
| US11812066B2 (en) | 2016-09-30 | 2023-11-07 | Interdigital Vc Holdings, Inc. | Methods, devices and stream to encode global rotation motion compensated images |
| CN114697678A (en) * | 2020-12-30 | 2022-07-01 | 中科寒武纪科技股份有限公司 | Image coding method, device, storage medium and device |
| CN113271464A (en) * | 2021-05-11 | 2021-08-17 | 北京奇艺世纪科技有限公司 | Video encoding method, decoding method and related devices |
| CN113271464B (en) * | 2021-05-11 | 2022-11-18 | 北京奇艺世纪科技有限公司 | Video encoding method, decoding method and related devices |
Also Published As
| Publication number | Publication date |
|---|---|
| CN1258925C (en) | 2006-06-07 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN1258925C (en) | Multi-view video coding and decoding prediction compensation method and device | |
| Li et al. | Pseudo-sequence-based 2-D hierarchical coding structure for light-field image compression | |
| Chen et al. | Learning for video compression | |
| CN1099656C (en) | Video compression method and device | |
| CN1774930A (en) | Video transcoding | |
| CN1874519A (en) | Intraprediction method and apparatus using video symmetry and encoding and decoding method and apparatus | |
| CN1757240A (en) | Video encoding | |
| CN1874521A (en) | Apparatus and method for image encoding and decoding | |
| CN101361371B (en) | Video encoding method and decoding method, device and program thereof, and storage medium for recording the program | |
| CN1875637A (en) | Method and apparatus for minimizing number of reference pictures used for inter-coding | |
| CN101036397A (en) | Method and system for decomposing multiview videos | |
| CN1525762A (en) | A Encoder/Decoder Bidirectional Prediction Method for Video Coding | |
| WO2006073116A1 (en) | Video encoding method and device, video decoding method and device, program thereof, and recording medium containing the program | |
| CN1669330A (en) | Motion estimation with weighting prediction | |
| CN101069429A (en) | Method and apparatus for multi-layered video encoding and decoding | |
| CN1929611A (en) | Methods and apparatus for video intraprediction encoding and decoding | |
| CN1497981A (en) | Method and apparatus for encoding and decoding motion vectors | |
| TW200952499A (en) | Apparatus and method for computationally efficient intra prediction in a video coder | |
| CN1658673A (en) | Video compression coding-decoding method | |
| WO2011088594A1 (en) | Video encoder, video decoder, method for video encoding and method for video decoding, separately for each colour plane | |
| CN1926884A (en) | Video encoding method and apparatus | |
| CN1615026A (en) | Data processing device, its method and encoding device | |
| CN1476253A (en) | An Intra-frame Coding Method Using Intra-frame Prediction Based on Prediction Block Group | |
| CN1925619A (en) | Apparatus and method for video encoding and decoding and recording medium | |
| CN1268136C (en) | Frame field adaptive coding method based on image slice structure |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| C06 | Publication | ||
| PB01 | Publication | ||
| C10 | Entry into substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| C14 | Grant of patent or utility model | ||
| GR01 | Patent grant | ||
| ASS | Succession or assignment of patent right |
Owner name: UNITED XINYUAN DIGITAL AUDIO-VIDEO TECHNOLOGY (BE Free format text: FORMER OWNER: INST. OF COMPUTING TECHN. ACADEMIA SINICA Effective date: 20080328 |
|
| C41 | Transfer of patent application or patent right or utility model | ||
| TR01 | Transfer of patent right |
Effective date of registration: 20080328 Address after: Beijing city Haidian District East Road No. 1 Yingchuang power building block A room 701 Patentee after: Lianhe Xinyuan Digit Audio-Frequency Technology Co., Ltd., Beijing Address before: Digital room (Institute of Physics), Institute of computing, Chinese Academy of Sciences, South Road, Zhongguancun, Haidian District, Beijing 6, China Patentee before: Institute of Computing Technology, Chinese Academy of Sciences |
|
| CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20060607 Termination date: 20210627 |
|
| CF01 | Termination of patent right due to non-payment of annual fee |