CN116800979B - Point cloud video coding method based on inter-frame implicit correlation - Google Patents
Point cloud video coding method based on inter-frame implicit correlation Download PDFInfo
- Publication number
- CN116800979B CN116800979B CN202310865197.4A CN202310865197A CN116800979B CN 116800979 B CN116800979 B CN 116800979B CN 202310865197 A CN202310865197 A CN 202310865197A CN 116800979 B CN116800979 B CN 116800979B
- Authority
- CN
- China
- Prior art keywords
- frame
- cube
- cubes
- point cloud
- inter
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/172—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/42—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/90—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
- H04N19/91—Entropy coding, e.g. variable length coding [VLC] or arithmetic coding
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
A point cloud video coding method based on inter-frame implicit correlation relates to the field of point cloud video compression and comprises the steps of firstly, performing motion compensation with minimized entropy, generating reference frames by adopting a motion compensation method, enabling the reference frames to align with topological structures in the inter-frame implicit correlation and simultaneously minimizing conditional entropy, dividing the frames into small cubes, evaluating matching degree between the two cubes by using an index, searching cubes with optimal matching degree from cubes of a previous frame for each cube of a current frame, splicing each optimal matching cube, generating the reference frames, selecting the reference frames capable of minimizing the conditional entropy as motion compensation output with minimized entropy, and performing inter-frame entropy coding. The invention fully utilizes the inter-frame redundant information of the dynamic frames to perform lossless compression on the point cloud video, effectively reduces the transmission bandwidth consumption of the point cloud video stream, and simultaneously utilizes the inter-frame implicit correlation to compress the point cloud video, thereby effectively reducing the video data volume.
Description
Technical Field
The invention relates to the technical field of point cloud video compression, in particular to a point cloud video coding method based on inter-frame implicit correlation.
Background
The mainstream point cloud video codec projects the point cloud video to the 2D video and then encodes, or directly encodes the point cloud .V-PCC(Sebastian Schwarz et al.2019.Emerging MPEG Standards for Point Cloud Compression.IEEE Journal on Emerging and Selected Topics in Circuits and Systems 9,1(2019),133–148.) encoder to project the geometry and properties of the point cloud video to the plurality of 2D video tracks .Vues(Yu Liu et al.2022.Vues:Practical Mobile Volumetric Video Streaming through Multiview Transcoding.In Proceedings of the 28th Annual International Conference on Mobile Computing And Networking(Sydney,NSW,Australia)(MobiCom'22).Association for Computing Machinery,New York,NY,USA,514–527.) and transcodes the point cloud video to a 2D video using an edge server. Recently, there has also been work to directly encode point cloud video. Draco derived from google utilizes kd-Tree to encode point cloud .GROOT(Kyungjin Lee et al.2020.GROOT:A Real-Time Streaming System of High-Fidelity Volumetric Videos.In Proceedings of the 26th Annual International Conference on Mobile Computing and Networking(London,United Kingdom)(MobiCom'20).Association for Computing Machinery,New York,NY,USA,Article 57,14pages.), and proposes that parallel octree improves decoding efficiency .YuZu(Anlan Zhang et al.2022.YuZu:Neural-Enhanced Volumetric Video Streaming.In 19th USENIX Symposium on Networked Systems Design and Implementation(NSDI 22).USENIX Association,Renton,WA,137–154.) utilizes 3D super-resolution technology to improve point cloud density .AITransfer(Yakun Huang et al.2021.AITransfer:Progressive AI-Powered Transmission for Real-Time Point Cloud Video Streaming.In Proceedings of the 29th ACM International Conference on Multimedia(Virtual Event,China)(MM'21).Association for Computing Machinery,New York,NY,USA,3989–3997.) and uses an AI model to encode point cloud. In general, these systems encode each frame of point cloud independently without considering inter-frame redundancy information.
In recent years, there have also been some advances .Kammerl(Julius Kammerl et al.2012.Real-time compression of point cloud streams.In 2012 IEEE International Conference on Robotics and Automation.778–785.) in the direction of inter-frame coding of point cloud video, et al, which propose to represent a point cloud frame as a distinction from the previous frame. However, this approach is only effective on point cloud frames that are full of static content. Many studies later compress inter-frame redundancy information using an entropy encoding-based compression method. Most of these methods rely on explicit correlation between frames, i.e. duplicate information of the same or adjacent position between adjacent frames. However, they are not specifically designed for point cloud video streaming and ignore video dynamics.
Disclosure of Invention
The invention aims to provide a point cloud video coding method based on inter-frame implicit correlation.
The technical scheme adopted by the invention for solving the technical problems is as follows:
The invention discloses a point cloud video coding method based on inter-frame implicit correlation, which comprises the following steps:
step S1, motion compensation with minimized entropy;
Step S1.1, voxelizing the point cloud, namely mapping the point cloud into a three-dimensional grid;
Step S1.2, generating a reference frame by adopting a motion compensation method, wherein the reference frame aligns a topological structure in implicit correlation between frames and simultaneously minimizes conditional entropy;
S1.2.1, dividing the frame into small cubes;
Step S1.2.2, evaluating the matching degree between the two cubes by using an index;
Step S1.2.3, searching a cube with the best matching degree from cubes of the previous frame for each cube of the current frame;
Step S1.2.4, splicing each best matching cube to generate a reference frame;
Step S1.2.5, selecting a reference frame capable of minimizing conditional entropy as motion compensation output with minimized entropy:
step S2, inter-frame entropy coding;
The current frame is encoded using the inter-frame entropy encoding algorithm S4D and the reference frame generated in step S1 capable of minimizing conditional entropy as a context.
Further, the specific operation flow of step S1.1 is as follows:
The voxelized point cloud frame is represented by a three-dimensional array, each member, namely the voxel, of the three-dimensional array is empty or occupied, and the definition of the conditional entropy H is as follows:
Wherein P 0 represents the probability that the voxel in the current frame is empty, P 1 represents the probability that the voxel in the current frame is occupied; And Representing the conditional probability that the voxel in the current frame is empty and occupied, respectively, when the voxel in the previous frame is empty; And Representing the conditional probability that the voxel in the current frame is empty and occupied when the voxel in the previous frame is occupied, respectively.
Further, the specific operation flow of step s1.2.1 is as follows:
the current frame and the previous frame are respectively denoted by I t and I t-1, and the current frame I t is divided into cubes which are mutually disjoint Side length is M voxels, each cubeLocated at position (x j,yj,zj).
Further, the specific operation flow of step S1.2.2 is as follows:
to give each cube Finding the best matching cube in the previous frame I t-1 At the same timeAn exhaustive search is performed in a search window of center and side length W voxels, whereinIs the union cube in the previous frame I t-1 The search space being represented as a set of candidate cubesMotion vectors are defined from the best matching cubeDirectional cubeWhen a candidate cubeAnd cubeWhen the matching degree is high, candidate cubes are usedVoxel prediction cube in (a)The corresponding voxels in (a) are used for representing the prediction result by using a confusion matrix, and the prediction result is shown in the following table;
Wherein, Represented in cubesIs occupied, i.e. has a value of 1,Represented in cubesIs empty, i.e. has a value of 0,Expressed in candidate cubesIs occupied, i.e. has a value of 1,Expressed in candidate cubesWherein the voxel is empty, i.e. has a value of 0, true Positive (TP) represents a True case, false Positive (FP) represents a False Positive, FALSE NEGATIVE (FN) represents a False Negative, true Negative (TN) represents a True Negative;
The candidate cube And cubeThe matching degree of (a) is represented by precision or recall, and precision refers to the proportion of true values in voxels predicted to be positive, i.e.Recall that the positive proportion of voxels with positive truth value is predicted to be positiveN (TP) represents the number of occurrences of TP sample, n (FP) represents the number of occurrences of FP sample, n (FN) represents the number of occurrences of FN sample, and the two indexes are balanced by using F-score to evaluate the matching degree F β of the two cubes:
Where β represents the balance coefficient, β balances the accuracy and importance of recall, precision represents the accuracy value, recall represents the recall value.
Further, the specific operation flow of step S1.2.3 is as follows:
At the collection of candidate cubes In, with cubesBest matching cube yielding highest F-scoreAnd k candidate matching indexes F 1,…,Fk are used for searching k best matching cubes for each cube of the current frame respectively.
Further, the specific operation flow of step S1.2.4 is as follows:
for a certain candidate matching index F k, all best matching cubes are obtained Thereafter, a best matching cube is usedReplacement cubeTo generate one reference frame, and k reference frames will be generated since k candidate matching indexes are used.
Further, the specific operation flow of step S1.2.5 is as follows:
A corresponding plurality of reference frames is generated using a series of beta values, 0< beta 1<…<βk < + > infinity, and calculating the conditional entropy corresponding to each reference frame And selecting a beta k value corresponding to the minimum conditional entropy, wherein a reference frame corresponding to the beta k value is the motion compensation output with minimum entropy.
The beneficial effects of the invention are as follows:
The point cloud video coding method based on the inter-frame implicit correlation fully utilizes the inter-frame redundant information of the dynamic frames to perform lossless compression on the point cloud video, effectively reduces the consumption of the streaming bandwidth of the point cloud video, and simultaneously utilizes the inter-frame implicit correlation to compress the point cloud video. Compared with the prior art, the method has obvious advantages in the point cloud video compression performance.
Detailed Description
The invention discovers that the inter-frame explicit correlation is obviously reduced on frames with stronger dynamics. Therefore, the invention finds an inter-frame implicit correlation, namely the consistency of the topological structure between adjacent frames. The present invention finds that the inter-frame implicit correlation is maintained at a relatively high level even on frames with strong dynamics. Thus, inter-frame implicit correlation has a high potential to help compress point cloud video. In order to fully utilize the implicit correlation between frames to compress the point cloud video, the invention adopts widely used entropy coding as a basic encoder model, and can use a reference frame as auxiliary information, and in the prior art, the last frame is directly used as the reference frame, so that the coding effect is poor. Meanwhile, smaller inter-frame conditional entropy can provide a higher theoretical upper limit of compression rate, and simply aligning adjacent frames using the existing motion estimation method cannot effectively reduce the inter-frame conditional entropy.
Therefore, the invention provides a point cloud video coding method based on inter-frame implicit correlation, which specifically comprises the following steps:
step S1, motion compensation with minimized entropy;
The goal of this step is to generate a reference frame that effectively increases the compression rate, and the effectiveness of the compression can be measured by the conditional entropy defined on the reference frame and the current frame. The specific operation steps are as follows:
s1.1, carrying out point cloud voxelization;
The point cloud is first voxelized, i.e., mapped into a three-dimensional grid. The voxelized point cloud frame is represented using a three-dimensional array, each member (voxel) of which is either empty (represented by 0) or occupied (represented by 1). The conditional entropy H is defined as follows:
Wherein P 0 represents the probability that the voxel in the current frame is empty, P 1 represents the probability that the voxel in the current frame is occupied; And Representing the conditional probability that the voxel in the current frame is empty and occupied, respectively, when the voxel in the previous frame is empty; And Representing the conditional probability that the voxel in the current frame is empty and occupied when the voxel in the previous frame is occupied, respectively.
Step S1.2, generating a reference frame by adopting a motion compensation method, wherein the reference frame aligns a topological structure in the implicit correlation between frames and simultaneously minimizes a conditional entropy H, and the method specifically comprises the following 5 steps:
S1.2.1, dividing the frame into small cubes;
the current frame and the previous frame are respectively denoted by I t and I t-1, and the current frame I t is divided into cubes which are mutually disjoint Side length is M voxels, each cubeAt position (x j,yj,zj), dividing the previous frame into cubes which are mutually disjoint in the same mannerCube bodyLocated at position (x j,yj,zj).
Step S1.2.2, evaluating the matching degree between the two cubes by using an index;
to give each cube Finding the best matching cube in the previous frame I t-1 At the same timeAn exhaustive search is performed in a search window of center and side length W voxels, whereinIs the union cube in the previous frame I t-1 The search space being represented as a set of candidate cubesMotion vectors are defined from the best matching cubeDirectional cube
When a candidate cubeAnd cubeWhen the matching degree is high, candidate cubes can be usedVoxel prediction cube in (a)Corresponding voxels of (b) a corresponding voxel of (c). Considering this process as binary prediction, since each voxel may be either 1 (Positive) or 0 (Negative), a confusion matrix is used to represent the prediction result, as shown in table 1.
TABLE 1
Wherein, Represented in cubesIs occupied, i.e. has a value of 1,Represented in cubesIs empty, i.e. has a value of 0,Expressed in candidate cubesIs occupied, i.e. has a value of 1,Expressed in candidate cubesIn (3) is empty, i.e. the value is 0, true Positive (TP) represents a True case, false Positive (FP) represents a False Positive, FALSE NEGATIVE (FN) represents a False Negative, and True Negative (TN) represents a True Negative.
N (·) is used to represent the number of occurrences of a particular element in the confusion matrix. Two cubes (candidate cubes)And cube) The degree of matching of (c) may be expressed in terms of precision or recall. Specifically, precision refers to the proportion of positive values in voxels predicted to be positive, i.eWhereas recall indicates that the positive proportion of voxels with positive true values is predicted, i.eWhere, n (FP) represents the number of occurrences of the FP sample, and n (FN) represents the number of occurrences of the FN sample. Because the two indexes are not uniform, the matching degree F β of the two cubes is better evaluated by using the F-fraction to balance the two indexes:
Where β represents the balance coefficient, β balances the accuracy and importance of recall, precision represents the accuracy value, recall represents the recall value.
Step S1.2.3, searching a cube with the best matching degree from cubes of the previous frame for each cube of the current frame;
At the collection of candidate cubes In, with cubesBest matching cube yielding highest F-scoreIs considered as the best match result. However, it is difficult to consider minimizing conditional entropy when searching for the best matching cube, as conditional entropy depends on the entire reference frame. To this end, the present invention uses k candidate matching indexes F 1,…,Fk to search k best matching cubes for each cube of the current frame, respectively.
Step S1.2.4, splicing each best matching cube to generate a reference frame;
Since step S1.2.3 uses k candidate matching indices, k reference frames will be generated. For a certain candidate matching index F k, all best matching cubes are obtained Thereafter, a best matching cube is usedReplacement cubeTo generate a reference frame.
Step S1.2.5, selecting a reference frame capable of minimizing conditional entropy;
The conditional entropy depends on the index of the reference frame and its corresponding degree of matching F β. However, in generating the reference frame, it is not determined which β value corresponds to the reference frame that minimizes conditional entropy. Therefore, the present invention simultaneously generates a plurality of corresponding reference frames using a plurality of beta values, and calculates a conditional entropy corresponding to each reference frame. The conditional entropy of the reference frame corresponding to the index using the degree of matching F β is denoted by H β. A series of beta values are used which, 0< beta 1<…<βk < + > infinity, then calculate the corresponding conditional entropy Corresponding conditional entropyThe smallest β k value is selected, then the corresponding reference frame is the entropy minimized motion compensated output.
Step S2, inter-frame entropy coding;
The current frame is encoded using the inter-frame entropy encoding algorithm (S4D) and the reference frame generated in step S1 capable of minimizing conditional entropy as a context. The specific operation steps are as follows:
S4D encodes each element of the point cloud three-dimensional array using context-adaptive binary arithmetic coding (CABAC). Specifically, for any element value (0 or 1) in the three-dimensional array of the current frame, S4D uses the element value of the same position in the three-dimensional array of the reference frame as the context of CABAC encoding.
In the prior art, the previous frame is directly used as a reference frame, only the explicit correlation (the same position between adjacent frames or the repeated information of the adjacent positions) between frames is utilized, the compression efficiency is poor on the frames with higher dynamic frames, and the compression performance is poor. The invention specially generates the reference frame corresponding to the minimized conditional entropy through the step S1, is very beneficial to the inter-frame entropy coding, effectively utilizes the inter-frame implicit correlation, and can effectively improve the compression rate on the dynamic frame. The prior art uses a plurality of voxels of the current frame and the reference frame as probability conditions in entropy coding, resulting in a high computational complexity, resulting in a decoded frame rate of less than 1FPS. The invention only uses the left adjacent voxel of the coded voxel and the voxel at the same position in the reference frame as probability conditions in entropy coding, thereby reducing the computational complexity. Thus, the present invention is more applicable to mobile streaming systems.
In order to verify the effect of the point cloud video coding method based on the inter-frame implicit correlation, the three public datasets Ricardo, pizza and Longdress are used for experimental comparison between the prior art and the method. The results show that the present invention reduces bandwidth consumption by 23.15%, 1.06% and 43.32% over three data sets, respectively, compared to existing inter-frame encoders that utilize explicit inter-frame correlation. Compared with the prior art, the method has obvious advantages in the point cloud video compression performance.
The foregoing is merely a preferred embodiment of the present invention and it should be noted that modifications and adaptations to those skilled in the art may be made without departing from the principles of the present invention, which are intended to be comprehended within the scope of the present invention.
Claims (7)
1. The point cloud video coding method based on the inter-frame implicit correlation is characterized by comprising the following steps of:
step S1, motion compensation with minimized entropy;
Step S1.1, voxelizing the point cloud, namely mapping the point cloud into a three-dimensional grid;
Step S1.2, generating a reference frame by adopting a motion compensation method, wherein the reference frame aligns a topological structure in implicit correlation between frames and simultaneously minimizes conditional entropy;
S1.2.1, dividing the frame into small cubes;
Step S1.2.2, evaluating the matching degree between the two cubes by using an index;
Step S1.2.3, searching a cube with the best matching degree from cubes of the previous frame for each cube of the current frame;
Step S1.2.4, splicing each best matching cube to generate a reference frame;
step S1.2.5, selecting a reference frame capable of minimizing conditional entropy as motion compensation output with minimized entropy;
step S2, inter-frame entropy coding;
The current frame is encoded using the inter-frame entropy encoding algorithm S4D and the reference frame generated in step S1 capable of minimizing conditional entropy as a context.
2. The method for point cloud video coding based on inter-frame implicit correlation according to claim 1, wherein the specific operation flow of step S1.1 is as follows:
The voxelized point cloud frame is represented by a three-dimensional array, each member, namely the voxel, of the three-dimensional array is empty or occupied, and the definition of the conditional entropy H is as follows:
Wherein P 0 represents the probability that the voxel in the current frame is empty, P 1 represents the probability that the voxel in the current frame is occupied; And Representing the conditional probability that the voxel in the current frame is empty and occupied, respectively, when the voxel in the previous frame is empty; And Representing the conditional probability that the voxel in the current frame is empty and occupied when the voxel in the previous frame is occupied, respectively.
3. The method for point cloud video coding based on inter-frame implicit correlation according to claim 1, wherein the specific operation flow of step s1.2.1 is as follows:
the current frame and the previous frame are respectively denoted by I t and I t-1, and the current frame I t is divided into cubes which are mutually disjoint Side length is M voxels, each cubeLocated at position (x j,yj,zj).
4. The method for point cloud video encoding based on inter-frame implicit correlation of claim 3, wherein the specific operation procedure of step S1.2.2 is as follows:
to give each cube Finding the best matching cube in the previous frame I t-1 At the same timeAn exhaustive search is performed in a search window of center and side length W voxels, whereinIs the union cube in the previous frame I t-1 The search space being represented as a set of candidate cubesMotion vectors are defined from the best matching cubeDirectional cubeWhen a candidate cubeAnd cubeWhen the matching degree is high, candidate cubes are usedVoxel prediction cube in (a)The corresponding voxels in (a) are used for representing the prediction result by using a confusion matrix, and the prediction result is shown in the following table;
wherein, positive in Represented in cubesIs occupied, i.e. has a value of 1, negative inRepresented in cubesIn which the voxel is empty, i.e. has a value of 0, positive inExpressed in candidate cubesIs occupied, i.e. has a value of 1, negative inExpressed in candidate cubesWherein the voxel is empty, i.e. has a value of 0, true Positive (TP) represents a True case, false Positive (FP) represents a False Positive, FALSE NEGATIVE (FN) represents a False Negative, true Negative (TN) represents a True Negative;
The candidate cube And cubeThe matching degree of (a) is represented by precision or recall, and precision refers to the proportion of true values in voxels predicted to be positive, i.e.Recall that the positive proportion of voxels with positive truth value is predicted to be positiveN (TP) represents the number of occurrences of TP sample, n (FP) represents the number of occurrences of FP sample, n (FN) represents the number of occurrences of FN sample, and the two indexes are balanced by using F-score to evaluate the matching degree F β of the two cubes:
Where β represents the balance coefficient, β balances the accuracy and importance of recall, precision represents the accuracy value, recall represents the recall value.
5. The method for point cloud video encoding based on inter-frame implicit correlation of claim 4, wherein the specific operation procedure of step S1.2.3 is as follows:
At the collection of candidate cubes In, with cubesBest matching cube yielding highest F-scoreAnd k candidate matching indexes F 1,…,Fk are used for searching k best matching cubes for each cube of the current frame respectively.
6. The method for point cloud video encoding based on inter-frame implicit correlation of claim 5, wherein the specific operation flow of step S1.2.4 is as follows:
for a certain candidate matching index F k, all best matching cubes are obtained Thereafter, a best matching cube is usedReplacement cubeTo generate one reference frame, and k reference frames will be generated since k candidate matching indexes are used.
7. The method for point cloud video encoding based on inter-frame implicit correlation of claim 6, wherein the specific operation procedure of step S1.2.5 is as follows:
A corresponding plurality of reference frames is generated using a series of beta values, 0< beta 1<…<βk < + > infinity, and calculating the conditional entropy corresponding to each reference frame And selecting a beta k value corresponding to the minimum conditional entropy, wherein a reference frame corresponding to the beta k value is the motion compensation output with minimum entropy.
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202310865197.4A CN116800979B (en) | 2023-07-14 | 2023-07-14 | Point cloud video coding method based on inter-frame implicit correlation |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202310865197.4A CN116800979B (en) | 2023-07-14 | 2023-07-14 | Point cloud video coding method based on inter-frame implicit correlation |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN116800979A CN116800979A (en) | 2023-09-22 |
| CN116800979B true CN116800979B (en) | 2025-05-27 |
Family
ID=88049693
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN202310865197.4A Active CN116800979B (en) | 2023-07-14 | 2023-07-14 | Point cloud video coding method based on inter-frame implicit correlation |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN116800979B (en) |
Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN112565764A (en) * | 2020-12-03 | 2021-03-26 | 西安电子科技大学 | Point cloud geometric information interframe coding and decoding method |
| CN113766228A (en) * | 2020-06-05 | 2021-12-07 | Oppo广东移动通信有限公司 | Point cloud compression method, encoder, decoder and storage medium |
Family Cites Families (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US11297346B2 (en) * | 2016-05-28 | 2022-04-05 | Microsoft Technology Licensing, Llc | Motion-compensated compression of dynamic voxelized point clouds |
| US11375194B2 (en) * | 2019-11-16 | 2022-06-28 | Uatc, Llc | Conditional entropy coding for efficient video compression |
| CN114095735B (en) * | 2020-08-24 | 2025-09-12 | 北京大学深圳研究生院 | A point cloud geometry inter-frame prediction method based on block motion estimation and motion compensation |
| CN116016953A (en) * | 2022-12-27 | 2023-04-25 | 中山大学·深圳 | A dynamic point cloud attribute compression method based on deep entropy coding |
-
2023
- 2023-07-14 CN CN202310865197.4A patent/CN116800979B/en active Active
Patent Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN113766228A (en) * | 2020-06-05 | 2021-12-07 | Oppo广东移动通信有限公司 | Point cloud compression method, encoder, decoder and storage medium |
| CN112565764A (en) * | 2020-12-03 | 2021-03-26 | 西安电子科技大学 | Point cloud geometric information interframe coding and decoding method |
Also Published As
| Publication number | Publication date |
|---|---|
| CN116800979A (en) | 2023-09-22 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN114095735B (en) | A point cloud geometry inter-frame prediction method based on block motion estimation and motion compensation | |
| US11769275B2 (en) | Method and device for predictive encoding/decoding of a point cloud | |
| CN1183767C (en) | Video encoding method based on the matching pursuit algorithm | |
| CN101610413B (en) | Video coding/decoding method and device | |
| CN100471278C (en) | A Multi-View Video Compression Codec Method Based on Distributed Source Coding | |
| CN116437089B (en) | A Deep Video Compression Method Based on Key Object | |
| CN103379333A (en) | Encoding and decoding method, encoding and decoding of video sequence code streams and device corresponding to methods | |
| Nguyen et al. | Deep probabilistic model for lossless scalable point cloud attribute compression | |
| CN116016951A (en) | Point cloud processing method, device, equipment and storage medium | |
| Zhai et al. | Disparity-based stereo image compression with aligned cross-view priors | |
| WO2022247705A1 (en) | Prediction coding and decoding method and apparatus for point cloud attribute information | |
| CN102833536A (en) | Distributed video encoding and decoding method facing to wireless sensor network | |
| Wang et al. | Bandwidth-efficient mobile volumetric video streaming by exploiting inter-frame correlation | |
| CN101980536A (en) | A Multi-eye Stereoscopic Video Compression Codec Method Based on Object and Fractal | |
| CN114025146A (en) | Dynamic point cloud geometric compression method based on scene flow network and time entropy model | |
| Rudolph et al. | Progressive coding for deep learning based point cloud attribute compression | |
| Tian et al. | Towards real-time neural video codec for cross-platform application using calibration information | |
| CN116800979B (en) | Point cloud video coding method based on inter-frame implicit correlation | |
| WO2023093377A1 (en) | Encoding method, decoding method and electronic device | |
| CN119341575B (en) | A fractal data compression and storage method, system, and related equipment | |
| Yu et al. | Hierarchical distortion learning for fast lossy compression of point clouds | |
| CN102263952B (en) | Quick fractal compression and decompression method for binocular stereo video based on object | |
| Ma et al. | Rate-distortion Optimized Point Cloud Preprocessing for Geometry-based Point Cloud Compression | |
| Hoang et al. | KeyNode-Driven Geometry Coding for Real-World Scanned Human Dynamic Mesh Compression | |
| Fei et al. | Review of distributed video coding |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant | ||
| GR01 | Patent grant |