Dynamic Multi-Reference Generative Prediction For Face Video Compression
Zhao Wang, Bolin Chen, Yan Ye, Shiqi Wang
-
SPS
IEEE Members: $11.00
Non-members: $15.00Length: 00:16:44
Efficient 3D point cloud compression plays critical role in immersive multimedia presentation and autonomous driving. The temporal redundancy between consecutive point cloud frames is obvious, but the exploration of inter-prediction is insufficient in the geometry-based point cloud compression (G-PCC) framework. in inter exploration model (inter-EM) of G-PCC, the reference information can only come from one previous frame and a consistent quantization parameter (QP) value is used for all frames. To perform a more efficient inter-prediction for geometry and attribute, a bidirectional inter-prediction (bi-prediction) scheme is proposed for G-PCC based on inter-EM. With the bi-prediction scheme, the reference information can come from two reference frames. For attribute compression, neighboring search is started from one search center derived by a Morton code distance and a two-threshold method is applied to constrain inter-prediction. Meanwhile, the dependencies between frames are designed according to a hierarchical group of frames (Gof) structure with corresponding hierarchical QP values. Experimental results demonstrate that our method outperforms G-PCC and inter-EM for both lossless and lossy compression. More specifically, the average coding gain over G-PCC is 14.1% on attribute and 11.8% on geometry under lossy condition. The designs on attribute compression have been adopted to the latest inter-EM.