TY - GEN
T1 - A Comparative Study of K-Planes vs. V-PCC for 6-DoF Volumetric Video Representation
AU - Li, Na
AU - Zhu, Mufeng
AU - Wang, Shuoqian
AU - Liu, Yao
N1 - Publisher Copyright:
© 2024 ACM.
PY - 2024/4/15
Y1 - 2024/4/15
N2 - With NeRF, neural scene representations have gained increased popularity in recent years. To date, many models have been designed to represent dynamic scenes that can be explored in 6 degrees-of-freedom (6-DoF) in immersive applications such as virtual reality (VR), augmented reality (AR), and mixed reality (MR). In this paper, we aim to evaluate how newer neural representations of 6-DoF video compare with more-traditional point cloud-based representations in terms of their representation and transmission efficiency. We design a new methodology for fair comparison between K-Planes, anew dynamic neural scene representation model, and video-based point cloud compression (V-PCC). We conduct extensive experiments using three datasets with a total of 11 sequences with different characteristics. Results show that the current K-Planes models excel for moderately dynamic content, but struggle with highly dynamic scenes. In addition, in emulated volumetric data capture scenarios, the recorded point cloud data can be highly noisy, and the visual quality of views rendered by trained K-Planes models are significantly better than V-PCC.
AB - With NeRF, neural scene representations have gained increased popularity in recent years. To date, many models have been designed to represent dynamic scenes that can be explored in 6 degrees-of-freedom (6-DoF) in immersive applications such as virtual reality (VR), augmented reality (AR), and mixed reality (MR). In this paper, we aim to evaluate how newer neural representations of 6-DoF video compare with more-traditional point cloud-based representations in terms of their representation and transmission efficiency. We design a new methodology for fair comparison between K-Planes, anew dynamic neural scene representation model, and video-based point cloud compression (V-PCC). We conduct extensive experiments using three datasets with a total of 11 sequences with different characteristics. Results show that the current K-Planes models excel for moderately dynamic content, but struggle with highly dynamic scenes. In addition, in emulated volumetric data capture scenarios, the recorded point cloud data can be highly noisy, and the visual quality of views rendered by trained K-Planes models are significantly better than V-PCC.
KW - 6-DoF
KW - neural scene representations
KW - point cloud
KW - volumetric videos
UR - http://www.scopus.com/inward/record.url?scp=85191439010&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85191439010&partnerID=8YFLogxK
U2 - 10.1145/3652212.3652227
DO - 10.1145/3652212.3652227
M3 - Conference contribution
AN - SCOPUS:85191439010
T3 - MMVE 2024 - Proceedings of the 2024 16th International Workshop on Immersive Mixed and Virtual Environment Systems
SP - 92
EP - 98
BT - MMVE 2024 - Proceedings of the 2024 16th International Workshop on Immersive Mixed and Virtual Environment Systems
PB - Association for Computing Machinery, Inc
T2 - 16th International Workshop on Immersive Mixed and Virtual Environment Systems, MMVE 2024
Y2 - 15 April 2024 through 18 April 2024
ER -