**Abstract — ** A universal multiscale conditional coding framework, Unicorn, is proposed to compress the geometry and attribute of any given point cloud. Geometry compression is addressed in Part I of this paper, while attribute compression is discussed in Part II.
For geometry compression, we construct the multiscale sparse tensors of each voxelized point cloud frame and properly leverage lower-scale priors in the current and (previously processed) temporal reference frames to improve the conditional probability approximation or content-aware predictive reconstruction of geometry occupancy in compression.
For attribute compression, Since attribute components exhibit very different intrinsic characteristics from the geometry element, e.g., 8-bit RGB color versus 1-bit occupancy, we process the attribute residual between lower-scale reconstruction and current-scale data. Similarly, we leverage spatially lower-scale priors in the current frame and (previously processed) temporal reference frame to improve the probability estimation of attribute intensity through conditional residual prediction in lossless mode or enhance the attribute reconstruction through progressive residual refinement in lossy mode for better performance.
The proposed Unicorn is a versatile, learning-based solution capable of compressing static and dynamic point clouds with diverse source characteristics in both lossy and lossless modes. Following the same evaluation criteria, Unicorn significantly outperforms standard-compliant approaches like MPEG G-PCC, V-PCC, and other learning-based solutions, yielding state-of-the-art compression efficiency while presenting affordable complexity for practical implementations.
For more information, please visit our homepage: https://njuvision.github.io/Unicorn/
- 2024.12.06 Open source Unicorn Pre (SparsePCGC)!
- 2024.10.28 Unicorn version 2 has responded to the Call for Proposals for AI-based Point Cloud Coding (m70061 & m70062 in MPEG).
- 2024.10.05 Initial release of part of the code and results. (The entire source code will be released to the public after the approval from the funding agency.)
- 2024.09.12 Unicorn version 1 was accepted by TPAMI. (https://ieeexplore.ieee.org/document/10682571 and https://ieeexplore.ieee.org/document/10682566)
- pytorch, MinkowskiEngine, etc.
- You can use docker to simply configure the environment:
docker pull jianqiang1995/pytorch:1.10.0-cuda11.1-cudnn8-devel
- You can use docker to simply configure the environment:
- ShapeNet: https://shapenet.org/
- RWTT: https://texturedmesh.isti.cnr.it/
- MPEG Dataset (Static Objects): http://mpegfs.int-evry.fr/MPEG/PCC/DataSets/pointCloud/CfP/datasets/ (MPEG password is required) (You can also access some of them on our NJU BOX. ( https://box.nju.edu.cn/d/51327ae7c2644c0fa1c4/ ))
- MPEG Dataset (Dynamic Objects): https://mpeg-pcc.org/index.php/pcc-content-database/
- KITTI: https://www.cvlibs.net/datasets/kitti/
- Ford: https://mpegfs.int-evry.fr/ws-mpegcontent/MPEG-I/Part05-PointCloudCompression/dataSets_new/Dynamic_Acquisition/Ford (MPEG password is required) (You can also access some of them on our NJU BOX. ( https://box.nju.edu.cn/d/2739fe997265478c8673/ ))
(Note: The training dataset generation methods and the amount of training dataset are not required to be fixed. We provide some examples in data_utils/datasets/READMe.sh
to show how to perform sampling, partition, quantization, and other operations on raw mesh or point cloud data to generate the training datasets.)
./results
These files are provided by Nanjing University Vision Lab. Thanks to Prof. Dandan Ding from Hangzhou Normal University and Prof. Yi Lin from Fudan University for their help. Please contact us ([email protected]) if you have any questions.