Volume 50 Issue 2
Feb.  2024
Turn off MathJax
Article Contents
LIU G X,ZHANG J T,DING D D. Lossy point cloud geometry compression based on Transformer[J]. Journal of Beijing University of Aeronautics and Astronautics,2024,50(2):634-642 (in Chinese) doi: 10.13700/j.bh.1001-5965.2022.0412
Citation: LIU G X,ZHANG J T,DING D D. Lossy point cloud geometry compression based on Transformer[J]. Journal of Beijing University of Aeronautics and Astronautics,2024,50(2):634-642 (in Chinese) doi: 10.13700/j.bh.1001-5965.2022.0412

Lossy point cloud geometry compression based on Transformer

doi: 10.13700/j.bh.1001-5965.2022.0412
Funds:  National Natural Science Foundation of China (6217010731); Zhejiang Provincial Nature Fund (LY20F010013)
More Information
  • Corresponding author: E-mail:DandanDing@hznu.edu.cn
  • Received Date: 26 May 2022
  • Accepted Date: 02 Jul 2022
  • Available Online: 07 Nov 2022
  • Publish Date: 04 Nov 2022
  • Point clouds are widely used for 3D object representation, however, real-world captured point clouds often have huge data, which is unfavorable for transmission and storage. To address the redundancy problem of point cloud data, an end-to-end Transformer-based multiscale point cloud geometry compression method is proposed by introducing the Transformer module based on the attention mechanism. The point cloud is voxelized, features are extracted using sparse convolution at the encoder, multi-scale gradual downsampling is performed, and the Transformer module is combined to enhance the point-space feature perception and extraction; at the decoder, the corresponding multi-scale up-sampling is performed for reconstruction, and the Transformer module is also used to enhance and recover the useful features, and the point cloud is progressively refined and reconstructed. Compared with two standard point cloud coding methods, the proposed method obtains 80% and 75% BD-Rate gain on average; compared with the deep learning-based point cloud compression method, it obtains 16% BD-Rate gain on average, and there is about 0.6 PSNR enhancement at the same bit rate. The experimental results demonstrate the feasibility and effectiveness of Transformer in the field of point cloud compression. In terms of subjective quality, the proposed method also has significant subjective effect improvement, and the reconstructed point cloud is closer to the original point cloud.

     

  • loading
  • [1]
    SCHWARZ S, PREDA M, BARONCINI V, et al. Emerging MPEG standards for point cloud compression[J]. IEEE Journal on Emerging and Selected Topics in Circuits and Systems, 2019, 9(1): 133-148. doi: 10.1109/JETCAS.2018.2885981
    [2]
    GUARDA A F R, RODRIGUES N M M, PEREIRA F. Point cloud coding: Adopting a deep learning-based approach[C]//Proceedings of the Picture Coding Symposium. Piscataway: IEEE Press, 2019: 1-5.
    [3]
    WANG J, ZHU H, LIU H J, et al. Lossy Point Cloud Geometry Compression via End-to-End Learning[J]. IEEE Transactions on Circuits and Systems for Video Technology, 2021, 31(12): 4909-4923. doi: 10.1109/TCSVT.2021.3051377
    [4]
    YAN W, SHAO Y T, LIU S, et al. Deep autoencoder-based lossy geometry compression for point clouds[J/OL]. Computer Vision and Pattern Recognition, 2019. (2019-4-18)[2022-1-22]. https://doi.org/10.48550/arXiv.1905.03691.
    [5]
    WEN X Z, WANG X, HOU J H, et al. Lossy geometry compression of 3d point cloud data via an adaptive octree-guided network[C]//Proceedings of the IEEE International Conference on Multimedia and Expo. Piscataway: IEEE Press, 2020: 1-6.
    [6]
    WANG J Q, DING D D, LI Z, et al. Multiscale point cloud geometry compression[C]//Proceedings of the Data Compression Conference. Piscataway: IEEE Press, 2021: 73-82.
    [7]
    ZHAO H S, JIANG L, JIA J Y, et al. Point Transformer[C]//Proceedings of the IEEE/CVF International Conference on Computer Vision. Piscataway: IEEE Press, 2021: 16239-16248.
    [8]
    GUO M H, CAI J X, LIU Z N, et al. PCT: Point cloud Transformer[J]. Computional Visual Media, 2021, 7(2): 187-199.
    [9]
    HUANG L L, WANG S L, WONG K, et al. Octsqueeze: Octree-structured entropy model for lidar compression[C]// Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE Press, 2020: 1310-1320.
    [10]
    QUE Z Z, LU G, XU D. Voxelcontext-net: An Octree based framework for point cloud compression[C]// Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE Press, 2021: 6038-6047.
    [11]
    NGUYEN D, QUACH M, VALENZISE G, et al. Lossless coding of point cloud geometry using a deep generative model[J]. IEEE Transactions on Circuits and Systems for Video Technology, 2021, 31(12): 4617-4629. doi: 10.1109/TCSVT.2021.3100279
    [12]
    NGUYEN D, QUACH M, VALENZISE G, et al. Multiscale deep context modeling for lossless point cloud geometry compression[C]//Proceedings of the IEEE International Conference on Multimedia & Expo Workshops. Piscataway: IEEE Press, 2021: 1-6.
    [13]
    HUANG T X, LIU Y. 3D point cloud geometry compression on deep learning[C]//Proceedings of the 27th ACM International Conference on Multimedia. New York: ACM, 2019: 890-898.
    [14]
    QUACH M, VALENZISE G, DUFAUX F. Learning convolutional Transforms for lossy point cloud geometry compression[C]// Proceedings of the IEEE International Conference on Image Processing. Piscataway: IEEE Press, 2019: 4320-4324.
    [15]
    VASWANI A, SHAZEER N, PARMAR N, et al. Attention is all you need[C]// Proceedings of the 31st International Conference on Neural Information Processing Systems. New York: ACM, 2017: 6000-6010.
    [16]
    DOSOVITSKIY A, BEYER L, KOLESNIKOV A, et al. An image is worth 16×16 words: Transformers for image recognition at scale[J/OL]. Computer Vision and Pattern Recognition, 2021. (2021-6-3)[2022-1-22]. https://doi.org/10.48550/arXiv.2010.11929.
    [17]
    PARMAR N, VASWANI A , USZKOREIT J, et al. Image transformer[EB/OL]. (2018-6-15)[2022-1-22]. https://arxiv.org/abs/1802.05751.
    [18]
    LIU Z, LIN Y T, CAO Y, et al. Swin Transformer: Hierarchical vision transformer using shifted windows[C]//Proceedings of the International Conference on Computer Vision. Piscataway: IEEE Press, 2021: 10012-10022.
    [19]
    QIU S, ANWARS, BARNES N. PU-Transformer: Point cloud upsampling transformer[C]//Proceedings of the Asian Conference on Computer Vision. Berlin: Springer, 2022: 2475-2493.
    [20]
    FU C Y, LI G, SONG R, et al. OctAttentioN: Octree-based large-scale contexts model for point cloud compression[C]//Prodeedings of the 36th AAAI conference on Artificial Intelligence. Palo Alto: AAAI Press, 2022: 625-633.
  • 加载中

Catalog

    通讯作者: 陈斌, bchen63@163.com
    • 1. 

      沈阳化工大学材料科学与工程学院 沈阳 110142

    1. 本站搜索
    2. 百度学术搜索
    3. 万方数据库搜索
    4. CNKI搜索

    Figures(5)  / Tables(5)

    Article Metrics

    Article views(977) PDF downloads(16) Cited by()
    Proportional views
    Related

    /

    DownLoad:  Full-Size Img  PowerPoint
    Return
    Return