1
|
Tohidi F, Paul M, Ulhaq A, Chakraborty S. Improved Video-Based Point Cloud Compression via Segmentation. SENSORS (BASEL, SWITZERLAND) 2024; 24:4285. [PMID: 39001064 PMCID: PMC11243880 DOI: 10.3390/s24134285] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/23/2024] [Revised: 06/19/2024] [Accepted: 06/28/2024] [Indexed: 07/16/2024]
Abstract
A point cloud is a representation of objects or scenes utilising unordered points comprising 3D positions and attributes. The ability of point clouds to mimic natural forms has gained significant attention from diverse applied fields, such as virtual reality and augmented reality. However, the point cloud, especially those representing dynamic scenes or objects in motion, must be compressed efficiently due to its huge data volume. The latest video-based point cloud compression (V-PCC) standard for dynamic point clouds divides the 3D point cloud into many patches using computationally expensive normal estimation, segmentation, and refinement. The patches are projected onto a 2D plane to apply existing video coding techniques. This process often results in losing proximity information and some original points. This loss induces artefacts that adversely affect user perception. The proposed method segments dynamic point clouds based on shape similarity and occlusion before patch generation. This segmentation strategy helps maintain the points' proximity and retain more original points by exploiting the density and occlusion of the points. The experimental results establish that the proposed method significantly outperforms the V-PCC standard and other relevant methods regarding rate-distortion performance and subjective quality testing for both geometric and texture data of several benchmark video sequences.
Collapse
Affiliation(s)
- Faranak Tohidi
- School of Computing Mathematics and Engineering, Charles Sturt University, Bathurst, NSW 2795, Australia
| | - Manoranjan Paul
- School of Computing Mathematics and Engineering, Charles Sturt University, Bathurst, NSW 2795, Australia
| | - Anwaar Ulhaq
- School of Engineering and Technology, Centre for Intelligent Systems, Central Queensland University, Sydney Campus, Rockhampton, QLD 4701, Australia
| | - Subrata Chakraborty
- Faculty of Science, Agriculture, Business and Law, University of New England, Armidale, NSW 2351, Australia
- Centre for Advanced Modelling and Geospatial Information Systems (CAMGIS), Faculty of Engineering and Information Technology, University of Technology Sydney, Ultimo, NSW 2007, Australia
- Griffith Business School, Griffith University, Brisbane, QLD 4111, Australia
| |
Collapse
|
2
|
Roriz R, Silva H, Dias F, Gomes T. A Survey on Data Compression Techniques for Automotive LiDAR Point Clouds. SENSORS (BASEL, SWITZERLAND) 2024; 24:3185. [PMID: 38794039 PMCID: PMC11125693 DOI: 10.3390/s24103185] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/25/2024] [Revised: 05/10/2024] [Accepted: 05/15/2024] [Indexed: 05/26/2024]
Abstract
In the evolving landscape of autonomous driving technology, Light Detection and Ranging (LiDAR) sensors have emerged as a pivotal instrument for enhancing environmental perception. They can offer precise, high-resolution, real-time 3D representations around a vehicle, and the ability for long-range measurements under low-light conditions. However, these advantages come at the cost of the large volume of data generated by the sensor, leading to several challenges in transmission, processing, and storage operations, which can be currently mitigated by employing data compression techniques to the point cloud. This article presents a survey of existing methods used to compress point cloud data for automotive LiDAR sensors. It presents a comprehensive taxonomy that categorizes these approaches into four main groups, comparing and discussing them across several important metrics.
Collapse
Affiliation(s)
| | | | | | - Tiago Gomes
- Centro ALGORITMI/LASI, Escola de Engenharia, Universidade do Minho, 4800-058 Guimarães, Portugal
| |
Collapse
|
3
|
Adnan M, Slavic G, Martin Gomez D, Marcenaro L, Regazzoni C. Systematic and Comprehensive Review of Clustering and Multi-Target Tracking Techniques for LiDAR Point Clouds in Autonomous Driving Applications. SENSORS (BASEL, SWITZERLAND) 2023; 23:6119. [PMID: 37447967 DOI: 10.3390/s23136119] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/29/2023] [Revised: 06/03/2023] [Accepted: 06/27/2023] [Indexed: 07/15/2023]
Abstract
Autonomous vehicles (AVs) rely on advanced sensory systems, such as Light Detection and Ranging (LiDAR), to function seamlessly in intricate and dynamic environments. LiDAR produces highly accurate 3D point clouds, which are vital for the detection, classification, and tracking of multiple targets. A systematic review and classification of various clustering and Multi-Target Tracking (MTT) techniques are necessary due to the inherent challenges posed by LiDAR data, such as density, noise, and varying sampling rates. As part of this study, the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) methodology was employed to examine the challenges and advancements in MTT techniques and clustering for LiDAR point clouds within the context of autonomous driving. Searches were conducted in major databases such as IEEE Xplore, ScienceDirect, SpringerLink, ACM Digital Library, and Google Scholar, utilizing customized search strategies. We identified and critically reviewed 76 relevant studies based on rigorous screening and evaluation processes, assessing their methodological quality, data handling adequacy, and reporting compliance. As a result of this comprehensive review and classification, we were able to provide a detailed overview of current challenges, research gaps, and advancements in clustering and MTT techniques for LiDAR point clouds, thus contributing to the field of autonomous driving. Researchers and practitioners working in the field of autonomous driving will benefit from this study, which was characterized by transparency and reproducibility on a systematic basis.
Collapse
Affiliation(s)
- Muhammad Adnan
- Department of Electrical, Electronic, Telecommunications Engineering and Naval Architecture (DITEN), University of Genova, Via Opera Pia 11a, I-16145 Genoa, Italy
- Departamento de Ingeniería de Sistemas y Automática, Universidad Carlos III de Madrid, Butarque 15, Leganés, 28911 Madrid, Spain
| | - Giulia Slavic
- Department of Electrical, Electronic, Telecommunications Engineering and Naval Architecture (DITEN), University of Genova, Via Opera Pia 11a, I-16145 Genoa, Italy
- Departamento de Ingeniería de Sistemas y Automática, Universidad Carlos III de Madrid, Butarque 15, Leganés, 28911 Madrid, Spain
| | - David Martin Gomez
- Departamento de Ingeniería de Sistemas y Automática, Universidad Carlos III de Madrid, Butarque 15, Leganés, 28911 Madrid, Spain
| | - Lucio Marcenaro
- Department of Electrical, Electronic, Telecommunications Engineering and Naval Architecture (DITEN), University of Genova, Via Opera Pia 11a, I-16145 Genoa, Italy
| | - Carlo Regazzoni
- Department of Electrical, Electronic, Telecommunications Engineering and Naval Architecture (DITEN), University of Genova, Via Opera Pia 11a, I-16145 Genoa, Italy
| |
Collapse
|
4
|
Lee DS, Kwon SK. Intra Prediction Method for Depth Video Coding by Block Clustering through Deep Learning. SENSORS (BASEL, SWITZERLAND) 2022; 22:9656. [PMID: 36560023 PMCID: PMC9787791 DOI: 10.3390/s22249656] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/02/2022] [Revised: 12/07/2022] [Accepted: 12/07/2022] [Indexed: 06/17/2023]
Abstract
In this paper, we propose an intra-picture prediction method for depth video by a block clustering through a neural network. The proposed method solves a problem that the block that has two or more clusters drops the prediction performance of the intra prediction for depth video. The proposed neural network consists of both a spatial feature prediction network and a clustering network. The spatial feature prediction network utilizes spatial features in vertical and horizontal directions. The network contains a 1D CNN layer and a fully connected layer. The 1D CNN layer extracts the spatial features for a vertical direction and a horizontal direction from a top block and a left block of the reference pixels, respectively. 1D CNN is designed to handle time-series data, but it can also be applied to find the spatial features by regarding a pixel order in a certain direction as a timestamp. The fully connected layer predicts the spatial features of the block to be coded through the extracted features. The clustering network finds clusters from the spatial features which are the outputs of the spatial feature prediction network. The network consists of 4 CNN layers. The first 3 CNN layers combine two spatial features in the vertical and horizontal directions. The last layer outputs the probabilities that pixels belong to the clusters. The pixels of the block are predicted by the representative values of the clusters that are the average of the reference pixels belonging to the clusters. For the intra prediction for various block sizes, the block is scaled to the size of the network input. The prediction result through the proposed network is scaled back to the original size. In network training, the mean square error is used as a loss function between the original block and the predicted block. A penalty for output values far from both ends is introduced to the loss function for clear network clustering. In the simulation results, the bit rate is saved by up to 12.45% under the same distortion condition compared with the latest video coding standard.
Collapse
Affiliation(s)
- Dong-seok Lee
- AI Grand ICT Research Center, Dong-eui University, Busan 47340, Republic of Korea
| | - Soon-kak Kwon
- Department of Computer Software Engineering, Dong-eui University, Busan 47340, Republic of Korea
| |
Collapse
|
5
|
Abstract
Building information modelling (BIM) is evolving significantly in the architecture, engineering and construction industries. BIM involves various remote-sensing tools, procedures and standards that are useful for collating the semantic information required to produce 3D models. This is thanks to LiDAR technology, which has become one of the key elements in BIM, useful to capture a semantically rich geometric representation of 3D models in terms of 3D point clouds. This review paper explains the ‘Scan to BIM’ methodology in detail. The paper starts by summarising the 3D point clouds of LiDAR and photogrammetry. LiDAR systems based on different platforms, such as mobile, terrestrial, spaceborne and airborne, are outlined and compared. In addition, the importance of integrating multisource data is briefly discussed. Various methodologies involved in point-cloud processing such as sampling, registration and semantic segmentation are explained in detail. Furthermore, different open BIM standards are summarised and compared. Finally, current limitations and future directions are highlighted to provide useful solutions for efficient BIM models.
Collapse
|
6
|
Deep Learning Geometry Compression Artifacts Removal for Video-Based Point Cloud Compression. Int J Comput Vis 2021. [DOI: 10.1007/s11263-021-01503-6] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
7
|
Wiesmann L, Milioto A, Chen X, Stachniss C, Behley J. Deep Compression for Dense Point Cloud Maps. IEEE Robot Autom Lett 2021. [DOI: 10.1109/lra.2021.3059633] [Citation(s) in RCA: 17] [Impact Index Per Article: 5.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
8
|
Sun X, Wang S, Wang M, Wang Z, Liu M. A Novel Coding Architecture for LiDAR Point Cloud Sequence. IEEE Robot Autom Lett 2020. [DOI: 10.1109/lra.2020.3010207] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|