Preprint Article Version 1 Preserved in Portico This version is not peer-reviewed

3D Point Cloud Enriched Semantic Segmentation through Knowledge-Enhanced Deep Learning Technique

Version 1 : Received: 27 June 2023 / Approved: 28 June 2023 / Online: 29 June 2023 (05:41:26 CEST)

How to cite: Ballouch, Z.; Hajji, R.; Kharroubi, A.; Poux, F.; Billen, R. 3D Point Cloud Enriched Semantic Segmentation through Knowledge-Enhanced Deep Learning Technique. Preprints 2023, 2023062022. https://doi.org/10.20944/preprints202306.2022.v1 Ballouch, Z.; Hajji, R.; Kharroubi, A.; Poux, F.; Billen, R. 3D Point Cloud Enriched Semantic Segmentation through Knowledge-Enhanced Deep Learning Technique. Preprints 2023, 2023062022. https://doi.org/10.20944/preprints202306.2022.v1

Abstract

Digital Twin Cities (DTCs) play a fundamental role in city planning and management. They allow three-dimensional modeling and simulation of cities. 3D semantic segmentation is the foundation for automatically creating enriched DTCs, as well as their updates. Past studies indicate that prior level fusion approaches demonstrate more promising precisions in 3D semantic segmentation compared to point level fusion, features level fusion, and decision level fusion families. In order to improve point cloud enriched semantic segmentation outcomes, this article proposes a new approach for 3D point cloud semantic segmentation through developing and benchmarking three prior level fusion scenarios. A reference approach based on point clouds and aerial images was proposed to compare it with the different developed scenarios. In each scenario, we inject a specific prior knowledge (geometric features,classified images ,etc) and aerial images as attributes of point clouds into the neural network’s learning pipeline. The objective is to find the one that integrates the most significant prior knowledge and enhances neural network knowledge more profoundly, which we have named the "smart fusion approach". The advanced Deep Learning algorithm "RandLaNet" was adopted to implement the different proposed scenarios and the reference approach, due to its excellent performance demonstrated in the literature. The introduction of some significant features associated with the label classes facilitated the learning process and improved the semantic segmentation results that can be achievable with the same neural network alone. Overall, our contribution provides a promising solution for addressing some challenges, in particular more accurate extraction of semantically rich objects from the urban fabric. An assessment of the semantic segmentation results obtained by the different scenarios is performed based on metrics computation and visual investigations. Finally,the smart fusion approach was derived based on the obtained qualitative and quantitative results.

Keywords

smart fusion approach; enriched semantic segmentation; LiDAR point clouds; images data; data fusion; prior knowledge; deep learning; urban environment

Subject

Environmental and Earth Sciences, Remote Sensing

Comments (0)

We encourage comments and feedback from a broad range of readers. See criteria for comments and our Diversity statement.

Leave a public comment
Send a private comment to the author(s)
* All users must log in before leaving a comment
Views 0
Downloads 0
Comments 0


×
Alerts
Notify me about updates to this article or when a peer-reviewed version is published.
We use cookies on our website to ensure you get the best experience.
Read more about our cookies here.