Urban object detection using sensor fusion for autonomous navigation

dc.contributor
Horváth, Dániel
dc.contributor
García, Fernando
dc.contributor.author
Yavuz, Selin
dc.date.accessioned
2026-03-07T19:50:53Z
dc.date.available
2026-03-07T19:50:53Z
dc.date.issued
2025-05
dc.identifier
http://hdl.handle.net/10256/28372
dc.identifier.uri
https://hdl.handle.net/10256/28372
dc.description.abstract
Urban object detection remains an important challenge for autonomous naviga tion, particularly in accurately identifying and locating thin and small structures such as poles. These structures are often difficult to detect due to their size, shape, and integration into complex urban scenes. In response to these challenges, this thesis proposes a dual-stage approach to urban object detection by integrating 2D deep learning-based object detection into 3D spatial representation. The methodology consists of two main stages: (1) 2D object detection, which involves fine-tuning a pretrained model to identify pole-like objects in urban environments, and (2) 3D object detection, where the 2D detections are projected into 3D space and further processed through point cloud clustering and outlier rejection to generate accurate 3D bounding boxes. The approach shows how combining segmentation-derived labels with geometric modeling can serve as an alternative to fully supervised 3D object inference. The model demonstrates promising performance on a custom dataset, benefiting from the SAHI framework, which enhances the detection of smaller or more challenging to-see objects. These results indicate that integrating 2D and 3D detection methods can provide valuable spatial information, even when detailed 3D annotations are limited or unavailable, supporting practical urban navigation applications. This thesis presents a foundational pipeline for pole detection in urban naviga tion, laying the groundwork for more adaptable and robust autonomous systems in complex real-world environments. Further development, including dataset growth and model refinement, can enhance detection accuracy and robustness, thereby ad vancing the effectiveness of autonomous systems in urban environments.
dc.description.abstract
9
dc.format
application/pdf
dc.language
eng
dc.publisher
Universitat de Girona. Institut de Recerca en Visió per Computador i Robòtica
dc.rights
Attribution-NonCommercial-NoDerivatives 4.0 International
dc.rights
http://creativecommons.org/licenses/by-nc-nd/4.0/
dc.rights
info:eu-repo/semantics/openAccess
dc.source
Erasmus Mundus Joint Master in Intelligent Field Robotic Systems (IFROS)
dc.subject
Aprenentatge profund (Aprenentatge automàtic)
dc.subject
Deep learning (Machine learning)
dc.subject
Sensors òptics tridimensionals
dc.subject
Sensors
dc.subject
Urban object detection
dc.subject
Vehicles autònoms -- Sistemes de navegació
dc.subject
Autonomous Vehicles -- Navigation systems
dc.subject
SAHI
dc.subject
Computer vision
dc.subject
Visió per ordinador
dc.title
Urban object detection using sensor fusion for autonomous navigation
dc.type
info:eu-repo/semantics/masterThesis


Ficheros en el ítem

FicherosTamañoFormatoVer

No hay ficheros asociados a este ítem.

Este ítem aparece en la(s) siguiente(s) colección(ones)