dc.contributor
Horváth, Dániel
dc.contributor
García, Fernando
dc.contributor.author
Yavuz, Selin
dc.date.accessioned
2026-03-07T19:50:53Z
dc.date.available
2026-03-07T19:50:53Z
dc.identifier
http://hdl.handle.net/10256/28372
dc.identifier.uri
https://hdl.handle.net/10256/28372
dc.description.abstract
Urban object detection remains an important challenge for autonomous naviga tion, particularly in accurately identifying and locating thin and small structures
such as poles. These structures are often difficult to detect due to their size, shape,
and integration into complex urban scenes.
In response to these challenges, this thesis proposes a dual-stage approach to
urban object detection by integrating 2D deep learning-based object detection into
3D spatial representation. The methodology consists of two main stages: (1) 2D
object detection, which involves fine-tuning a pretrained model to identify pole-like
objects in urban environments, and (2) 3D object detection, where the 2D detections
are projected into 3D space and further processed through point cloud clustering
and outlier rejection to generate accurate 3D bounding boxes.
The approach shows how combining segmentation-derived labels with geometric
modeling can serve as an alternative to fully supervised 3D object inference. The
model demonstrates promising performance on a custom dataset, benefiting from
the SAHI framework, which enhances the detection of smaller or more challenging to-see objects. These results indicate that integrating 2D and 3D detection methods
can provide valuable spatial information, even when detailed 3D annotations are
limited or unavailable, supporting practical urban navigation applications.
This thesis presents a foundational pipeline for pole detection in urban naviga tion, laying the groundwork for more adaptable and robust autonomous systems in
complex real-world environments. Further development, including dataset growth
and model refinement, can enhance detection accuracy and robustness, thereby ad vancing the effectiveness of autonomous systems in urban environments.
dc.description.abstract
9
dc.format
application/pdf
dc.publisher
Universitat de Girona. Institut de Recerca en Visió per Computador i Robòtica
dc.rights
Attribution-NonCommercial-NoDerivatives 4.0 International
dc.rights
http://creativecommons.org/licenses/by-nc-nd/4.0/
dc.rights
info:eu-repo/semantics/openAccess
dc.source
Erasmus Mundus Joint Master in Intelligent Field Robotic Systems (IFROS)
dc.subject
Aprenentatge profund (Aprenentatge automàtic)
dc.subject
Deep learning (Machine learning)
dc.subject
Sensors òptics tridimensionals
dc.subject
Urban object detection
dc.subject
Vehicles autònoms -- Sistemes de navegació
dc.subject
Autonomous Vehicles -- Navigation systems
dc.subject
Computer vision
dc.subject
Visió per ordinador
dc.title
Urban object detection using sensor fusion for autonomous navigation
dc.type
info:eu-repo/semantics/masterThesis