Annotation-free Learning for Sensor Fusion in ADAS
| dc.contributor.author | Björkman, Maria | |
| dc.contributor.author | Tvingby, Ludvig | |
| dc.contributor.department | Chalmers tekniska högskola / Institutionen för elektroteknik | sv |
| dc.contributor.examiner | Hammarstrand, Lars | |
| dc.contributor.supervisor | Wang, Tzu-Jui | |
| dc.contributor.supervisor | Priisalu, Maria | |
| dc.date.accessioned | 2025-12-01T13:59:47Z | |
| dc.date.issued | 2025 | |
| dc.date.submitted | ||
| dc.description.abstract | Vehicle automation has the potential to significantly improve road safety. Achieving comprehensive vehicle perception requires systems that optimally combine information from multiple sensor modalities. Such systems leverage the strengths of each modality while compensating for their weaknesses. By continuously encoding and fusing information from cameras, LiDARs, RADARs and the motion of the egovehicle, a dynamic representation of the surrounding environment can be created and maintained. A major challenge for these systems is the large amount of annotated data required for training, as manual labelling creates a significant bottleneck for scalability. In this study, a pre-training task for a multi-modal machine learning model was implemented and evaluated. To circumvent labour-intensive labelling, self-supervision was employed, with both the model input and the supervision signal involving annotation-free data. The pre-training aimed to learn general features related to sensor pose changes by predicting ego-vehicle pose changes using odometry data. To assess pre-training performance, the features were then used as initial weights for fine-tuning a perception model. The performance of the perception model using baseline weights trained on annotated data was similar to that using weights trained on annotation-free data, indicating that the proposed method is viable. However, further testing is required to establish statistical significance. Future work could explore implementing attention-based methods for feature matching between scene representations to improve model performance. | |
| dc.identifier.coursecode | EENX30 | |
| dc.identifier.uri | http://hdl.handle.net/20.500.12380/310781 | |
| dc.language.iso | eng | |
| dc.setspec.uppsok | Technology | |
| dc.subject | ADAS | |
| dc.subject | Annotation-free | |
| dc.subject | Ego-vehicle | |
| dc.subject | Multi-modal | |
| dc.subject | Perception | |
| dc.subject | Pretraining | |
| dc.subject | Sensor Fusion | |
| dc.subject | Transformer | |
| dc.title | Annotation-free Learning for Sensor Fusion in ADAS | |
| dc.type.degree | Examensarbete för masterexamen | sv |
| dc.type.degree | Master's Thesis | en |
| dc.type.uppsok | H | |
| local.programme | Information and communication technology (MPICT), MSc |
