Periodic Reporting for period 1 - VeVuSafety (Artificial Intelligence for Traffic Safety between Vehicles and Vulnerable Road Users)
Berichtszeitraum: 2022-10-01 bis 2024-09-30
Furthermore, the project’s scientific contributions have strong potential for real-world applications. For example, we presented an uncertainty-based method for selecting collective perception messages by manipulating the generated evidential bird’s-eye-view maps (WP1). This method improves communication efficiency among connected autonomous vehicles (CAVs) by sharing only the information the ego vehicle is uncertain about, while other CAVs provide more certain data from their perspectives. This approach reduces communication overhead by 87%, with only a slight drop in performance for collective perception, which is critical for real-time communication in autonomous driving systems. In WP3, for trajectory prediction, we demonstrated that dynamic scene context from traversed trajectories can compensate for the lack of map data to achieve scene-compliant trajectory predictions. Moreover, to facilitate interactive autonomous driving (WP4), we explored various human-machine interfaces (HMI), including visual, textual, auditory, and multimodal interfaces, to ensure clear communication between autonomous vehicles and other road users in diverse driving situations, such as shared spaces or bottleneck roads.
Looking ahead, with the development of Vision Language Models (VLMs), we aim to explore their potential for generating both visual and textual cues to enhance the explainability of deep learning models' decision-making processes. We also plan to leverage Large Language Models (LLMs), trained on vast amounts of text data, to address long-tail scenarios for behavior modeling, motion prediction, and situation-aware detection, further advancing end-to-end autonomous driving.