Development of a Unified Evaluation Framework for Autonomous Driving in CARLA
- Institute
- Professur für autonome Fahrzeugsysteme (TUM-ED)
- Type
- Semester Thesis Master's Thesis
- Content
- Description
Motivation
The validation and safety assurance of modern autonomous vehicle systems is a cornerstone in the development of highly automated driving functions (Level 4+). In this context, there is a fundamental trade-off between simulation and real-world testing: while physical tests offer high physical validity, they are capital-intensive, time-consuming, and limited in scalability. In contrast, simulation platforms such as CARLA enable reproducible testing of complex traffic scenarios at a comparatively low cost and with high scalability. Across both real-world environments and simulations, a critical question persists: what defines 'good' autonomous driving?
Simply "completing a route" is not enough to qualify a self-driving agent as capable. To truly assess an agent's performance, we need a standardized "Digital Referee" that evaluates driving behavior across multiple dimensions.
While many existing benchmarks focus solely on goal reaching or collision avoidance (Safety), a sophisticated evaluator must consider the human element (Comfort) and operational costs (Efficiency). This thesis aims to build a robust, metric-based evaluator that processes CARLA telemetry data to provide a holistic driving assessment.
Objectives
The goal of this thesis is to design and implement a standardized evaluation tool that interfaces with CARLA and calculates key performance indicators (KPIs) in three core areas: Safety, Comfort, and Efficiency.
Planned work packages include:
- Metric Definition: Researching and selecting state-of-the-art mathematical formulations for driving quality.
- System Integration: Developing a Python-based toolchain that extracts real-time telemetry from the CARLA Python API.
- Evaluator Implementation: Building the logic to transform raw data (coordinates, velocities, accelerations) into normalized scores.
- Scenario Benchmarking: Testing the evaluator against different AD agents across various weather and traffic conditions.
- Visualization: Creating a "Reporting Dashboard" that summarizes the agent’s performance at the end of a run.
Requirements
- Interest in the simulation or evaluation of autonomous systems.
- Strong Python programming skills.
- A structured and independent working style.
- Proficiency in German or English.
- Prior experience with CARLA, ROS, or autonomous software stacks is helpful but not mandatory.
Start Date: Immediate / As soon as possible.
How to Apply: If interested, please submit your application via email, including your CV, current transcript of records, and a brief cover letter explaining why you would like to work on this subject.
- Tags
- AVS Bank
- Possible start
- sofort
- Contact
-
Christoph Bank
christoph.banktum.de