← Back to Paper List

Champion-level drone racing using deep reinforcement learning

Elia Kaufmann, L. Bauersfeld, Antonio Loquercio, Matthias Müller, V. Koltun, Davide Scaramuzza
Robotics and Perception Group, University of Zurich, Intel Labs
Nature (2023)
RL Agent

📝 Paper Summary

Autonomous Drone Racing Sim-to-Real Transfer Robotic Control
Swift achieves world-champion drone racing performance by training a reinforcement learning policy in a simulator that is augmented with data-driven residual models to correct for real-world perception and dynamics discrepancies.
Core Problem
Autonomous drones struggle to match human champions because policies trained in simulators fail when transferring to the real world due to unmodeled aerodynamic effects and noisy sensory perception (the 'sim-to-real' gap).
Why it matters:
  • Demonstrates that autonomous mobile robots can reach physical performance limits previously exclusive to expert humans
  • Solves the challenge of controlling high-speed systems where accurate state estimation is difficult due to motion blur and delays
  • Traditional methods (like optimal control) fail when their rigid physical models do not perfectly match reality or when state estimation is noisy
Concrete Example: A standard simulation assumes a drone executes a sharp turn instantly. In reality, at 100 km/h, aerodynamic drag and motor lag cause the drone to slide (drift). A policy trained only in standard sim will command a turn, expect to be on track, but actually crash into a wall. Swift learns a 'residual' model from real data to predict this specific drift and adjusts its training to anticipate it.
Key Novelty
Hybrid Sim-to-Real with Residual Modeling
  • Combines deep reinforcement learning (RL) with physics-based simulation, but augments the simulation using real-world data
  • Uses 'residual models' (corrections) learned from sparse real-world data: Gaussian Processes for perception noise and k-Nearest Neighbors for aerodynamic discrepancies
  • Trains the flight policy inside this augmented simulation, allowing the drone to adapt to real-world imperfections without requiring massive amounts of dangerous real-world training
Evaluation Highlights
  • Won 15 out of 25 head-to-head races (60% win rate) against three human champions, including the Drone Racing League world champion
  • Achieved the fastest race time of 17.47s, beating the best human time (Alex Vanover) by ~0.5s
  • Consistently faster start reaction times (120ms average advantage) and tighter turning radiuses in complex maneuvers like the 'Split-S'
Breakthrough Assessment
10/10
A landmark achievement in mobile robotics. The first time an autonomous system has beaten human world champions in a real-world physical sport, solving extreme sim-to-real challenges.
×