Avatar

Antonin Raffin

Research Engineer in Robotics and Machine Learning

German Aerospace Center (DLR)

Bio

Antonin Raffin is a research engineer at the German Aerospace Center (DLR) who specializes in reinforcement learning (RL). He is the lead developer of Stable-Baselines3 (SB3), an open-source library that implements Deep RL algorithms. His main focus is on learning controllers directly on real robots and improving the reproducibility of RL.

Interests

  • Robotics
  • Reinforcement Learning
  • State Representation Learning
  • Machine Learning

Projects

*

SBX: Stable Baselines Jax

Proof of concept version of Stable-Baselines3 in Jax.

Datasaurust

Blazingly fast implementation of the Datasaurus paper in Rust. Same Stats, Different Graphs.

Stable Baselines3

A set of improved implementations of reinforcement learning algorithms in PyTorch.

Learning to Drive Smoothly in Minutes

Learning to drive smoothly in minutes using reinforcement learning on a Donkey Car.

RL Baselines Zoo

A collection of 70+ pre-trained RL agents using Stable Baselines

S-RL Toolbox

S-RL Toolbox: Reinforcement Learning (RL) and State Representation Learning (SRL) for Robotics

Stable Baselines

A fork of OpenAI Baselines, implementations of reinforcement learning algorithms

Racing Robot

Autonomous Racing Robot With an Arduino, a Raspberry Pi and a Pi Camera

Arduino Robust Serial

A simple and robust serial communication protocol. Implementation in C Arduino, C++, Python and Rust.

Recent & Upcoming Talks

Stable-Baselines3 (SB3) Tutorial: Getting Started With Reinforcement Learning

This tutorial will present the basics of the Gymnasium and Stable-Baselines3 (SB3) libraries in order to apply reinforcement learning …

PhD Defense: Enabling Reinforcement Learning on Real Robots

This dissertation makes several contributions to the training of reinforcement learning agents directly on real robots. It introduces a …

Designing (Robot) Software That Is Easy to Use

Hardware without software is like an instrument without a musician. Based on my experience maintaining the Stable-Baselines3 library and working with real robots, I will present key principles for creating easy-to-use interfaces.

Recent Advances in RL for Continuous Control

A presentation on recent advances in RL, in terms of algorithms, software, and simulators.

Recent Posts

RL103: From Deep Q-Learning (DQN) to Soft Actor-Critic (SAC) and Beyond

This second blog post continues my practical introduction to (deep) reinforcement learning, presenting the main concepts and providing intuitions to understand the more recent Deep RL algorithms. In a first post (RL102), I started from tabular Q-learning and worked my way up to Deep Q-learning (DQN).

RL102: From Tabular Q-Learning to Deep Q-Learning (DQN)

This blog post is meant to be a practical introduction to (deep) reinforcement learning1, presenting the main concepts and providing intuitions to understand the more recent Deep RL algorithms. For a more in-depth and theoretical introduction, I recommend reading the RL Bible by Sutton and Barto.

Getting SAC to Work on a Massive Parallel Simulator: Tuning for Speed (Part II)

This second post details how I tuned the Soft-Actor Critic (SAC) algorithm to learn as fast as PPO in the context of a massively parallel simulator (thousands of robots simulated in parallel).

Automatic Hyperparameter Tuning - In Practice (Part 2)

This is the second (and last) post on automatic hyperparameter optimization. In the first part, I introduced the challenges and main components of hyperparameter tuning (samplers, pruners, objective function, …). This second part is about the practical application of this technique with the Optuna library, in a reinforcement learning setting (using the Stable-Baselines3 (SB3) library).

Experience

 
 
 
 
 

Researcher

German Aerospace Center (DLR)

October 2018 – Present Munich
Machine Learning for Robots.
 
 
 
 
 

PhD in Robotics

Technical University of Munich (TUM)

October 2018 – October 2025 Munich
PhD Thesis: Enabling Reinforcement Learning on Real Robots
 
 
 
 
 

Research Engineer

ENSTA ParisTech - U2IS robotics lab

October 2017 – October 2018 Palaiseau
Working on Reinforcement Learning and State Representation Learning for the DREAM project.
 
 
 
 
 

Research Intern

Riminder

April 2017 – September 2017 Paris
Deep Learning for Human Resources.
 
 
 
 
 

Research Intern

TU Berlin - RBO lab

May 2016 – August 2016 Berlin
Research internship in representation and reinforcement learning.