I developed a machine-learning autopilot as part of my Master’s thesis. The autopilot can control an aircraft even when unexpected failures occur, such as a plane losing part of its wing. By adapting the control strategy, the autopilot ensures the aircraft can still fly safely. If you are interested in the technical details, check:
The Challenge
While air travel is among the safest modes of transportation, traditional autopilot systems can struggle with unexpected malfunctions or structural failures. These systems are designed for operation under specific ranges and conditions, making them less effective when facing unexpected scenarios.
The solution
My project addresses this issue by developing an autopilot that can adapt to unforeseen conditions. To achieve this, I created a novel algorithm that combines Distributional Soft Actor-Critic (DSAC)1 and Incremental Dual Heuristic Programming (IDHP)2. DSAC excels in offline learning by using precomputed knowledge to make quick decisions. In contrast, IDHP adapts in real-time, known as online learning. By bringing the two together, the resulting hybrid algorithm allows the autopilot system to adjust in real-time to adverse conditions, enhancing its immediate and future response capabilities.

Topology of hybrid actor network.
Experiments
In my experiments, I demonstrated how this autopilot can retain control under sensor noise, variations in flight conditions, and failure cases, such as partial loss of rudder and aileron.
Technology Stack
- PyTorch: Used to develop and train the autopilot model.
- Gymnasium Used to build the reinforcement learning environment.
- Weights & Biases (W&B) Used for project tracking and MLOps.