News Release

Drone racing prepares neural-network AI for space

In collaboration with the European Space Agency, drones are being raced against the clock at Delft University of Technology’s ‘Cyber Zoo’ to test the performance of neural-network-based AI control systems planned for next-generation space missions.

Peer-Reviewed Publication

Delft University of Technology

Drone takeoff controlled by AI

image: 

Drone takeoff controlled by AI

view more 

Credit: Guido de Croon / TU Delft|MAV Lab

Drones are being raced against the clock at Delft University of Technology’s ‘Cyber Zoo’ to test the performance of neural-network-based AI control systems planned for next-generation space missions. The research – undertaken by ESA’s Advanced Concepts Team together with the Micro Air Vehicle Laboratory, MAVLab, of TU Delft –  was published in Science Robotics. 

“Through a long-term collaboration, we’ve been looking into the use of trainable neural networks for the autonomous oversight of all kinds of demanding spacecraft manoeuvres, such as interplanetary transfers, surface landings and dockings,” notes Dario Izzo, scientific coordinator of ESA’s ACT.

“In space every onboard resource must be utilised as efficiently as possible – including propellant, available energy, computing resources, and often time. Such a neural network approach could enable optimal onboard operations, boosting mission autonomy and robustness. But we needed a way to test it in the real world, ahead of planning actual space missions.

“That’s when we settled on drone racing as the ideal gym environment to test end-to-end neural architectures on real robotic platforms, to increase confidence in their future use in space.”

Drones have been competing to achieve the best time through a set course within the Cyber Zoo at TU Delft, a 10x10 m test area maintained by the University’s Faculty of Aerospace Engineering, ESA’s partner in this research. Human-steered ‘Micro Air Vehicle’ quadcopters were alternated with autonomous counterparts with neural networks trained in various ways.

“The traditional way that spacecraft manoeuvres work is that they are planned in detail on the ground then uploaded to the spacecraft to be carried out,” explains ACT Young Graduate Trainee Sebastien Origer. “Essentially, when it comes to mission Guidance and Control, the Guidance part occurs on the ground, while the Control part is undertaken by the spacecraft.”

The space environment is inherently unpredictable however, with the potential for all kinds of unforeseen factors and noise, such as gravitational variations, atmospheric turbulence or planetary bodies that turn out to be shaped differently from on-ground modelling.

Whenever the spacecraft deviates from its planned path for whatever reason, its control system works to return it to the set profile. The problem is that such an approach can be quite costly in resource terms, requiring a whole set of brute force corrections.

Sebastien adds: “Our alternative end-to-end Guidance & Control Networks, G&C Nets, approach involves all the work taking place on the spacecraft. Instead of sticking a single set course, the spacecraft continuously replans its optimal trajectory, starting from the current position it finds itself at, which proves to be much more efficient.”

In computer simulations neural nets composed of interlinked neurons – mimicking the setup of animal brains – performed well when trained using  ‘behavioural cloning’, based on prolonged exposure to expert examples. But then came the question of how to build trust in this approach in the real world. At this point the researchers turned to drones.

“There’s quite a lot of synergies between drones and spacecraft, although the dynamics involved in flying drones are much faster and noisier,” comments Dario.

“When it comes to racing obviously the main scarce resource is time, but we can use that as a substitute for other variables that a space mission might have to prioritise, such as propellant mass. Satellite CPUs are quite constrained, but our G&CNETs are surprisingly modest, perhaps storing up to 30 000 parameters in memory, which can be done using only a few hundred kilobytes, involving less than 360 neurons in all.”

In order to be optimal, the G&CNet should be able to send commands directly to the actuators. For a spacecraft these are the thrusters and in the case of drones their propellers.

“The main challenge that we tackled for bringing G&CNets to drones is the reality gap between the actuators in simulation and in reality”, says Christophe De Wagter, principal investigator at TU Delft. “We deal with this by identifying the reality gap while flying and teaching the neural network to deal with it. For example, if the propellers give less thrust than expected, the drone can notice this via its accelerometers. The neural network will then regenerate the commands to follow the new optimal path.”

“There’s a whole academic community of drone racing, and it all comes down to winning races,” says Sebastien. “For our G&CNets approach, the use of drones represents a way to build trust, develop a solid theoretical framework and establish safety bounds, ahead of planning an actual space mission demonstrator.”


Disclaimer: AAAS and EurekAlert! are not responsible for the accuracy of news releases posted to EurekAlert! by contributing institutions or for the use of any information through the EurekAlert system.