Researchers from the US Naval Research Laboratory (NRL) have achieved a groundbreaking milestone in space robotics. Samantha Chapin, Kenneth Stewart, Roxana Leontie, and Carl Glen Henshaw have successfully demonstrated the first-ever reinforcement learning (RL) control of a free-flying robot in space. Their experiment, known as the Autonomous Planning In-space Assembly Reinforcement-learning free-flYer (APIARY), utilized the NASA Astrobee robot aboard the International Space Station (ISS) to showcase the potential of RL in enhancing robotic autonomy in zero-gravity environments.
The APIARY team developed a robust 6-degrees of freedom (DOF) control policy using an actor-critic Proximal Policy Optimization (PPO) network within the NVIDIA Isaac Lab simulation environment. This policy was trained by randomizing over goal poses and mass distributions to ensure robustness. The simulation testing, ground testing, and subsequent flight validation of this experiment have collectively demonstrated the transformative potential of RL for space exploration, logistics, and real-time mission needs.
The successful on-orbit demonstration of RL control marks a significant advancement in robotic autonomy. The APIARY experiment highlights the capability of RL to enable rapid development and deployment of tailored behaviors for space robots, a process that can now be accomplished in minutes to hours. This breakthrough could revolutionize how robotic systems are utilized in space, making them more adaptable and efficient in responding to mission requirements.
The implications of this research extend beyond the confines of the ISS. The ability to deploy RL-controlled robots in space opens up new possibilities for in-space assembly, maintenance, and exploration. These advancements could lead to more sophisticated and autonomous robotic systems that can operate independently in the challenging environment of space, reducing the need for human intervention and increasing the safety and efficiency of space missions.
As the maritime industry increasingly looks to space for navigation, communication, and environmental monitoring, the lessons learned from the APIARY experiment could also have practical applications on Earth. The principles of RL and autonomous control can be adapted to improve the performance of maritime robots and autonomous vessels, enhancing their ability to operate in complex and dynamic environments. This could lead to more efficient and sustainable maritime operations, contributing to the broader goals of decarbonization and digital transformation in the sector.
The APIARY experiment represents a significant step forward in the field of space robotics and highlights the potential of reinforcement learning to revolutionize robotic autonomy. As the technology continues to evolve, it is likely to have far-reaching implications for both space exploration and maritime operations, driving innovation and efficiency in these critical sectors. Read the original research paper here.

