A gymnasium style library for standardized Reinforcement Learning research in Air Traffic Management developed in Python. Build on BlueSky and The Farama Foundation's Gymnasium
An example trained agent attempting the merge environment available in BlueSky-Gym.
For a complete list of the currently available environments click here
pip install bluesky-gym
Note that the pip package is bluesky-gym
, for usage however, import as bluesky_gym
.
Using the environments follows the standard API from Gymnasium, an example of which is given below:
import gymnasium as gym
import bluesky_gym
bluesky_gym.register_envs()
env = gym.make('MergeEnv-v0', render_mode='human')
obs, info = env.reset()
done = truncated = False
while not (done or truncated):
action = ... # Your agent code here
obs, reward, done, truncated, info = env.step(action)
Additionally you can directly use algorithms from standardized libraries such as Stable-Baselines3 or RLlib to train a model:
import gymnasium as gym
import bluesky_gym
from stable_baselines3 import DDPG
bluesky_gym.register_envs()
env = gym.make('MergeEnv-v0', render_mode=None)
model = DDPG("MultiInputPolicy",env)
model.learn(total_timesteps=2e6)
model.save()
If you use BlueSky-Gym in your work, please cite it using:
@misc{bluesky-gym,
author = {Groot, DJ and Leto, G and Vlaskin, A and Moec, A and Ellerbroek, J},
title = {BlueSky-Gym: Reinforcement Learning Environments for Air Traffic Applications},
year = {2024},
journal = {SESAR Innovation Days 2024},
}
List of publications & preprints using BlueSky-Gym
(please open a pull request to add missing entries):
- missing entry