Learning to Fly with a Video Generator
Chia-Chun Chung, Wen-Hsiao Peng, Teng-Hu Cheng and Chia-Hau Yu
Abstract
This paper demonstrates a model-based reinforcement learning framework for training a self-flying drone. We implement the Dreamer proposed in a prior work as an environment model that responds to the action taken by the drone by predicting the next video frame as a new state signal. The Dreamer is a conditional video sequence generator. This model-based environment avoids the time-consuming interactions between the agent and the environment, speeding up largely the training process. This demonstration showcases for the first time the application of the Dreamer to train an agent that can finish the racing task in the Airsim simulator.
Experimental Results
Dreamer: conditional video generator
Demo video