A deep learning model using Q-learning successfully learns control policies from raw pixels and outperforms previous methods and human experts in Atari 2600 games.
This paper introduces a deep learning model that leverages a convolutional neural network trained with a variant of Q-learning to learn control policies directly from high-dimensional sensory input, specifically raw pixels. This approach was tested on seven Atari 2600 games within the Arcade Learning Environment. The results demonstrate that the model outperforms all previous methods on six of the games and even surpasses human expert performance on three of the games, without requiring any adjustments to the architecture or learning algorithm for different games.
Potential limitations include the need for significant computational resources for training deep learning models and the challenge of generalizing this approach to more complex or varied environments beyond the specific set of Atari 2600 games. Future work may explore optimization of the architecture and learning algorithm for broader applications and more efficient training processes.
Human-level control through deep reinforcement learning
semantic_scholar
A general reinforcement learning algorithm that masters chess, shogi, and Go through self-play
semantic_scholar
PyTorch: An Imperative Style, High-Performance Deep Learning Library
semantic_scholar
RCSB Protein Data Bank (RCSB.org): delivery of experimentally-determined PDB structures alongside one million computed structure models of proteins from artificial intelligence/machine learning
semantic_scholar
Omnilingual MT: Machine Translation for 1,600 Languages
arxiv