1. 程式人生 > >Continuous control with deep reinforcement learning

Continuous control with deep reinforcement learning

(Submitted on 9 Sep 2015 (v1), last revised 29 Feb 2016 (this version, v5))
We adapt the ideas underlying the success of Deep Q-Learning to the continuous action domain. We present an actor-critic, model-free algorithm based on the deterministic policy gradient that can operate over continuous action spaces. Using the same learning algorithm, network architecture and hyper-parameters, our algorithm robustly solves more than 20 simulated physics tasks, including classic problems such as cartpole swing-up, dexterous manipulation, legged locomotion and car driving. Our algorithm is able to find policies whose performance is competitive with those found by a planning algorithm with full access to the dynamics of the domain and its derivatives. We further demonstrate that for many of the tasks the algorithm can learn policies end-to-end: directly from raw pixel inputs.
Comments: 10 pages + supplementary
Subjects: Learning (cs.LG); Machine Learning (stat.ML)

Submission history

From: Jonathan Hunt [view email
[v1] Wed, 9 Sep 2015 23:01:36 GMT (344kb,D)
[v2] Wed, 18 Nov 2015 17:34:41 GMT (338kb,D)
[v3] Thu, 7 Jan 2016 19:09:07 GMT (338kb,D)
[v4] Tue, 19 Jan 2016 20:30:47 GMT (339kb,D)
[v5]
 Mon, 29 Feb 2016 18:45:53 GMT (339kb,D)