http://arxiv.org/abs/1507.04888
This paper presents a general framework for exploiting
the representational capacity of neural
networks to approximate complex, nonlinear reward
functions in the context of solving the inverse
reinforcement learning (IRL) problem. We
show in this context that the Maximum Entropy
paradigm for IRL lends itself naturally to the effi-
cient training of deep architectures. At test time,
the approach leads to a computational complexity
independent of the number of demonstrations,
which makes it especially well-suited for applications
in life-long learning scenarios. Our approach
achieves performance commensurate to
the state-of-the-art on existing benchmarks while
exceeding on an alternative benchmark based on
highly varying reward structures.Finally, we extend
the basic architecture - which is equivalent
to a simplified subclass of Fully Convolutional
Neural Networks (FCNNs) with width one - to
include larger convolutions in order to eliminate
dependency on precomputed spatial features and
work on raw input representations.
Maximum Entropy Deep Inverse Reinforcement Learning
Paste_Image.png
Paste_Image.png
Paste_Image.png
Paste_Image.png
Paste_Image.png
最后编辑于 :
©著作权归作者所有,转载或内容合作请联系作者
- 文/潘晓璐 我一进店门,熙熙楼的掌柜王于贵愁眉苦脸地迎上来,“玉大人,你说我怎么就摊上这事。” “怎么了?”我有些...
- 文/花漫 我一把揭开白布。 她就那样静静地躺着,像睡着了一般。 火红的嫁衣衬着肌肤如雪。 梳的纹丝不乱的头发上,一...
- 文/苍兰香墨 我猛地睁开眼,长吁一口气:“原来是场噩梦啊……” “哼!你这毒妇竟也来了?” 一声冷哼从身侧响起,我...