Adaptive Importance Sampling with Automatic Model Selection in Value Function Approximation

阅读量:

37

作者:

H HachiyaT AkiyamaM SugiyamaJ Peters

展开

摘要:

Off-policy reinforcement learning is aimed at efficiently reusing data samples gathered in the past. A common approach is to use importance sampling techniques for compensating for the bias caused by the difference between data-collecting policies and the target policy. However, existing off-policy methods do not often take the variance of value function estimators explicitly into account and therefore their performance tends to be unstable. To cope with this problem, we propose using an adaptive importance sampling technique which allows us to actively control the trade-off between bias and variance. We further provide a method for optimally determining the trade-off parameter based on a statistical machine learning theory.

展开

被引量:

46

年份:

2007

通过文献互助平台发起求助,成功后即可免费获取论文全文。

相似文献

参考文献

引证文献

引用走势

2009
被引量:24

辅助模式

0

引用

文献可以批量引用啦~
欢迎点我试用!

引用