Adaptive Importance Sampling with Automatic Model Selection in Value Function Approximation
摘要:
Off-policy reinforcement learning is aimed at efficiently reusing data samples gathered in the past. A common approach is to use importance sampling techniques for compensating for the bias caused by the difference between data-collecting policies and the target policy. However, existing off-policy methods do not often take the variance of value function estimators explicitly into account and therefore their performance tends to be unstable. To cope with this problem, we propose using an adaptive importance sampling technique which allows us to actively control the trade-off between bias and variance. We further provide a method for optimally determining the trade-off parameter based on a statistical machine learning theory.
展开
被引量:
年份:
2007
通过文献互助平台发起求助,成功后即可免费获取论文全文。
相似文献
参考文献
引证文献
引用走势
辅助模式
引用
文献可以批量引用啦~
欢迎点我试用!