期刊论文详细信息
JOURNAL OF MATHEMATICAL ANALYSIS AND APPLICATIONS 卷:435
Near optimality of quantized policies in stochastic control under weak continuity conditions
Article
Saldi, Naci1  Yueksel, Serdar1  Linder, Tamas1 
[1] Queens Univ, Dept Math & Stat, Kingston, ON K7L 3N6, Canada
关键词: Stochastic control;    Quantization;    Approximation;    Partially observed Markov decision processes;   
DOI  :  10.1016/j.jmaa.2015.10.008
来源: Elsevier
PDF
【 摘 要 】

This paper studies the approximation of optimal control policies by quantized (discretized) policies for a very general class of Markov decision processes (MDPs). The problem is motivated by applications in networked control systems, computational methods for MDPs, and learning algorithms for MDPs. We consider the finite-action approximation of stationary policies for a discrete-time Markov decision process with discounted and average costs under a weak continuity assumption on the transition probability, which is a significant relaxation of conditions required in earlier literature. The discretization is constructive, and quantized policies are shown to approximate optimal deterministic stationary policies with arbitrary precision. The results are applied to the fully observed reduction of a partially observed Markov decision process, where weak continuity is a much more reasonable assumption than more stringent conditions such as strong continuity or continuity in total variation. (C) 2015 Elsevier Inc. All rights reserved.

【 授权许可】

Free   

【 预 览 】
附件列表
Files Size Format View
10_1016_j_jmaa_2015_10_008.pdf 422KB PDF download
  文献评价指标  
  下载次数:11次 浏览次数:1次