学位论文详细信息
State Clustering in Markov Decisions Processes with an Application in Information Sharing
Markov Decision Process with Restricted Observati;Information Sharing;Markov Decision Process
Berrings Davis, Lauren Marie ; Dr. Donald Bitzer, Committee Member,Dr. Henry Nuttle, Committee Member,Dr. Russell King, Committee Co-Chair,Dr. Thom Hodgson, Committee Co-Chair,Berrings Davis, Lauren Marie ; Dr. Donald Bitzer ; Committee Member ; Dr. Henry Nuttle ; Committee Member ; Dr. Russell King ; Committee Co-Chair ; Dr. Thom Hodgson ; Committee Co-Chair
University:North Carolina State University
关键词: Markov Decision Process with Restricted Observati;    Information Sharing;    Markov Decision Process;   
Others  :  https://repository.lib.ncsu.edu/bitstream/handle/1840.16/5880/etd.pdf?sequence=1&isAllowed=y
美国|英语
来源: null
PDF
【 摘 要 】

This research examines state clustering in Markov Decision processes, specifically addressing the problem referred to as Markov Decision process with restricted observations. The general problem is a special case of a Partially Observable Markov Decision process where the state space is partitioned into mutually exclusive sets representing the observable portion of the process.The goal is to find an optimal policy defined over the partition of the state space that minimizes (maximizes) some performance objective.Algorithms presented to solve this problem for the infinite horizon undiscounted average cost case have largely been based on enumerative procedures.A heuristic solution procedure based on Howard's (1960) policy iteration method is presented.Applications of Markov decision processes with restricted observations exist in networks of queues, retrial queues, maintenance problems and queuing networks with server control.A new application area is proposed in the field of information sharing to measure the value of information sharing in a supply chain under optimal control.This is achieved by representing a model of full information sharing as a completely observable Markov Decision process (MDP), while no information sharing is represented as an MDP with restricted observations. Solution procedures are presented for the general Markov Decision process with restricted observations.Heuristic solutions are evaluated against the optimal solution obtained via total enumeration.Both random Markov Decision processes and information sharing problems are studied.The value of sharing information in a two-stage supply chain system is studied.The influence of capacity, demand, cost and retailer policy on the value of information sharing is considered. Insight on the structure of the optimal policy with and without information sharing is provided.

【 预 览 】
附件列表
Files Size Format View
State Clustering in Markov Decisions Processes with an Application in Information Sharing 739KB PDF download
  文献评价指标  
  下载次数:10次 浏览次数:37次