Abstract

The variance-penalized metric in Markov decision processes (MDPs) seeks to maximize the average reward minus a scalar time the variance of rewards. in this paper, our goal is to study the same metric in the context of the semi-Markov decision process (SMDP). in the SMDP, unlike the MDP, the time spent in each transition is not identical and may in fact be a random variable. We first develop an expression for the variance of rewards in the SMDPs, and then formulate the VP-SMDP. Our interest here is in solving the problem without generating the underlying transition probabilities of the Markov chains. We propose the use of two stochastic search techniques, namely simultaneous perturbation and learning automata, to solve the problem; these techniques use stochastic policies and can be used within simulators, thereby avoiding the generation of the transition probabilities. © 2011 IEEE.

Department(s)

Engineering Management and Systems Engineering

International Standard Book Number (ISBN)

978-145772108-3

International Standard Serial Number (ISSN)

0891-7736

Document Type

Article - Conference proceedings

Document Version

Citation

File Type

text

Language(s)

English

Rights

© 2024 Institute of Electrical and Electronics Engineers, All rights reserved.

Publication Date

01 Dec 2011

Share

 
COinS