@article{Schlosser2020, author = {Schlosser, Rainer}, title = {Risk-sensitive control of Markov decision processes}, series = {Computers \& operations research : and their applications to problems of world concern}, volume = {123}, journal = {Computers \& operations research : and their applications to problems of world concern}, publisher = {Elsevier}, address = {Oxford}, issn = {0305-0548}, doi = {10.1016/j.cor.2020.104997}, pages = {14}, year = {2020}, abstract = {In many revenue management applications risk-averse decision-making is crucial. In dynamic settings, however, it is challenging to find the right balance between maximizing expected rewards and minimizing various kinds of risk. In existing approaches utility functions, chance constraints, or (conditional) value at risk considerations are used to influence the distribution of rewards in a preferred way. Nevertheless, common techniques are not flexible enough and typically numerically complex. In our model, we exploit the fact that a distribution is characterized by its mean and higher moments. We present a multi-valued dynamic programming heuristic to compute risk-sensitive feedback policies that are able to directly control the moments of future rewards. Our approach is based on recursive formulations of higher moments and does not require an extension of the state space. Finally, we propose a self-tuning algorithm, which allows to identify feedback policies that approximate predetermined (risk-sensitive) target distributions. We illustrate the effectiveness and the flexibility of our approach for different dynamic pricing scenarios. (C) 2020 Elsevier Ltd. All rights reserved.}, language = {en} }