Abstract
In the literature, the problem of maximizing the expected discounted reward over all stopping rules has been explicitly solved for a number of reward functions (including $(\max\{x,0\})^{\nu}$, $\nu>0$, in particular) when the underlying process is either a random walk in discrete time or a L\'evy process in continuous time. All of such reward functions are increasing and logconcave while the corresponding optimal stopping rules have the threshold form. In this paper, we explore the close connection between increasing and logconcave reward functions and optimal stopping rules of threshold form. In the discrete case, we show that if a reward function defined on $\mathbb{Z}$ is nonnegative, increasing and logconcave, then the optimal stopping rule is of threshold form provided the the underlying random walk is skip-free to the right. In the continuous case, it is shown that for a reward function defined on $\mathbb{R}$ which is nonnegative, increasing, logconcave and right-continuous, the optimal stopping rule is of threshold form provided the underlying process is a spectrally negative L\'evy process. Furthermore, we also establish the necessity of logconcavity and monotonicity of a reward function in order for the optimal stopping rule to be of threshold form in the discrete (continuous, \$textit{resp.}$) case when the underlying process belongs to the class of Bernoulli random walks (Brownian motions, \textit{resp.}) with a downward drift.
Citation
Shoou-Ren Hsiau. Yi-Shen Lin. Yi-Ching Yao. "Logconcave reward functions and optimal stopping rules of threshold form." Electron. J. Probab. 19 1 - 18, 2014. https://doi.org/10.1214/EJP.v19-3745
Information