Access the full text.
Sign up today, get DeepDyve free for 14 days.
B. Kröse (1995)
Learning from delayed rewardsRobotics Auton. Syst., 15
J. Kelly (1956)
A new interpretation of information rateIRE Trans. Inf. Theory, 2
(2001)
Prospect Theory and Asset Prices
D. Bertsekas (2009)
Neuro-Dynamic Programming
Colin Camerer (2003)
Behavioral Game Theory: Experiments in Strategic Interaction
C. Watkins, P. Dayan (2004)
Technical Note: Q-LearningMachine Learning, 8
Nicholas Polson, M. Sorensen (2011)
A simulation-based approach to stochastic dynamic programmingApplied Stochastic Models in Business and Industry, 27
Charles Holt, Susan Laury (2002)
Risk Aversion and Incentive EffectsAndrew Young School of Policy Studies Research Paper Series
R. Thaler, Eric Johnson (1990)
Gambling with the house money and trying to break even: the effects of prior outcomes on risky choiceManagement Science, 36
David Banks, F. Petralia, Shouqiang Wang (2011)
Adversarial risk analysis: Borel gamesApplied Stochastic Models in Business and Industry, 27
N. Barberis (2009)
A Model of Casino GamblingBehavioral & Experimental Finance eJournal
H. Jeffreys (1922)
A Treatise on ProbabilityNature, 109
(2012)
Asset Allocation: A Bayesian Perspective
David Kreps, Evan Porteus (1978)
Temporal Resolution of Uncertainty and Dynamic Choice TheoryEconometrica, 46
F. Ramsey (2001)
The Foundations of Mathematics and Other Logical Essays
F. Wörgötter, B. Porr (2019)
Reinforcement learning
T. Post, Martijn Assem, Guido Baltussen, R. Thaler (2008)
Deal or No Deal? Decision Making under Risk in a Large-Payoff Game ShowEFA Submission Session (check box to submit to EFA 2006 Zurich Meeting)
M. Puterman (1994)
Markov Decision Processes: Discrete Stochastic Dynamic Programming
C. Smith, John Neumann, O. Morgenstern (1945)
Theory of Games and Economic Behavior.Journal of the American Statistical Association, 40
Michael Brandt (2010)
Measuring the Time-Varying Risk-Return Relation from the Cross-Section of Equity Returns
M. Bombardini, Francesco Trebbi (2012)
RISK AVERSION AND EXPECTED UTILITY THEORY: AN EXPERIMENT WITH LARGE AND SMALL STAKESJournal of the European Economic Association, 10
R. Howard (1960)
Dynamic Programming and Markov Processes
J. Campbell, J. Cochrane (1995)
By Force of Habit: A Consumption‐Based Explanation of Aggregate Stock Market BehaviorJournal of Political Economy, 107
P. Whittle (1982)
Optimization Over Time
(1952)
Annali di Economia
(1952)
Sulla preferibilitá
J. Neumann, O. Morgenstern (1944)
The Theory of Games and Economic Behaviour
J. Gittins (1979)
Bandit processes and dynamic allocation indicesJournal of the royal statistical society series b-methodological, 41
Noname manuscript No. (will be inserted by the editor) Risk, Ambiguity, and State-Preference Theory
William Uther (2004)
Markov Decision Processes
D. Bertsekas (1995)
Dynamic programming and optimal control, 3rd Edition
In this paper, we derive an optimal strategy for the popular Deal or No Deal game show. To do this, we use Q‐learning methods, which quantify the continuation value inherent in sequential decision making in the game. We then analyze two contestants, Frank and Susanne, risky choices from the European version of the game. Given their choices and our optimal strategy, we find what their implied bounds would be on their levels of risk aversion. Previous empirical evidence in risky decision making has suggested that past outcomes affect future choices and that contestants have time‐varying risk aversion. We demonstrate that the strategies of Frank and Susanne are consistent with constant risk aversion levels except for their final risk‐seeking choice. We conclude with directions for future research. Copyright © 2013 John Wiley & Sons, Ltd.
Applied Stochastic Models in Business and Industry – Wiley
Published: Jan 1, 2014
Keywords: ; ;
Read and print from thousands of top scholarly journals.
Already have an account? Log in
Bookmark this article. You can see your Bookmarks on your DeepDyve Library.
To save an article, log in first, or sign up for a DeepDyve account if you don’t already have one.
Copy and paste the desired citation format or use the link below to download a file formatted for EndNote
Access the full text.
Sign up today, get DeepDyve free for 14 days.
All DeepDyve websites use cookies to improve your online experience. They were placed on your computer when you launched this website. You can change your cookie settings through your browser.