Access the full text.
Sign up today, get DeepDyve free for 14 days.
K. Chatterjee (2019)
Stochastic gamesGame Theory
Y. Moses, Moshe Tennenholtz (1996)
Multi-entity Models
(1995)
Artificial Intelligence: A Modern Approach
L. Kaelbling, M. Littman, A. Moore (1996)
Reinforcement Learning: A SurveyJ. Artif. Intell. Res., 4
B. Bollobás (2015)
The Probabilistic MethodFundamentals of Ramsey Theory
D. Blackwell (1956)
An analog of the minimax theorem for vector payoffs.Pacific Journal of Mathematics, 6
R. Aumann (1995)
Repeated Games with Incomplete Information
M. Wellman, J. Doyle (1992)
Proceedings of the 1st International Conference on AI Planning Systems
Michael Wellman (1985)
REASONING ABOUT PREFERENCE MODELS
J. Milnor (1954)
Decision Processes
L.G. Valiant (1984)
A theory of the learnableComm. ACM, 27
J.C. Harsanyi (1967)
Games with incomplete information played by bayesian players, Parts i, ii, iiiManagement Science, 14
J. Harsanyi (1968)
Games with Incomplete Information Played by “Bayesian” Players Part II. Bayesian Equilibrium PointsManagement Science, 14
Michael Wellman, J. Doyle (1992)
Modular utility representation for decision-theoretic planning
D. Fudenberg, D. Levine (1998)
The Theory of Learning in Games
S. Sen (1996)
IJCAI-95 Workshop on Adaptation and Learning in Multiagent SystemsAi Magazine, 17
David Kreps (2020)
A Course in Microeconomic Theory
H. Chernoff (1952)
A Measure of Asymptotic Efficiency for Tests of a Hypothesis Based on the sum of ObservationsAnnals of Mathematical Statistics, 23
D. Lindley, L. Savage (1955)
The Foundations of StatisticsThe Mathematical Gazette, 57
David Kreps (1988)
Notes On The Theory Of Choice
M. Littman (1994)
Markov Games as a Framework for Multi-Agent Reinforcement Learning
D. Fudenberg, J. Tirole (1991)
Game Theory
C. Papadimitriou (1983)
Games against nature24th Annual Symposium on Foundations of Computer Science (sfcs 1983)
Edward Madden, R. Luce, H. Raiffa (1958)
GAMES AND DECISIONS; INTRODUCTION AND CRITICAL SURVEY.American Sociological Review, 23
David Editor
Artificial Intelligence and Language Processing a Theory of the Learnable
Dov Monderer, Moshe Tennenholtz (1997)
Dynamic Non-Bayesian Decision MakingArXiv, cs.AI/9711104
C. Papadimitriou, M. Yannakakis (1989)
Shortest Paths Without a MapTheor. Comput. Sci., 84
L.S. Shapley (1953)
Stochastic gamesProc. Nat. Acad. Sci. U.S.A., 39
R. Brafman, Moshe Tennenholtz (1997)
On the Axiomatization of Qualitative Decision Criteria
We consider a group of several non-Bayesian agents that can fully coordinate their activities and share their past experience in order to obtain a joint goal in face of uncertainty. The reward obtained by each agent is a function of the environment state but not of the action taken by other agents in the group. The environment state (controlled by Nature) may change arbitrarily, and the reward function is initially unknown. Two basic feedback structures are considered. In one of them — the perfect monitoring case — the agents are able to observe the previous environment state as part of their feedback, while in the other — the imperfect monitoring case — all that is available to the agents are the rewards obtained. Both of these settings refer to partially observable processes, where the current environment state is unknown. Our study refers to the competitive ratio criterion. It is shown that, for the imperfect monitoring case, there exists an efficient stochastic policy that ensures that the competitive ratio is obtained for all agents at almost all stages with an arbitrarily high probability, where efficiency is measured in terms of rate of convergence. It is also shown that if the agents are restricted only to deterministic policies then such a policy does not exist, even in the perfect monitoring case.
Annals of Mathematics and Artificial Intelligence – Springer Journals
Published: Sep 30, 2004
Read and print from thousands of top scholarly journals.
Already have an account? Log in
Bookmark this article. You can see your Bookmarks on your DeepDyve Library.
To save an article, log in first, or sign up for a DeepDyve account if you don’t already have one.
Copy and paste the desired citation format or use the link below to download a file formatted for EndNote
Access the full text.
Sign up today, get DeepDyve free for 14 days.
All DeepDyve websites use cookies to improve your online experience. They were placed on your computer when you launched this website. You can change your cookie settings through your browser.