Access the full text.
Sign up today, get DeepDyve free for 14 days.
E. Moser, E. Kropff, M. Moser (2008)
Place cells, grid cells, and the brain's spatial representation system.Annual review of neuroscience, 31
S. Haykin (2007)
Cognitive Dynamic Systems2007 4th International Conference on Electrical and Electronics Engineering
D. Kirsh, P. Maglio (1994)
On Distinguishing Epistemic from Pragmatic ActionCogn. Sci., 18
Daniella Laureiro-Martínez, S. Brusoni, M. Zollo (2010)
The neuro-scientific foundations of the exploration-exploitation dilemmaJournal of Neuroscience, Psychology, and Economics, 3
D. Ognibene, G. Baldassarre (2015)
Ecological Active Vision: Four Bioinspired Principles to Integrate Bottom–Up and Adaptive Top–Down Attention Tested With a Simple Camera-Arm RobotIEEE Transactions on Autonomous Mental Development, 7
B. Broek, W. Wiegerinck, H. Kappen (2010)
Risk Sensitive Path Integral Control
R. Howard (1966)
Information Value TheoryIEEE Trans. Syst. Sci. Cybern., 2
(2010)
Perception-reason-action cycle: Models, algorithms and systems
W. Schultz, P. Apicella, T. Ljungberg (1993)
Responses of monkey dopamine neurons to reward and conditioned stimuli during successive steps of learning a delayed response task, 13
(1996)
Activity of dopamine neurons during learning in a familiar task context
(2012)
closing, various attempts have already been made to use predictive coding as a framework with which to interpret the role of dopamine in the context of music listening
(1999)
Coastal navigation: Robot navigation under uncertainty in dynamic environments
G. Talland (1961)
Plans and the Structure of BehaviorPsychosomatic Medicine, 23
Samuel McClure, N. Daw, P. Montague (2003)
A computational substrate for incentive salienceTrends in Neurosciences, 26
Dirk Walther, Ueli Rutishauser, C. Koch, P. Perona (2005)
Selective visual attention enables learning and recognition of multiple objects in cluttered scenesComput. Vis. Image Underst., 100
Alexander Andreopoulos, John Tsotsos (2012)
A Computational Learning Theory of Active Object Recognition Under UncertaintyInternational Journal of Computer Vision, 101
C. Souchay, M. Isingrini, Laurence Espagnet (2000)
Aging, episodic memory feeling-of-knowing, and frontal functioning.Neuropsychology, 14 2
H. Helmholtz, J. Southall (1962)
Treatise on Physiological Optics
Matthew Beal (2003)
Variational algorithms for approximate Bayesian inference
Feng Cao, Soumya Ray (2012)
Bayesian Hierarchical Reinforcement Learning
Susanne Still (2007)
Information-theoretic approach to interactive learningEPL (Europhysics Letters), 85
G. Pezzulo, M. Meer, C. Lansink, C. Pennartz (2014)
Internally generated sequences in learning and executing goal-directed behaviorTrends in Cognitive Sciences, 18
D. Ognibene, E. Chinellato, Miguel Sarabia, Y. Demiris (2013)
Contextual action recognition and target localization with an active allocation of attention on a humanoid robotBioinspiration & Biomimetics, 8
Pedro Ortega, Daniel Braun, Naftali Tishby (2014)
Monte Carlo methods for exact & efficient solution of the generalized optimality equations2014 IEEE International Conference on Robotics and Automation (ICRA)
P. Schwartenbeck, T. FitzGerald, R. Dolan, Karl Friston (2013)
Exploration, novelty, surprise, and free energy minimizationFrontiers in Psychology, 4
B. Richmond, L. Optican, M. Podell, H. Spitzer (1987)
Temporal encoding of two-dimensional patterns by single units in primate inferior temporal cortex. I. Response characteristics.Journal of neurophysiology, 57 1
C. Thornton (2010)
Some puzzles relating to the free-energy principle: comment on FristonTrends in Cognitive Sciences, 14
F. Wörgötter, B. Porr (2019)
Reinforcement learning
A. Clark (2013)
Whatever next? Predictive brains, situated agents, and the future of cognitive science.The Behavioral and brain sciences, 36 3
S. Kakade, P. Dayan (2002)
Dopamine: generalization and bonusesNeural networks : the official journal of the International Neural Network Society, 15 4-6
Daniel Braun, Pedro Ortega (2011)
Information, Utility and Bounded Rationality
M. Littman, R. Sutton, Satinder Singh (2001)
Predictive Representations of State
W. Bialek, I. Nemenman, Naftali Tishby (2000)
Predictability, Complexity, and LearningNeural Computation, 13
D. Bach, R. Dolan (2012)
Knowing how much you don't know: a neural organization of uncertainty estimatesNature Reviews Neuroscience, 13
J. Tani, S. Nolfi (1998)
Learning to perceive the world as articulated: an approach for hierarchical learning in sensory-motor systemsNeural networks : the official journal of the International Neural Network Society, 12 7-8
J. Schmidhuber (1991)
Curious model-building control systems[Proceedings] 1991 IEEE International Joint Conference on Neural Networks
N. Lepora, Uriel Martinez-Hernandez, T. Prescott (2013)
Active touch for robust perception under position uncertainty2013 IEEE International Conference on Robotics and Automation
R. Brooks (1991)
Intelligence without RepresentationArtif. Intell., 47
M. Botvinick, Marc Toussaint (2012)
Planning as inferenceTrends in Cognitive Sciences, 16
R. Kass, D. Steffey (1989)
Approximate Bayesian Inference in Conditionally Independent Hierarchical Models (Parametric Empirical Bayes Models)Journal of the American Statistical Association, 84
V. Salimpoor, M. Benovoy, K. Larcher, A. Dagher, R. Zatorre (2011)
Anatomically distinct dopamine release during anticipation and experience of peak emotion to musicNature Neuroscience, 14
John Henderson (2001)
A sensorimotor account of vision and visual consciousness.The Behavioral and brain sciences, 24 5
Ivan Herreros-Alonso, P. Verschure (2013)
Nucleo-olivary inhibition balances the interaction between the reactive and adaptive layers in motor controlNeural networks : the official journal of the International Neural Network Society, 47
(2010)
thus, this paper’s main contribution can also be seen as extending risk-sensitive control to a model that includes non-observable states
P. Dayan, Geoffrey Hinton, Radford Neal, R. Zemel (1995)
The Helmholtz MachineNeural Computation, 7
P. Verschure, T. Voegtlin, R. Douglas (2003)
Environmentally mediated synergy between perception and behaviour in mobile robotsNature, 425
HighWire Press (1781)
Philosophical Transactions of the Royal Society of LondonThe London Medical Journal, 2
P. Redgrave, K. Gurney (2006)
The short-latency dopamine signal: a role in discovering novel actions?Nature Reviews Neuroscience, 7
Proceedings of the Twenty-Third International Joint Conference on Artificial Intelligence Towards Active Event Recognition ∗
Karl Friston, Francesco Rigoli, D. Ognibene, C. Mathys, Thomas FitzGerald, G. Pezzulo (2016)
Corrigendum to "Active inference and epistemic value" [Cognitive Neuroscience, 6 (2015), 187-214. doi: 10.1080/17588928.2015.1020053]Cognitive Neuroscience, 7
Daniel Braun, Pedro Ortega (2014)
Information-Theoretic Bounded Rationality and ε-OptimalityEntropy, 16
Ethan Bromberg-Martin, Okihide Hikosaka (2009)
Midbrain Dopamine Neurons Signal Preference for Advance Information about Upcoming RewardsNeuron, 63
Y. Niv (2007)
Cost, Benefit, Tonic, PhasicAnnals of the New York Academy of Sciences, 1104
Daniel Braun, Pedro Ortega, Evangelos Theodorou, S. Schaal (2011)
Path integral control and bounded rationality2011 IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning (ADPRL)
J. Najemnik, W. Geisler (2005)
Optimal eye movement strategies in visual searchNature, 434
M. Humphries, M. Khamassi, K. Gurney (2012)
Dopaminergic Control of the Exploration-Exploitation Trade-Off via the Basal GangliaFrontiers in Neuroscience, 6
N. Bunzeck, E. Düzel (2006)
Absolute Coding of Stimulus Novelty in the Human Substantia Nigra/VTANeuron, 51
K. Jezek, Espen Henriksen, A. Treves, E. Moser, M. Moser (2011)
Theta-paced flickering between place-cell maps in the hippocampusNature, 478
Pedro Ortega, Daniel Braun (2012)
Thermodynamics as a theory of decision-making with information-processing costsProceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences, 469
Karl Friston, P. Schwartenbeck, T. FitzGerald, M. Moutoussis, T. Behrens, R. Dolan (2014)
The anatomy of choice: dopamine and decision-makingPhilosophical Transactions of the Royal Society B: Biological Sciences, 369
B. Olshausen, D. Field (1996)
Emergence of simple-cell receptive field properties by learning a sparse code for natural imagesNature, 381
A. Barto, Satinder Singh, N. Chentanez (2004)
Intrinsically Motivated Learning of Hierarchical Collections of Skills
R. Sutton, A. Barto (1998)
Reinforcement Learning: An IntroductionIEEE Trans. Neural Networks, 9
H. Mushiake, N. Saito, K. Sakamoto, Y. Itoyama, J. Tanji (2006)
Activity in the Lateral Prefrontal Cortex Reflects Multiple Steps of Future Events in Action PlansNeuron, 50
M. Lungarella, O. Sporns (2006)
Mapping Information Flow in Sensorimotor NetworksPLoS Computational Biology, 2
P. Cisek (2011)
Modelling Natural Action Selection: Cortical mechanisms of action selection: the affordance competition hypothesis
G. Pezzulo, Francesco Rigoli, Fabian Chersi (2013)
The Mixed Instrumental Controller: Using Value of Information to Combine Habitual Choice and Mental SimulationFrontiers in Psychology, 4
A. Klyubin, D. Polani, Chrystopher Nehaniv (2008)
Keep Your Options Open: An Information-Based Driving Principle for Sensorimotor SystemsPLoS ONE, 3
P. Verschure, C. Pennartz, G. Pezzulo (2014)
The why, what, where, when and how of goal-directed choice: neuronal and computational principlesPhilosophical Transactions of the Royal Society B: Biological Sciences, 369
(1972)
Epistemics and economics
Massimiliano Schembri, M. Mirolli, G. Baldassarre (2007)
Evolving internal reinforcers for an intrinsically motivated reinforcement-learning robot2007 IEEE 6th International Conference on Development and Learning
D. Ognibene, G. Pezzulo, G. Baldassarre (2010)
Learning to Look in Different Environments: An Active-Vision Model Which Learns and Readapts Visual Routines
Manuel Sánchez-Montañés, P. König, P. Verschure (2002)
Learning sensory maps with real-world stimuli in real time using a biophysically realistic learning ruleIEEE transactions on neural networks, 13 3
Ece Kamar, E. Horvitz (2013)
Light at the end of the tunnel: a Monte Carlo approach to computing value of information
V. Ferrari, Andrew Zisserman (2007)
Advances in Neural Information Processing Systems (NIPS)
A. Battersby (1968)
Plans and the Structure of BehaviorJournal of the Operational Research Society, 19
R. Krebs, B. Schott, H. Schütze, E. Düzel (2009)
The novelty exploration bonus and its attentional modulationNeuropsychologia, 47
S. Vijayakumar, Marc Toussaint, Georgios Petkos, M. Howard (2009)
Planning and Moving in Dynamic Environments
L. Itti, C. Koch (2001)
Computational modelling of visual attentionNature Reviews Neuroscience, 2
G. Viswanathan, S. Buldyrev, S. Havlin, M. Luz, E. Raposo, H. Stanley (1999)
Optimizing the success of random searchesNature, 401
N. Daw, J. O’Doherty, P. Dayan, B. Seymour, R. Dolan (2006)
Cortical substrates for exploratory decisions in humansNature, 441
W. Penny, P. Zeidman, N. Burgess (2013)
Forward and Backward Inference in Spatial CognitionPLoS Computational Biology, 9
L. Itti, P. Baldi (2005)
Bayesian surprise attracts human attentionVision Research, 49
Alexander Schneider, Jürgen Sturm, C. Stachniss, M. Reisert, H. Burkhardt, Wolfram Burgard (2009)
Object identification with tactile sensors using bag-of-features2009 IEEE/RSJ International Conference on Intelligent Robots and Systems
N. Lepora, K. Gurney (2012)
The Basal Ganglia Optimize Decision Making over General Perceptual HypothesesNeural Computation, 24
Daniel Little, F. Sommer (2013)
Learning and exploration in action-perception loopsFrontiers in Neural Circuits, 7
Alec Solway, M. Botvinick (2012)
Goal-directed decision making as probabilistic inference: a computational framework and potential neural correlates.Psychological review, 119 1
G. Baldassarre, M. Mirolli (2013)
Intrinsically Motivated Learning in Natural and Artificial SystemsIntrinsically Motivated Learning in Natural and Artificial Systems
J. Daunizeau, H. Ouden, M. Pessiglione, S. Kiebel, K. Stephan, Karl Friston (2010)
Observing the Observer (I): Meta-Bayesian Models of Learning and Decision-MakingPLoS ONE, 5
Karl Friston (2010)
The free-energy principle: a unified brain theory?Nature Reviews Neuroscience, 11
Pierre-Yves Oudeyer, F. Kaplan (2007)
What is Intrinsic Motivation? A Typology of Computational ApproachesFrontiers in Neurorobotics, 1
H. Barlow (2012)
Possible Principles Underlying the Transformations of Sensory Messages
Karl Friston, C. Thornton, A. Clark (2012)
Free-Energy Minimization and the Dark-Room ProblemFrontiers in Psychology, 3
R. Beer (1995)
A Dynamical Systems Perspective on Agent-Environment InteractionArtif. Intell., 72
Nantachai Sornkarn, M. Howard, T. Nanayakkara (2014)
Internal impedance control helps information gain in embodied perception2014 IEEE International Conference on Robotics and Automation (ICRA)
T. Lee, D. Mumford (2003)
Hierarchical Bayesian inference in the visual cortex.Journal of the Optical Society of America. A, Optics, image science, and vision, 20 7
P. Dayan, Geoffrey Hinton (1997)
Using Expectation-Maximization for Reinforcement LearningNeural Computation, 9
Sieve Flatt (1994)
Nursing process: I think, therefore I am.Nursing standard (Royal College of Nursing (Great Britain) : 1987), 8 47
Thomas FitzGerald, R. Dolan, Karl Friston (2014)
Model averaging, optimal inference, and habit formationFrontiers in Human Neuroscience, 8
D. Ognibene, Nicola Volpi, G. Pezzulo, G. Baldassarre (2013)
Learning Epistemic Actions in Model-Free Memory-Free Reinforcement Learning: Experiments with a Neuro-robotic Model
C. Gros (2010)
Cognition and Emotion: Perspectives of a Closing GapCognitive Computation, 2
M. Botvinick, James An (2008)
Goal-directed decision making in prefrontal cortex: a computational frameworkAdvances in neural information processing systems, 21
D. Ballard, Dmitry Kit, C. Rothkopf, Brian Sullivan (2013)
A hierarchical modular architecture for embodied cognition.Multisensory research, 26 1-2
(2011)
A tutorial on variational Bayes
H. Harlow (1950)
Learning and satiation of response in intrinsically motivated complex puzzle performance by monkeys.Journal of comparative and physiological psychology, 43 4
D. Berlyne (2014)
Conflict, arousal, and curiosity
C. Moulin, C. Souchay (2014)
Epistemic Feelings and Memory
J. Fuster (2004)
Upper processing stages of the perception–action cycleTrends in Cognitive Sciences, 8
H. Barlow (1974)
Inductive Inference, Coding, Perception, and LanguagePerception, 3
R. Dolan, P. Dayan (2013)
Goals and Habits in the BrainNeuron, 80
Marc Toussaint, A. Storkey (2006)
Probabilistic inference for solving discrete and continuous state Markov Decision ProcessesProceedings of the 23rd international conference on Machine learning
Amarjeet Singh, Andreas Krause, Carlos Guestrin, W. Kaiser (2014)
Efficient Informative Sensing using Multiple RobotsJ. Artif. Intell. Res., 34
P. Dayan (2009)
Dopamine, reinforcement learning, and addiction.Pharmacopsychiatry, 42 Suppl 1
B. Wittmann, N. Daw, B. Seymour, R. Dolan (2008)
Striatal Activity Underlies Novelty-Based Choice in HumansNeuron, 58
R. Canolty, E. Edwards, S. Dalal, M. Soltani, S. Nagarajan, H. Kirsch, M. Berger, N. Barbaro, R. Knight (2006)
High Gamma Power Is Phase-Locked to Theta Oscillations in Human NeocortexScience, 313
W. Schultz (1998)
Predictive reward signal of dopamine neurons.Journal of neurophysiology, 80 1
C. Atkeson, J. Santamaría (1997)
A comparison of direct and model-based reinforcement learningProceedings of International Conference on Robotics and Automation, 4
Karl Friston, P. Schwartenbeck, T. FitzGerald, M. Moutoussis, T. Behrens, R. Dolan (2013)
The anatomy of choice: active inference and agencyFrontiers in Human Neuroscience, 7
(2011)
Interpretations of such data have been made variously in the context of the Incentive Salience Theory and/or in the context of the Reward
César Rennó-Costa, J. Lisman, P. Verschure (2014)
A Signature of Attractor Dynamics in the CA3 Region of the HippocampusPLoS Computational Biology, 10
J. O’Doherty, P. Dayan, J. Schultz, R. Deichmann, Karl Friston, R. Dolan (2004)
Dissociable Roles of Ventral and Dorsal Striatum in Instrumental ConditioningScience, 304
G. Miller, E. Galanter, K. Pribram (1960)
Plans and the structure of behavior
William Cohen, R. Schapire, Y. Singer
Âóùöòòð Óó Öøø¬ Blockin
H. Kappen, V. Gómez, M. Opper (2009)
Optimal control as a graphical model inference problemMachine Learning, 87
(2013)
facilitate the identification of the value of FEP in understanding mind and brain
L. Optican, B. Richmond (1987)
Temporal encoding of two-dimensional patterns by single units in primate inferior temporal cortex. III. Information theoretic analysis.Journal of neurophysiology, 57 1
T. Furmston, D. Barber (2010)
Variational methods for Reinforcement Learning
L. Kaelbling, M. Littman, A. Cassandra (1998)
Planning and Acting in Partially Observable Stochastic DomainsArtif. Intell., 101
A. Clark (2013)
The many faces of precision (Replies to commentaries on “Whatever next? Neural prediction, situated agents, and the future of cognitive science”)Frontiers in Psychology, 4
Kimberlee D'ardenne, Samuel McClure, L. Nystrom, J. Cohen (2008)
BOLD Responses Reflecting Dopaminergic Signals in the Human Ventral Tegmental AreaScience, 319
Zachary Pezzementi, E. Plaku, Caitlin Reyda, Gregory Hager (2011)
Tactile-Object Recognition From Appearance InformationIEEE Transactions on Robotics, 27
H. Attias (2003)
Planning by Probabilistic Inference
Naftali Tishby, D. Polani (2011)
Information Theory of Decisions and Actions
Neurorobotics Original Research Article
Karl Friston, Spyridon Samothrakis, P. Montague (2012)
Active inference and agency: optimal control without cost functionsBiological Cybernetics, 106
A. Koriat, H. Ma'ayan, Ravit Nussinson (2006)
The intricate relationships between monitoring and control in metacognition: lessons for the cause-and-effect relation between subjective experience and behavior.Journal of experimental psychology. General, 135 1
B. Balleine, A. Dickinson (1998)
Goal-directed instrumental action: contingency and incentive learning and their cortical substratesNeuropharmacology, 37
J. Lau, F. Rijsdijk, T. Eley (2006)
I think, therefore I am: a twin study of attributional style in adolescents.Journal of child psychology and psychiatry, and allied disciplines, 47 7
Susanne Still, Doina Precup (2012)
An information-theoretic approach to curiosity-driven reinforcement learningTheory in Biosciences, 131
P. Schwartenbeck, Thomas FitzGerald, C. Mathys, R. Dolan, Karl Friston (2014)
The Dopaminergic Midbrain Encodes the Expected Certainty about Desired OutcomesCerebral Cortex (New York, NY), 25
(2010)
acknowledge, hyperpriors on the (expected) precision are crucial for weighting prediction errors (Clark
Karl Friston, Rick Adams, Laurent Perrinet, M. Breakspear (2012)
Perceptions as Hypotheses: Saccades as ExperimentsFrontiers in Psychology, 3
K. Scherer, E. Coutinho, T. Cochrane, B. Fantini, K. Scherer (2013)
How music creates emotion: a multifactorial process approach
E. Deci, R. Ryan (1975)
Intrinsic Motivation and Self-Determination in Human Behavior
Rajesh Rao, D. Ballard (1999)
Predictive coding in the visual cortex: a functional interpretation of some extra-classical receptive-field effects.Nature Neuroscience, 2
R. Linsker (1990)
Perceptual neural organization: some approaches based on network models and information theory.Annual review of neuroscience, 13
E. Oja (1989)
Neural Networks, Principal Components, and SubspacesInt. J. Neural Syst., 1
K. Gurney, T. Prescott, P. Redgrave (2001)
A computational model of action selection in the basal ganglia. I. A new functional anatomyBiological Cybernetics, 84
G. Pezzulo, C. Castelfranchi (2009)
Thinking as the control of imagination: a conceptual framework for goal-directed systemsPsychological Research PRPF, 73
N. Daw, K. Doya (2006)
The computational neurobiology of learning and rewardCurrent Opinion in Neurobiology, 16
C. Fiorillo, P. Tobler, W. Schultz (2003)
Discrete Coding of Reward Probability and Uncertainty by Dopamine NeuronsScience, 299
Blai Bonet, Hector Geffner (2014)
Belief Tracking for Planning with Sensing: Width, Complexity and ApproximationsJ. Artif. Intell. Res., 50
J. Cohen, Samuel McClure, Angela Yu (2007)
Should I stay or should I go? How the human brain manages the trade-off between exploitation and explorationPhilosophical Transactions of the Royal Society B: Biological Sciences, 362
K. Berridge (2007)
The debate over dopamine’s role in reward: the case for incentive saliencePsychopharmacology, 191
L. Gebauer, M. Kringelbach, P. Vuust (2012)
Ever-changing cycles of musical pleasure: The role of dopamine and anticipation.Psychomusicology: Music, Mind and Brain, 22
Neil Bruce, John Tsotsos (2009)
Saliency, attention, and visual search: an information theoretic approach.Journal of vision, 9 3
B. Martino, S. Fleming, Neil Garrett, R. Dolan (2012)
Confidence in value-based choiceNature neuroscience, 16
W. Wendland (1991)
Variational Methods for BEM
T. Nelson (1990)
Metamemory: A Theoretical Framework and New FindingsPsychology of Learning and Motivation, 26
N. Daw, Y. Niv, P. Dayan (2005)
Uncertainty-based competition between prefrontal and dorsolateral striatal systems for behavioral controlNature Neuroscience, 8
Andreas Krause, Carlos Guestrin (2005)
Optimal Nonmyopic Value of Information in Graphical Models - Efficient Algorithms and Theoretical Limits
M. Pessiglione, B. Seymour, G. Flandin, R. Dolan, C. Frith (2006)
Dopamine-dependent prediction errors underpin reward-seeking behaviour in humansNature, 442
S. Fleming, R. Dolan, C. Frith (2012)
Metacognition: computation, biology and functionPhilosophical Transactions of the Royal Society B: Biological Sciences, 367
We offer a formal treatment of choice behavior based on the premise that agents minimize the expected free energy of future outcomes. Crucially, the negative free energy or quality of a policy can be decomposed into extrinsic and epistemic (or intrinsic) value. Minimizing expected free energy is therefore equivalent to maximizing extrinsic value or expected utility (defined in terms of prior preferences or goals), while maximizing information gain or intrinsic value (or reducing uncertainty about the causes of valuable outcomes). The resulting scheme resolves the exploration-exploitation dilemma: Epistemic value is maximized until there is no further information gain, after which exploitation is assured through maximization of extrinsic value. This is formally consistent with the Infomax principle, generalizing formulations of active vision based upon salience (Bayesian surprise) and optimal decisions based on expected utility and risk-sensitive (Kullback-Leibler) control. Furthermore, as with previous active inference formulations of discrete (Markovian) problems, ad hoc softmax parameters become the expected (Bayes-optimal) precision of beliefs about, or confidence in, policies. This article focuses on the basic theory, illustrating the ideas with simulations. A key aspect of these simulations is the similarity between precision updates and dopaminergic discharges observed in conditioning paradigms.
Cognitive Neuroscience – Taylor & Francis
Published: Oct 2, 2015
Keywords: Active inference; Agency; Bayesian inference; Bounded rationality; Free energy; Utility theory; Information gain; Bayesian surprise; Epistemic value; Exploration; Exploitation
Read and print from thousands of top scholarly journals.
Already have an account? Log in
Bookmark this article. You can see your Bookmarks on your DeepDyve Library.
To save an article, log in first, or sign up for a DeepDyve account if you don’t already have one.
Copy and paste the desired citation format or use the link below to download a file formatted for EndNote
Access the full text.
Sign up today, get DeepDyve free for 14 days.
All DeepDyve websites use cookies to improve your online experience. They were placed on your computer when you launched this website. You can change your cookie settings through your browser.