IDEAS home Printed from https://ideas.repec.org/a/plo/pcbi00/1007805.html
   My bibliography  Save this article

Learning action-oriented models through active inference

Author

Listed:
  • Alexander Tschantz
  • Anil K Seth
  • Christopher L Buckley

Abstract

Converging theories suggest that organisms learn and exploit probabilistic models of their environment. However, it remains unclear how such models can be learned in practice. The open-ended complexity of natural environments means that it is generally infeasible for organisms to model their environment comprehensively. Alternatively, action-oriented models attempt to encode a parsimonious representation of adaptive agent-environment interactions. One approach to learning action-oriented models is to learn online in the presence of goal-directed behaviours. This constrains an agent to behaviourally relevant trajectories, reducing the diversity of the data a model need account for. Unfortunately, this approach can cause models to prematurely converge to sub-optimal solutions, through a process we refer to as a bad-bootstrap. Here, we exploit the normative framework of active inference to show that efficient action-oriented models can be learned by balancing goal-oriented and epistemic (information-seeking) behaviours in a principled manner. We illustrate our approach using a simple agent-based model of bacterial chemotaxis. We first demonstrate that learning via goal-directed behaviour indeed constrains models to behaviorally relevant aspects of the environment, but that this approach is prone to sub-optimal convergence. We then demonstrate that epistemic behaviours facilitate the construction of accurate and comprehensive models, but that these models are not tailored to any specific behavioural niche and are therefore less efficient in their use of data. Finally, we show that active inference agents learn models that are parsimonious, tailored to action, and which avoid bad bootstraps and sub-optimal convergence. Critically, our results indicate that models learned through active inference can support adaptive behaviour in spite of, and indeed because of, their departure from veridical representations of the environment. Our approach provides a principled method for learning adaptive models from limited interactions with an environment, highlighting a route to sample efficient learning algorithms.Author summary: Within the popular framework of ‘active inference’, organisms learn internal models of their environments and use the models to guide goal-directed behaviour. A challenge for this framework is to explain how such models can be learned in practice, given (i) the rich complexity of natural environments, and (ii) the circular dependence of model learning and sensory sampling, which may lead to behaviourally suboptimal models being learned. Here, we develop an approach in which organisms selectively model those aspects of the environment that are relevant for acting in a goal-directed manner. Learning such ‘action-oriented’ models requires that agents balance information-seeking and goal-directed actions in a principled manner, such that both learning and information seeking are contextualised by goals. Using a combination of theory and simulation modelling, we show that this approach allows simple but effective models to be learned from relatively few interactions with the environment. Crucially, our results suggest that action-oriented models can support adaptive behaviour in spite of, and indeed because of, their departure from accurate representations of the environment.

Suggested Citation

  • Alexander Tschantz & Anil K Seth & Christopher L Buckley, 2020. "Learning action-oriented models through active inference," PLOS Computational Biology, Public Library of Science, vol. 16(4), pages 1-30, April.
  • Handle: RePEc:plo:pcbi00:1007805
    DOI: 10.1371/journal.pcbi.1007805
    as

    Download full text from publisher

    File URL: https://journals.plos.org/ploscompbiol/article?id=10.1371/journal.pcbi.1007805
    Download Restriction: no

    File URL: https://journals.plos.org/ploscompbiol/article/file?id=10.1371/journal.pcbi.1007805&type=printable
    Download Restriction: no

    File URL: https://libkey.io/10.1371/journal.pcbi.1007805?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    References listed on IDEAS

    as
    1. Amir Mitchell & Gal H. Romano & Bella Groisman & Avihu Yona & Erez Dekel & Martin Kupiec & Orna Dahan & Yitzhak Pilpel, 2009. "Adaptive prediction of environmental changes by microorganisms," Nature, Nature, vol. 460(7252), pages 220-224, July.
    2. Karl J Friston & Jean Daunizeau & Stefan J Kiebel, 2009. "Reinforcement Learning or Active Inference?," PLOS ONE, Public Library of Science, vol. 4(7), pages 1-13, July.
    3. Paul F. M. J. Verschure & Thomas Voegtlin & Rodney J. Douglas, 2003. "Environmentally mediated synergy between perception and behaviour in mobile robots," Nature, Nature, vol. 425(6958), pages 620-624, October.
    4. Guido Montúfar & Keyan Ghazi-Zahedi & Nihat Ay, 2015. "A Theory of Cheap Control in Embodied Systems," PLOS Computational Biology, Public Library of Science, vol. 11(9), pages 1-22, September.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Benjamin Patrick Evans & Mikhail Prokopenko, 2021. "A maximum entropy model of bounded rational decision-making with prior beliefs and market feedback," Papers 2102.09180, arXiv.org, revised May 2021.
    2. Mateus Joffily & Giorgio Coricelli, 2013. "Emotional Valence and the Free-Energy Principle," Post-Print halshs-00834063, HAL.
    3. David A Sivak & Matt Thomson, 2014. "Environmental Statistics and Optimal Regulation," PLOS Computational Biology, Public Library of Science, vol. 10(9), pages 1-12, September.
    4. Peter A. Corning, 2014. "Systems Theory and the Role of Synergy in the Evolution of Living Systems," Systems Research and Behavioral Science, Wiley Blackwell, vol. 31(2), pages 181-196, March.
    5. Jaroslav Vítků & Petr Dluhoš & Joseph Davidson & Matěj Nikl & Simon Andersson & Přemysl Paška & Jan Šinkora & Petr Hlubuček & Martin Stránský & Martin Hyben & Martin Poliak & Jan Feyereisl & Marek Ros, 2020. "ToyArchitecture: Unsupervised learning of interpretable models of the environment," PLOS ONE, Public Library of Science, vol. 15(5), pages 1-50, May.
    6. Francesco Donnarumma & Domenico Maisto & Giovanni Pezzulo, 2016. "Problem Solving as Probabilistic Inference with Subgoaling: Explaining Human Successes and Pitfalls in the Tower of Hanoi," PLOS Computational Biology, Public Library of Science, vol. 12(4), pages 1-30, April.
    7. Jennifer A. Loughmiller-Cardinal & James Scott Cardinal, 2023. "The Behavior of Information: A Reconsideration of Social Norms," Societies, MDPI, vol. 13(5), pages 1-27, April.
    8. Stefano Palminteri & Germain Lefebvre & Emma J Kilford & Sarah-Jayne Blakemore, 2017. "Confirmation bias in human reinforcement learning: Evidence from counterfactual feedback processing," PLOS Computational Biology, Public Library of Science, vol. 13(8), pages 1-22, August.
    9. Sébastien Boyer & Lucas Hérissant & Gavin Sherlock, 2021. "Adaptation is influenced by the complexity of environmental change during evolution in a dynamic environment," PLOS Genetics, Public Library of Science, vol. 17(1), pages 1-27, January.
    10. Gianluigi Mongillo & Hanan Shteingart & Yonatan Loewenstein, 2014. "The Misbehavior of Reinforcement Learning," Discussion Paper Series dp661, The Federmann Center for the Study of Rationality, the Hebrew University, Jerusalem.
    11. Ismael T Freire & Clement Moulin-Frier & Marti Sanchez-Fibla & Xerxes D Arsiwalla & Paul F M J Verschure, 2020. "Modeling the formation of social conventions from embodied real-time interactions," PLOS ONE, Public Library of Science, vol. 15(6), pages 1-22, June.
    12. Dongqi Han & Kenji Doya & Dongsheng Li & Jun Tani, 2024. "Synergizing habits and goals with variational Bayes," Nature Communications, Nature, vol. 15(1), pages 1-14, December.

    More about this item

    Statistics

    Access and download statistics

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:plo:pcbi00:1007805. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: ploscompbiol (email available below). General contact details of provider: https://journals.plos.org/ploscompbiol/ .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.