Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 134 tok/s
Gemini 2.5 Pro 41 tok/s Pro
GPT-5 Medium 23 tok/s Pro
GPT-5 High 18 tok/s Pro
GPT-4o 86 tok/s Pro
Kimi K2 194 tok/s Pro
GPT OSS 120B 432 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

Reframing the Expected Free Energy: Four Formulations and a Unification (2402.14460v1)

Published 22 Feb 2024 in cs.AI

Abstract: Active inference is a leading theory of perception, learning and decision making, which can be applied to neuroscience, robotics, psychology, and machine learning. Active inference is based on the expected free energy, which is mostly justified by the intuitive plausibility of its formulations, e.g., the risk plus ambiguity and information gain / pragmatic value formulations. This paper seek to formalize the problem of deriving these formulations from a single root expected free energy definition, i.e., the unification problem. Then, we study two settings, each one having its own root expected free energy definition. In the first setting, no justification for the expected free energy has been proposed to date, but all the formulations can be recovered from it. However, in this setting, the agent cannot have arbitrary prior preferences over observations. Indeed, only a limited class of prior preferences over observations is compatible with the likelihood mapping of the generative model. In the second setting, a justification of the root expected free energy definition is known, but this setting only accounts for two formulations, i.e., the risk over states plus ambiguity and entropy plus expected energy formulations.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (24)
  1. Learning perception and planning with deep active inference. In 2020 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2020, Barcelona, Spain, May 4-8, 2020, pages 3952–3956. IEEE, 2020. doi: 10.1109/ICASSP40776.2020.9054364. URL https://doi.org/10.1109/ICASSP40776.2020.9054364.
  2. Branching time active inference: Empirical study and complexity class analysis. Neural Networks, 152:450–466, 2022a. ISSN 0893-6080. doi: https://doi.org/10.1016/j.neunet.2022.05.010. URL https://www.sciencedirect.com/science/article/pii/S0893608022001824.
  3. Branching time active inference: The theory and its generality. Neural Networks, 151:295–316, 2022b. ISSN 0893-6080. doi: https://doi.org/10.1016/j.neunet.2022.03.036. URL https://www.sciencedirect.com/science/article/pii/S0893608022001149.
  4. Multi-modal and multi-factor branching time active inference, 2022c. URL https://arxiv.org/abs/2206.12503.
  5. Branching Time Active Inference with Bayesian Filtering. Neural Computation, 34(10):2132–2144, 09 2022d. ISSN 0899-7667. doi: 10.1162/neco˙a˙01529. URL https://doi.org/10.1162/neco_a_01529.
  6. Deconstructing deep active inference, 2023. URL https://arxiv.org/abs/2303.01618.
  7. Active inference in openai gym: A paradigm for computational investigations into psychiatric illness. Biological Psychiatry: Cognitive Neuroscience and Neuroimaging, 3(9):809 – 818, 2018. ISSN 2451-9022. doi: https://doi.org/10.1016/j.bpsc.2018.06.010. URL http://www.sciencedirect.com/science/article/pii/S2451902218301617. Computational Methods and Modeling in Psychiatry.
  8. Dopamine, reward learning, and active inference. Frontiers in Computational Neuroscience, 9:136, 2015. ISSN 1662-5188. doi: 10.3389/fncom.2015.00136. URL https://www.frontiersin.org/article/10.3389/fncom.2015.00136.
  9. Deep active inference agents using Monte-Carlo methods. In Hugo Larochelle, Marc’Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin, editors, Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, 2020. URL https://proceedings.neurips.cc/paper/2020/hash/865dfbde8a344b44095495f3591f7407-Abstract.html.
  10. A tutorial on variational Bayesian inference. Artificial Intelligence Review, 38(2):85–95, Aug 2012. ISSN 1573-7462. doi: 10.1007/s10462-011-9236-8. URL https://doi.org/10.1007/s10462-011-9236-8.
  11. Active inference and learning. Neuroscience & Biobehavioral Reviews, 68:862 – 879, 2016. ISSN 0149-7634. doi: https://doi.org/10.1016/j.neubiorev.2016.06.022.
  12. Sophisticated Inference. Neural Computation, 33(3):713–763, 03 2021. ISSN 0899-7667. doi: 10.1162/neco˙a˙01351. URL https://doi.org/10.1162/neco_a_01351.
  13. The graphical brain: Belief propagation and active inference. Network Neuroscience, 1(4):381–414, 12 2017. ISSN 2472-1751. doi: 10.1162/NETN˙a˙00018. URL https://doi.org/10.1162/NETN_a_00018.
  14. Bayesian surprise attracts human attention. Vision Research, 49(10):1295 – 1306, 2009. ISSN 0042-6989. doi: https://doi.org/10.1016/j.visres.2008.09.007. URL http://www.sciencedirect.com/science/article/pii/S0042698908004380. Visual Attention: Psychophysics, electrophysiology and neuroimaging.
  15. Edwin T Jaynes. Information theory and statistical mechanics. Physical review, 106(4):620, 1957a.
  16. Edwin T Jaynes. Information theory and statistical mechanics. ii. Physical review, 108(2):171, 1957b.
  17. Probabilistic graphical models: principles and techniques. MIT press, 2009.
  18. Factor graphs and the sum-product algorithm. IEEE Transactions on information theory, 47(2):498–519, 2001.
  19. Beren Millidge. Combining active inference and hierarchical predictive coding: A tutorial introduction and case study., 2019. URL https://doi.org/10.31234/osf.io/kf6wc.
  20. Neuronal message passing using mean-field, bethe, and marginal approximations. Scientific reports, 9(1):1889, 2019.
  21. Active inference: the free energy principle in mind, brain, and behavior. MIT Press, 2022.
  22. End-to-end pixel-based deep active inference for body perception and action. In Joint IEEE 10th International Conference on Development and Learning and Epigenetic Robotics, ICDL-EpiRob 2020, Valparaiso, Chile, October 26-30, 2020, pages 1–8. IEEE, 2020. doi: 10.1109/ICDL-EpiRob48136.2020.9278105. URL https://doi.org/10.1109/ICDL-EpiRob48136.2020.9278105.
  23. Computational mechanisms of curiosity and goal-directed exploration. bioRxiv, 2018. doi: 10.1101/411272. URL https://www.biorxiv.org/content/early/2018/09/07/411272.
  24. Variational message passing. Journal of Machine Learning Research, 6(4), 2005.
Citations (3)

Summary

  • The paper unifies four formulations of Expected Free Energy by deriving them from a singular foundational definition.
  • It employs a generative model with variational inference to balance observational accuracy with efficient policy exploration.
  • It highlights limitations in prior preferences and likelihood mappings, paving the way for more robust theoretical frameworks.

Reframing the Expected Free Energy: Four Formulations and a Unification

Introduction

The concept of active inference is rooted in the modeling of perception, decision-making, and learning, offering applications that extend into fields like neuroscience, robotics, and machine learning. At the core of active inference is the Expected Free Energy (EFE), which comprises various formulations that are largely derived for their intuitive appeal. The unification problem as discussed in this paper addresses the challenge of deriving these formulations from a singular foundational definition of EFE.

Generative Model and Variational Framework

Active inference employs a generative model encapsulating hidden states, observations, and actions across time, adhering to a Partially Observable Markov Decision Process (POMDP) structure. To infer hidden states from observations, a variational inference approach is used, replacing the intractable true posterior with a more computationally feasible variational distribution. The agent minimizes the Variational Free Energy (VFE) to achieve this approximation, balancing accuracy and complexity.

Planning with Expected Free Energy

The decision-making component in active inference leverages EFE to evaluate and choose policies. EFE quantifies the trade-off between exploration and exploitation, aiming to maximize information gain alongside pragmatic value from the agent's prior preferences. The computational complexity of this process necessitates efficient policy exploration strategies such as Monte-Carlo tree search or sophisticated inference.

Unification Problem and EFE Formulations

The unification problem seeks to derive four known EFE formulations: Risk Over States and Ambiguity, Risk Over Observations and Ambiguity, Information Gain and Pragmatic Value, and Expected Energy and Entropy, from a unified definition. By establishing commonality through forecast and target distributions in theoretical frameworks, the paper aims to trace these formulations back to a foundational EFE description.

Analysis of EFE Formulations

  • Risk Over Observations Plus Ambiguity: The paper posits Risk Over Observations plus Ambiguity as a potential root EFE definition. It explores the equivalences and derivations between different formulations, revealing that while this formulation allows recovery of all four EFE expressions, it lacks a formal theoretical justification.
  • Risk Over States Plus Ambiguity: This formulation emerges as an upper bound of EFE and is currently the only one with a theoretical underpinning. However, it only supports two of the four EFE expressions.

The derivations rely critically on assumptions about distributions and marginalization, illustrating limitations like the restricted class of valid prior preferences over observations and potential inconsistency issues.

Limitations and Challenges

One prominent limitation is the restriction on arbitrary prior preferences due to compatibility issues with likelihood mappings. Additionally, there is a lack of justification for the Risk Over Observations plus Ambiguity formulation, raising questions about the theoretical integrity of the unified EFE derivation problem.

Conclusion

The attempt to unify varying formulations of EFE presents significant insights but also reveals critical constraints in theoretical assumptions and justifications. Future research should focus on developing a robust theoretical framework that encompasses all EFE formulations while resolving inconsistencies between prior preferences and likelihood mappings. Further exploration is required to bridge the theoretical discussions with practical implementations, particularly in the context of deep active inference applications.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube