MDP: Markov Decision Process—a mathematical framework for modeling decision-making where outcomes are partly random and partly under the control of a decision maker
KL divergence: A measure of how one probability distribution differs from a second, reference probability distribution
Free Energy: A thermodynamic quantity representing the amount of internal energy available to perform work; in this context, used to balance reward and information cost
Stochastic Thermodynamics: A branch of physics dealing with thermodynamic quantities (heat, work, entropy) at the level of individual trajectories in stochastic systems
BAMDP: Bayesian Adaptive MDP—an extension of MDPs where the transition probabilities are unknown and learned via Bayesian inference
Maxwell's Demon: A thought experiment where an entity uses information about particle speeds to reduce entropy, demonstrating the link between information and energy
HJB principle: Hamilton-Jacobi-Bellman equation—a partial differential equation central to optimal control theory
Drift dynamics: The deterministic or directed component of a system's movement
Diffusion process: The random, spreading component of a system's movement, often modeling exploration or noise