Reward Shaping: The process of modifying the reward function to provide more frequent feedback to the agent, accelerating learning without altering the optimal policy
Reward Engineering: The broader task of defining the reward function from scratch, often involving domain knowledge, constraints, or heuristics
PBRS: Potential-Based Reward Shaping—a technique adding a shaping term derived from the difference in potential functions φ(s') - φ(s), guaranteeing policy invariance
IRD: Inverse Reward Design—a method that infers the true objective by treating the specified proxy reward as an observation rather than the ground truth
Reward Hacking: A failure mode where the agent exploits loopholes in the reward function to maximize points without achieving the intended task
Sparse Rewards: Environments where non-zero rewards are rare (e.g., only upon winning), making it difficult for agents to learn which actions are beneficial
Sim-to-Real: The challenge of transferring policies trained in simulation to the real world, often requiring robust reward engineering to handle domain discrepancies
Intrinsic Motivation: Rewards generated internally by the agent (e.g., for curiosity or novelty) rather than provided by the external environment
PGRD: Policy Gradient for Reward Design—an algorithm that optimizes reward parameters via gradient ascent to maximize the designer's objective