World Model: An internal neural network that predicts the next state of the environment given the current state and action
Prediction Error (PE): The difference between the world model's predicted next state and the actual observed next state; used here as the anomaly signal
Drift: A gradual corruption added to observations over time (e.g., slowly increasing velocity bias)
Sinusoidal Blindness: The phenomenon where world models learn to predict periodic noise as part of normal dynamics, making it invisible to anomaly detectors
Collapse Before Awareness (CBA): A failure mode where the agent's policy fails (e.g., robot falls) due to drift before the anomaly detector accumulates enough evidence to flag the issue
Doubt Index (DI): A detector family that tracks the z-score of prediction error using an exponential moving average
PPO: Proximal Policy Optimization—a standard reinforcement learning algorithm used here to train the agent's policy
MuJoCo: A physics engine used for simulating robot control environments (HalfCheetah, Hopper, Walker2d, Ant)
MSE: Mean Squared Error—used to measure the magnitude of prediction error
ROC curve: Receiver Operating Characteristic curve—a plot illustrating the diagnostic ability of a binary classifier system as its discrimination threshold is varied
Wilson score interval: A confidence interval for a binomial proportion (e.g., detection rate), used for statistical reporting
EMA: Exponential Moving Average—a type of temporal smoothing used in the Doubt Index detector