IEEE 754: The technical standard for floating-point arithmetic, representing numbers using three parts: a sign bit, an exponent, and a significand (fraction)
Significand: The part of a floating-point number that contains the significant digits (also called the mantissa)
xVal: A method that represents numbers by scaling a single learnable token embedding by the number's value
FoNE: Fourier Number Embedding—a method encoding numbers using sinusoidal functions (sines and cosines) of different frequencies
sMAPE: Symmetric Mean Absolute Percentage Error—a metric for measuring prediction accuracy relative to the magnitude of the target
nanoGPT: A small-scale implementation of the GPT architecture used for efficient experimentation
Reasoning Chain: The process where an LLM generates intermediate text steps (like 'Chain of Thought') to solve a problem, rather than outputting the answer immediately
RMSNorm: Root Mean Square Normalization—a technique used in neural networks to stabilize training by normalizing the input vector
BCE Loss: Binary Cross Entropy Loss—a loss function used here to train the model to predict each bit of the floating-point representation correctly