PLM: Protein Language Model—a transformer-based model trained on protein sequences to predict properties or generate new proteins
NLM: Natural Language Model—a transformer-based model trained on human text
early-exit: An inference strategy where a model can output a prediction from an intermediate layer if confidence is high, rather than processing all layers
attention heads: Components in transformer models that learn relationships between different parts of the input sequence
positional information: Information derived from the relative or absolute location of tokens (amino acids or words) in a sequence
semantic information: Information derived from the identity and context of tokens (meaning of words or physicochemical properties of amino acids)
MLP: Multi-Layer Perceptron—a simple feed-forward neural network used here as a classification head attached to PLM layers
ESM2: Evolutionary Scale Modeling 2—a state-of-the-art protein language model
ProtBERT: A BERT-based protein language model trained on the UniRef100 dataset
ProtAlBERT: An AlBERT-based protein language model, designed to be more parameter-efficient