PLM: Pre-trained Language Model—models like BERT trained on vast text to learn language representations
ECP: Entity Concept Prediction—the novel pre-training objective where the model predicts the taxonomic class of an entity mention
long-tail entities: Entities that appear very rarely in the training corpus, making them hard for models to learn purely from context statistics
isA relation: A semantic relationship indicating that an entity is an instance of a concept (e.g., Socrates isA Philosopher)
WikiTaxo: The specific taxonomy constructed in this paper from Wikidata, containing entities and their popular/basic-level concepts
KEPLMs: Knowledge-Enhanced Pre-trained Language Models—PLMs that integrate external structured knowledge (like KGs) into their training
Tagme: A tool used to link entity mentions in text to Wikipedia pages
MLM: Masked Language Modeling—the standard pre-training task where models fill in hidden tokens in a sentence
GELU: Gaussian Error Linear Unit—an activation function used in BERT and modern neural networks