← Back to Paper List

Bring Your Own Knowledge: A Survey of Methods for LLM Knowledge Expansion

Mingyang Wang, Alisa Stoll, Lukas Lange, Heike Adel, Hinrich Schutze, Jannik Strotgen
Bosch Center for Artificial Intelligence (BCAI), Ludwig Maximilian University of Munich, Munich Center for Machine Learning, Karlsruhe University of Applied Sciences, Hochschule der Medien, Stuttgart
Proceedings of the First Workshop on Large Language Model Memorization (L2M2) (2025)
Pretraining Factuality RAG P13N RL

📝 Paper Summary

Knowledge expansion Continual learning Model editing Retrieval-augmented generation
This survey provides a taxonomy and overview of methods for expanding LLM knowledge—categorizing techniques into continual learning, model editing, and retrieval—across factual, domain, language, and preference dimensions.
Core Problem
LLMs are typically trained once with a cutoff date, making their internal knowledge static and unable to adapt to evolving facts, specialized domains, new languages, or changing user preferences without intervention.
Why it matters:
  • Static models become obsolete as real-world information changes (factual decay), limiting their utility in time-sensitive applications
  • General-purpose models often fail in specialized fields like medicine or law without targeted domain adaptation
  • Re-training fully is computationally prohibitive, creating a need for efficient adaptation strategies that mitigate catastrophic forgetting
Concrete Example: An LLM trained in 2021 will not know about the 2023 Nobel Prize winners. Without knowledge expansion, it hallucinates or refuses to answer. A retrieval-based method would fetch the news, while model editing would directly modify the weights associated with 'Nobel Prize 2023'.
Key Novelty
Task-Oriented Knowledge Expansion Taxonomy
  • Classifies expansion methods not just by technique (continual learning vs. editing vs. retrieval) but by the *type* of knowledge being integrated (factual, domain, language, preference)
  • Contrasts 'implicit' knowledge expansion (modifying internal parameters via continual learning or editing) with 'explicit' expansion (retrieval-based access during inference) to guide selection based on use-case needs
Evaluation Highlights
  • Review of Continual Pretraining (CPT) effectiveness for language expansion, citing Glot500 extending support to 500 languages
  • High-level comparison of methods showing retrieval is best for 'Plug-and-Play' flexibility, while Continual Learning excels at 'Generalization' (Table 1 summary)
  • Survey of programming language expansion showing domain-specific models like CodeLLaMA and StarCoder 2 consistently outperform general-purpose LLMs on code benchmarks
Breakthrough Assessment
7/10
A comprehensive survey that structures a fragmented field. While it doesn't propose a new model, its clear taxonomy (Facts/Domain/Language/Preference × CL/Editing/Retrieval) is a valuable contribution for researchers navigating adaptation choices.
×