← Back to Paper List

Targeted Lexical Injection: Unlocking Latent Cross-Lingual Alignment in Lugha-Llama via Early-Layer LoRA Fine-Tuning

Stanley Ngugi
arXiv (2025)
Pretraining Benchmark

📝 Paper Summary

Cross-lingual alignment Low-resource languages (LRLs) Parameter-Efficient Fine-Tuning (PEFT)
The paper improves Swahili-English translation capabilities in a Llama-based model by identifying that early layers already understand these translations perfectly, then fine-tuning the model to preserve this knowledge until the final output.
Core Problem
Multilingual LLMs often show poor output-level alignment for low-resource languages like Swahili, despite possessing strong latent knowledge of these translations in their internal layers.
Why it matters:
  • Low-resource languages like Swahili are underserved by current LLMs, hindering equitable access to technology.
  • Suboptimal lexical alignment degrades performance in critical downstream tasks like translation and cross-lingual information retrieval.
  • Current methods often assume the model lacks knowledge, rather than realizing the knowledge exists but is lost during information propagation through deep networks.
Concrete Example: The model Lugha-Llama inherently knows that Swahili 'mkate' means 'bread' (near-perfect similarity in Layer 2), but by the time the information reaches the final output layer (Layer 31), the representation has degraded to a low similarity (~0.32), failing to reflect this knowledge.
Key Novelty
Targeted Lexical Injection (TLI)
  • Empirically identifies a specific early layer (Layer 2) where cross-lingual alignment is naturally maximal before it degrades deeper in the network.
  • Uses LoRA (Low-Rank Adaptation) fine-tuning with a contrastive objective specifically on embeddings from this optimal early layer to reinforce and propagate this existing knowledge to the output.
Evaluation Highlights
  • +28.08% improvement in average cosine similarity for trained Swahili-English word pairs (from 0.3211 to 0.4113).
  • +28.32% improvement generalizes to unseen control word pairs (from 0.3143 to 0.4033), showing the method improves the mechanism rather than just memorizing pairs.
  • Statistical significance confirmed with extremely low p-values (p < 10^-240) for both trained and control sets.
Breakthrough Assessment
7/10
Novel insight about layer-wise alignment degradation and a targeted, parameter-efficient fix. Strong empirical results for Swahili, though limited to lexical alignment on one language pair so far.
×