Fine-Tuning Is Not Enough — How to Really Teach LLMs New Knowledge

August 03, 2025 Deep Technical

Research confirms that most factual knowledge in an LLM is acquired during the pre‑training phase, not later during fine-tuning.

Blog Banner

Many people think fine-tuning is all you need to teach an LLM (like LLaMA, GPT, or Mistral) new facts. But if you're trying to update your model with recent events, product changes, new policies etc., fine-tuning alone won’t do the job. Here’s the truth: fine-tuning mostly changes how a model speaks — not what it knows.


What does "teaching knowledge" mean?

Let’s say:

In all of these, the model needs to remember and reason with these facts — not just mimic your prompt. That’s what I mean by teaching/infusing knowledge.


Why Fine-Tuning Doesn’t Work Well (According to Research)


1. Fine-tuned models learn facts slowly

Ovadia et al., EMNLP 2024 - "Fine-Tuning or Retrieval?"

Key takeaway: Fine-tuning is inefficient and brittle for teaching new factual knowledge.


2. Fine-tuning causes forgetting of old knowledge

Li et al., ACL Findings 2024 - "Revisiting Catastrophic Forgetting"

Key takeaway: Fine-tuning often "overwrites" existing knowledge unless very carefully controlled.


3. Fine-tuning increases hallucination risk

Gekhman et al., 2023 - "On Factuality and Memorization"

Key takeaway: Fine-tuning may teach wrong generalizations and lead to confident hallucinations.


4. Real developers agree

Developers across forums like Reddit and GitHub report similar findings:

Example posts:

Key takeaway: Practitioners seems to confirm what the papers show — fine-tuning is fragile for knowledge updates.


So Why Continued Pre‑Training Works for Knowledge Infusion


Research confirms that most factual knowledge in an LLM is acquired during the pre‑training phase, not later during fine-tuning. During pre-training, the model organically builds deep internal representations—a sort of "mental map" linking facts and concepts together. That’s where real knowledge infusion happens.

📌 Research Backing This

Singhal et al. 2022/2023, in “Large Language Models Encode Clinical Knowledge”, evaluated PaLM and Med-PaLM on medical question-answering across six benchmarks (MultiMedQA, HealthSearchQA, MedQA, PubMedQA, MMLU topics). They found that larger models internalize clinical facts primarily through pre-training, while instruction tuning improves usability—not core knowledge retention.

Sources: arxiv.org, sapien.io, openreview.net, semanticscholar.org

Chang et al. 2024 (KAIST, later published at NeurIPS) studied how knowledge is acquired across training steps. They showed that factual memorization occurs as the model sees facts repeatedly during pre-training; performance scales with exposure frequency, dataset deduplication, and larger batch sizes—and that later-step checkpoints didn’t add new memorization power but instead plateaued.

Source: arxiv.org

This explains why continued pre-training (also known as Domain-Adaptive Pretraining or DAPT) is more effective for updating a model’s knowledge base. It essentially mimics the original pre-training process, allowing the model to absorb new facts in a similar way.

Continued pre-training, also called Domain-Adaptive Pretraining (DAPT), is the best way to truly teach an LLM new knowledge.

This method involves:

Why it's better:

Supported by top industry leaders:


The Cons of Continued Pre-Training

While more effective than fine-tuning, continued pre-training is not perfect:

Because of these limitations, continued pre-training is better suited for major knowledge upgrades or domain-specific models — not quick fixes.


When RAG Might Be Better

In many real-world use cases, Retrieval-Augmented Generation (RAG) is the most practical option.

Use RAG if:

RAG is often the best starting point for adding knowledge — and continued pre-training should be reserved for large, stable, and domain-wide updates.


When fine-tuning is useful

But these methods do not scale well to thousands of new facts or major domain shifts.


Summary

If you're serious about building an LLM that truly understands new information, don’t just fine-tune it. Choose the right method for the job.


1 like

Comments (0)

💬 Leave a Comment

No comments yet

Be the first to share your thoughts!

Back to Blogs

Subscribe to AI Tarun

Get notified about latest in AI
Get notified for AI tutorials and Articles
Get notified for My latest AI projects
Subscribe
Buy me a Coffee