Skip to main content
Early access β€” new tools and guides added regularly
Core AI

Transfer Learning

Last reviewed: April 2026

A technique where a model trained on one task is reused as the starting point for a different but related task, dramatically reducing the data and time needed.

Transfer learning is the technique of taking a model that has already been trained on one task and adapting it for a different task. Instead of training a new model from scratch β€” which requires massive data and compute β€” you start with a pre-trained model and fine-tune it for your specific needs.

The intuition

Consider how humans learn. A person who speaks French finds it easier to learn Spanish than someone starting from zero. Their knowledge of grammar, vocabulary patterns, and Romance language structure transfers. Transfer learning works the same way β€” a model that has learned general patterns about language, images, or data can apply that foundational knowledge to new, specific tasks.

How transfer learning works

  1. Pre-training: A large model is trained on a broad dataset to learn general features. For language: grammar, facts, reasoning patterns. For images: edges, textures, shapes, objects.
  2. Transfer: The pre-trained model's learned representations are preserved.
  3. Fine-tuning: The model is further trained on a smaller, task-specific dataset. Only the final layers (or all layers with a small learning rate) are updated.

Why it matters

Transfer learning solves the fundamental constraint of AI development: most organisations do not have millions of labelled examples. With transfer learning:

  • A model trained on millions of general images can be adapted to identify product defects with just a few hundred labelled examples
  • A large language model trained on the internet can be fine-tuned to understand your company's specific terminology with a few thousand examples
  • A speech recognition model can be adapted to your industry's vocabulary with modest training data

The foundation model paradigm

Modern AI is built on transfer learning. Models like GPT-4, Claude, and Gemini are general-purpose foundation models designed to be adapted. The pre-training captures broad capabilities. Fine-tuning, prompting, and retrieval-augmented generation are all forms of transfer β€” applying general knowledge to specific tasks.

Practical applications

  • Custom classifiers: Fine-tune a pre-trained model to classify your specific document types
  • Domain adaptation: Adapt a general model to understand medical, legal, or financial language
  • Low-resource languages: Transfer knowledge from well-resourced languages to less common ones
  • Small businesses: Access state-of-the-art AI without the budget for training from scratch

Limitations

Transfer learning works best when the source and target tasks are related. A model trained on English text will transfer well to sentiment analysis in English, but may transfer poorly to medical image analysis. The greater the gap between the original and target domains, the less useful the transfer.

Want to go deeper?
This topic is covered in our Practitioner level. Access all 60+ lessons free.

Why This Matters

Transfer learning is what makes AI accessible to organisations without massive datasets or computing budgets. Understanding it helps you evaluate how AI vendors customise models for your industry and assess the feasibility of building custom AI solutions with your available data.

Related Terms

Learn More

Continue learning in Practitioner

This topic is covered in our lesson: Customising AI for Your Organisation