Skip to main content
Early access β€” new tools and guides added regularly
Core AI

Zero-Shot Learning

Last reviewed: April 2026

An AI model's ability to perform a task it was not explicitly trained on, using general knowledge to handle new categories or instructions without any examples.

Zero-shot learning is an AI model's ability to perform a task correctly without having been trained on specific examples of that task. The model uses its general knowledge to handle entirely new situations β€” much like a human who can identify a platypus after reading a description, without ever seeing one before.

How it works with LLMs

When you ask Claude to "classify this customer email as a complaint, enquiry, or compliment," the model can do this accurately despite never being explicitly trained on your specific classification task. It draws on its general understanding of language, context, and meaning.

This contrasts with traditional machine learning, where you would need to:

  1. Collect hundreds of labelled emails
  2. Train a custom classification model
  3. Test and refine the model
  4. Deploy it

With zero-shot learning, you simply describe the task and the model performs it immediately.

Zero-shot vs few-shot vs fine-tuning

These represent a spectrum of task adaptation:

  • Zero-shot: No examples provided. Rely entirely on the model's general knowledge. Fast to deploy but may have lower accuracy on specialised tasks.
  • Few-shot: A handful of examples included in the prompt. Slightly better accuracy with minimal effort.
  • Fine-tuning: Hundreds or thousands of examples used to further train the model. Highest accuracy but requires more data, time, and expertise.

Why zero-shot learning is revolutionary

Before large language models, every new AI task required a new model with new training data. Want to detect spam? Train a spam classifier. Want to extract entities? Train an NER model. Want to summarise text? Train a summarisation model.

Zero-shot learning collapses all of this into a single model that can attempt any text task based on instructions alone. This dramatically reduces the barrier to deploying AI for new use cases.

Limitations

Zero-shot performance depends on how well the task aligns with the model's training:

  • Common tasks: Sentiment analysis, translation, and summarisation work well zero-shot because the model has seen similar tasks during training.
  • Specialised tasks: Classifying legal documents by jurisdiction or identifying specific medical conditions may require examples or fine-tuning.
  • Novel categories: The model can only generalise from what it knows. Completely novel concepts may confuse it.

Zero-shot in computer vision

Models like CLIP (OpenAI) can classify images into categories they have never seen during training by matching images to text descriptions. This enables flexible image classification without training a new model for each set of categories.

Want to go deeper?
This topic is covered in our Essentials level. Access all 60+ lessons free.

Why This Matters

Zero-shot learning is what makes modern AI immediately useful out of the box. It means you can prototype AI solutions in minutes instead of months. Understanding zero-shot capabilities helps you quickly identify which business problems AI can solve immediately and which require additional investment in examples or fine-tuning.

Related Terms

Learn More

Continue learning in Essentials

This topic is covered in our lesson: Writing Effective Prompts