Transfer Learning
A machine learning technique where a model trained on one task or dataset is adapted to perform a different but related task, leveraging the knowledge acquired during initial training. Transfer learning is the foundational principle behind fine-tuning and the use of pre-trained foundation models across diverse applications.
Definition
Transfer learning is the practice of applying knowledge learned from one task to improve performance on a different task. In deep learning, this typically means using a model pre-trained on a large, general-purpose dataset as the starting point for training on a smaller, specialised dataset. The pre-trained model’s learned representations — its understanding of language structure, visual features, or other patterns — transfer to the new task, dramatically reducing the data and computation required. Transfer learning is the principle that makes modern AI practical: rather than training every model from scratch, organisations start with foundation models (GPT, Claude, Llama, Gemini) and adapt them through fine-tuning, prompt engineering, or RAG to their specific needs.
How It Relates to AI Threats
Transfer learning has a direct security implication within the Security and Cyber Threats domain: adversarial transferability. Adversarial examples crafted against one model often succeed against different models trained on similar data, because transfer learning causes models to share similar internal representations and decision boundaries. This means an attacker can craft adversarial perturbations against a publicly available model and use them to attack proprietary models that share architectural lineage or training data. Transfer learning also means that vulnerabilities in widely used foundation models propagate to every downstream application built on them.
Why It Occurs
- Training AI models from scratch requires enormous datasets and computational resources that most organisations cannot afford
- Pre-trained models contain general-purpose knowledge representations that are useful across many tasks
- The transformer architecture’s learned representations transfer effectively between language, vision, and multimodal tasks
- Foundation model providers offer APIs and model weights specifically designed for transfer learning workflows
- The economic incentive structure of AI development (pre-train once, deploy many times) makes transfer learning the dominant paradigm
Real-World Context
Transfer learning is used in virtually all modern AI deployments. Security researchers have demonstrated adversarial transferability across model families: attacks crafted against open-weight models (Llama, Mistral) can transfer to closed models (GPT, Claude) with varying success rates. This has implications for AI security assessments, which must account for attacks developed against proxy models. On the beneficial side, transfer learning enables organisations to build specialised AI applications without the billions of dollars required for pre-training, democratising access to AI capabilities.
Related Threat Patterns
Related Terms
Last updated: 2026-04-03