What is Foundation model
A foundation model is a large, general-purpose model trained on broad datasets so it can be adapted to many downstream tasks. Instead of training a separate model from scratch for every problem, teams start with a foundation model and then specialize it using fine-tuning, prompting, or task-specific heads.
Foundation models are most commonly discussed for:
- Language: large language models used for writing, reasoning, and retrieval
- Vision: models trained on broad image-text pairs for classification, detection, and search
- Multimodal: models that handle combinations of text, images, audio, and video
Why foundation models changed ML workflows
- They reduce the need for massive labeled datasets for every new task
- They enable faster iteration: prompting, few-shot learning, and lightweight tuning
- They shift effort toward data quality, evaluation, and alignment (making outputs reliable)
Common ways teams adapt foundation models
- Prompting: give instructions and examples in the input
- Fine-tuning: update weights using task data (often supervised)
- Retrieval-augmented generation (RAG): ground responses using external documents
- Evaluation + guardrails: measure and control outputs before deployment
Foundation models are powerful, but they still depend on clean requirements, strong evaluation sets, and domain-specific data to behave reliably in production.