Search⌘ K

What are Foundation Models

Understand what foundation models are, what they can do, how they are made, and their implications.

Foundation models represent a major shift in AI. Instead of being trained for a single task, they learn broad patterns from large datasets, allowing them to adapt to many different applications with minimal additional training.

We’ve seen GPT generate fluent, coherent text, but GPT itself is more than just a model. It is an example of a foundation model, a powerful system trained on vast amounts of data that can adapt to many tasks with minimal extra training. But what exactly are foundation models, and why are they so transformative?

Foundation models

Traditionally, AI models were designed from scratch for narrow, specialized tasks like spam detection, language translation, or image classification. They worked well but lacked flexibility—you had to build a new model for every new job.

The term “foundation model” emerged precisely because AI now extends beyond language alone, spanning various domains such as vision, audio, and multimodal applications. The diversity of these models underscores their expansive capabilities and potential.

Foundation models flip this idea. Instead of being trained for a single purpose, they are trained once on massive, diverse datasets: text, images, audio, even code. These models learn general knowledge and patterns that can be adapted to countless downstream tasks.

Think of GPT again: it wasn’t designed just for one task, but as a general-purpose system capable of summarizing, translating, ...