Docs

What Is a Foundation Model?

A foundation model is a large-scale artificial intelligence system trained on massive amounts of data to perform a wide range of tasks. Think of it as the "brain" that powers many of the AI tools you use — from chatbots to coding assistants to image generators. When you use ChatGPT, Claude, or Gemini, you are interacting with a foundation model (or a variant of one).

Understanding foundation models matters because the model powering a tool largely determines what it can and cannot do. If you know which model sits under the hood, you can make smarter choices about which tools belong in your stack.

Definition and Examples

The term "foundation model" was popularized by Stanford's Institute for Human-Centered AI in 2021. It describes models that are:

Major foundation models as of 2026 include:

How Foundation Models Are Trained

Foundation models are built through pre-training: the model learns patterns from huge datasets (often billions of tokens of text or millions of images). This phase is expensive and done by large labs. After pre-training, models may be further refined with:

You do not train a foundation model yourself. You use one (via API or a product built on top of it) and adapt it with prompts, fine-tuning, or RAG.

Key Concepts That Affect Tool Selection

Parameters

Parameters are the learned weights inside the model. More parameters generally mean more capacity, but also higher cost and latency. A "70B" model has roughly 70 billion parameters. When comparing tools, check which model size they use — it affects quality and price.

Context Window

The context window is how much input the model can consider at once (measured in tokens; roughly 4 characters per token for English). A 200K context window can hold a long document; a 1M window can hold entire codebases or large reports. Tools built on models with larger context windows can handle longer conversations and bigger files.

Training Data Cutoff

Models have a knowledge cutoff — the date after which their training data ends. A model trained on data through mid-2025 will not know about events or products from 2026. For time-sensitive tasks, check the cutoff and consider tools that use retrieval (RAG) to pull in fresh information.

Why This Matters for Tool Selection

When you browse the >Model Directory, you will see tools built on different foundation models. A writing assistant powered by Claude may excel at nuanced prose; one powered by GPT may be stronger at structured output. A coding tool might use a model optimized for code. The underlying model is a major factor in:

Use the directory's filters and tool profiles to see which foundation model powers each tool. That information helps you compare apples to apples and avoid paying for more (or less) than you need.

The Bottom Line

Foundation models are the core AI systems that power most modern AI tools. They are pre-trained at scale by large labs, and products layer interfaces, workflows, and integrations on top of them. Understanding parameters, context windows, and training cutoffs gives you a clearer picture of what each tool can do — and whether it fits your stack.

Related Reading