Site icon Gradient Flow

Foundation Models: A Primer for Investors and Builders

A non-technical guide and market map.

By Kenn So and Ben Lorica.

What are foundation models?

Foundation models (FM) are a class of machine learning models that are trained on diverse data and can be adapted or fine-tuned for a wide range of downstream tasks. The term “foundation” is controversial among some researchers, but setting aside disagreements over terminology,  these models already have had a significant impact. They are already used in large-scale applications in various areas including search, natural language processing, and software development.

Figure 1: Large Language Models, Transformers, and Foundation Models.

Scale is a key element of these models. Scale is enabled by improvements in hardware, the emergence of powerful models that can be parallelized (transformers), and the availability of large amounts of data for training (typically by self-supervised learning). In Figure 2 below, we show some of the open data sources used in training a large language model. The fact that such models are trained on a broad knowledge base makes them adaptable to a variety of downstream tasks.

Figure 2: Distribution of data sources for the 800GB text dataset “The Pile”. The size of each box is the size relative to other data sources. Source: EleutherAI.

To illustrate their broad utility, we present a list of common NLP applications that foundation models are being used for today. Many of these applications (Q&A systems, chatbots, etc.) are well-known, but others, like code generation, are more innovative.

Figure 3: Foundation models are being used in a variety of NLP applications and downstream tasks.

Why should investors and builders care

  1. Foundation models are already powering the products we’re using daily: BERT is used in Google’s search engine. As we show in Figure 4 below, foundation models are not novel technologies looking for problems to solve. 
  2. Foundation models accelerate product development: In areas where foundation models are available, the focus shifts from training models from scratch, to acquiring data to fine-tune models for specific applications and tasks. FMs are also helping software engineers write code (e.g. Github Copilot). 
  3. Underlying technologies have made significant advances and models will only continue to improve: Google’s PALM was shown to be better than the average human across a wide range of language benchmarks from Q&A to code explanation to logical deductions. And researchers are just getting started: experiments with more efficient algorithms, new distributed computing tools,  and multimodal models continue to yield steady improvements. 
  4. Foundation models could lead to more pegacorns: Deep learning breakthroughs led to startups with significant revenues (“pegacorns”). Most of the AI Pegacorns we uncovered are vertical or domain specific applications, not general purpose platforms We believe that foundation models will spur a similar wave of successful startups. 
Figure 4: Some recent applications powered by foundation models.

Startup Ecosystem

Given the potential of FMs, investors have begun funding startups in this space, most notably in companies building copywriting software. In Figure 5 below, we list over 60 companies that have raised over $3.8 billion to date and have been publicly described as using or building foundation models, which include large language models like GPT-3,  Jurassic, etc.

Figure 5: Representative sample of companies that have publicly stated that they are using, building, or enabling foundation models.

Implications for product builders, entrepreneurs, and investors

Here’s the slide deck version of the guide where you can also find the spreadsheet of 60+ companies in the market map.


Kenn So is an investor at Shasta Ventures, an early-stage VC, and was previously a data scientist. He also writes Quild, a newsletter about startups. 

Ben Lorica is a principal at Gradient Flow. He helps organize the Data+AI Summit, Ray Summit, and is co-chair of the NLP Summit and K1st World. He is an advisor to several startups.


To stay up to date, subscribe to the Gradient Flow newsletter.


Related content


[Image: Foundations by Ben Lorica; original photos from Unsplash, via Infogram.]

Exit mobile version