Now we discuss the technology behind building better model architectures,
training and adaptation procedures, and of course scaling up the systems.
One crucial but often overlooked topic is data—where does it come from and
what is its composition?
In addition, we want foundation models to be robust to distribution shifts
and secure against attackers.
Finally, we wish to understand why foundation models work from both a mathematical perspective
as well as an empirical perspective.
Modeling
To learn more, see
this section in the report.
Authors: Drew A. Hudson, Antoine Bosselut, Alex Tamkin, Omar Khattab, Jared Quincy Davis, Jiaxuan You, Trevor Gale
What structural properties give rise to a foundation model? In the modeling section, we explore the underlying architectures behind foundation models and identify 5 key attributes. First, we start by discussing
expressivity of the computational model — to capture and assimilate real-world information, and
scalability — to adeptly handle large quantities of high-dimensional data. These properties are successfully realized by existing architectures such as the transformer network that underpins most foundation models to date. We then proceed to attributes may be essential for the next generation of models, including:
multimodallity — to consume, process and potentially produce content from different sources and domains,
memory capacity — to effectively store and retrieve the acquired knowledge, and finally,
compositionality, to foster successful generalization to novel settings and environments. We believe that realizing the full potential envisioned for foundation models will hinge on modelling advances to fulfill these desiderata.
Training
Training objectives mathematically specify how models should learn and acquire capabilities from their training data.
The current status quo for training foundation models involves modality-specific objectives (
e.g., masked language modeling for text and SimCLR for images) that are often chosen heuristically.
We envision that future training objectives for foundation models will reflect two changes:
principled selection derived from systematic evidence and evaluation, and
domain-generality to provide rich, scalable, and unified training signal across data sources and modalities. We also discuss important design trade-offs, including generative vs discriminative training, the choice of input data representation, and the potential of future training objectives that involve explicit representations of goals.
Adaptation
To learn more, see
this section in the report.
Authors: Xiang Lisa Li*, Eric Mitchell*, Sang Michael Xie, Xuechen Li, Tatsunori Hashimoto
Foundation models are intermediary assets; they are unfinished and generally should not be used directly, instead requiring adaptation for specific downstream tasks.
The
de facto approach for adaptation has been fine-tuning, with recent work suggesting that lightweight fine-tuning alternatives and prompting-based methods may achieve favorable accuracy-efficiency tradeoffs.
Moving forward, we envision a more expansive view of adaptation that goes beyond just specializing foundation models to perform the task of interest: adaptation will alleviate deficiencies of stand-alone foundation models (
e.g., temporal adaptation to reflect changes over time in the world) or introduce
constraints (
e.g., GDPR compliance relating to the
right to be forgotten); this broader perspective on adaptation coincides with a need for new evaluation protocols that systematically evaluate adaptation methods while controlling for resources (
e.g., runtime, memory) and access requirements involved in adaptation.
Evaluation
Evaluation offers context to foundation models by providing a means to track progress, understand models, and document their capabilities and biases.
Foundation models challenge the ability of standard evaluation paradigms in machine learning to achieve these goals since they are one step removed from specific tasks.
To envision new paradigms in evaluation that suit foundation models, we discuss (i) evaluating foundation models
directly to measure their
inherent capabilities and inform how foundation models are trained, (ii) evaluating task-specific models by
controlling for adaptation resources and access, and (iii) broader
evaluation design to provide richer context beyond measures of accuracy (
e.g., robustness, fairness, efficiency, environmental impact).
Reform of evaluation practices will allow for evaluation that adequately serves both the diverse goals and stakeholders involved in the foundation model paradigm.
Systems
To learn more, see
this section in the report.
Authors: Deepak Narayanan, Trevor Gale, Keshav Santhanam, Omar Khattab, Tianyi Zhang, Matei Zaharia
While the training data determines the theoretical information available for foundation models, and model architectures and training objectives determine how much of this information can be extracted, computer systems determine what is practically achievable.
Systems are a key bottleneck for scaling in terms of data and model size, both of which appear to reliably track with improvements in capabilities.
To ensure that we can train the next generation of foundation models efficiently (with respect to time and cost), we will require the co-design of algorithms, models, software, and hardware.
This co-design is already starting to happen to in various forms, from carefully tuned parallelism strategies to new architectures such as retrieval-based and mixture-of-expert models.
Beyond training, we consider what will be required to deploy applications on top of foundation models (
e.g., efficient inference).
Data
To learn more, see
this section in the report.
Authors: Laurel Orr, Simran Arora, Karan Goel, Avanika Narayan, Michael Zhang, Christopher Ré
Data is the lifeblood of foundation models; the training data of these models largely determines what these capabilities these models can acquire.
The centrality of data is not unique to foundation models; recent calls for
data-centric AI indicate the pervasive importance of managing, understanding, and documenting data used to train machine learning models.
For foundation models specifically, the current
modus operandi is for training data to be selected using unspecified or unclear principles with a general lack of transparency regarding the nature of training data.
We believe an alternative approach is needed to re-imagine the data ecosystem surrounding foundation models: we draw upon work on data visualization and management to propose a
data hub for foundation models.
We articulate how this proposal relates to many of the relevant data-centric considerations for foundation models: selection, curation, documentation, access, visualization and inspection, quality assessment, and legal regulation.
Security and privacy
Security and privacy for foundation models is largely uncharted at present.
Fundamentally, foundation models are a high-leverage
single point of failure, making them a prime target for attack: existing work demonstrates a variety of security vulnerabilities (
e.g., adversarial triggers to generate undesirable outputs) or privacy risks (
e.g., memorization of training data) for these models.
Further, the generality of foundation models compounds these concerns, intensifying the risk for
function creep or dual use (
i.e., use for unintended purposes).
For security, we view foundation models as akin to
operating systems in traditional software systems; we discuss steps towards secure foundation models which, if achieved, would provide a strong abstraction layer to build upon for reliable ML applications.
For privacy, by leveraging knowledge transfer from public data, foundation models may enable more sample efficient adaptation to sensitive data distributions,
i.e., privacy-preserving applications may incur less degradation in accuracy when built using foundation models.
Robustness to distribution shifts
To learn more, see
this section in the report.
Authors: Sang Michael Xie, Ananya Kumar, Rohan Taori, Tony Lee, Shiori Sagawa, Pang Wei Koh, Tatsunori Hashimoto
A major limitation of standard machine learning is that it produces models that are not robust to
distribution shifts, where the training distribution does not match the test distribution (for the downstream task).
Existing work shows that adapting a foundation model trained on a broad range of unlabeled data improves the robustness of adapted models across a wide variety of shifts.
This opens a new set of promising directions for improving training and adaptation of foundation models for robustness.
However, we do not believe that foundation models are a panacea for robustness—challenges such as extrapolation across time and spurious correlations are not likely to be fully addressed.
AI safety and alignment
Ensuring foundation models are reliable, robust, and interpretable is increasingly important when considering the potential real-world applications of these models.
In addition to critical and immediate considerations, we also consider the relationship between foundation models and larger-scale risks, hazards, and harms that have the potential for increased relevance as model capabilities continue to advance.
For example, we consider the importance of
aligning foundation models such that they are not deployed with
misspecified goals or values. We also discuss the relevance of
forecasting the emergent behaviors of foundation models (
e.g., the ability to deceive or plan strategically), which may complicate attempts to adapt them to particular tasks, and may require new approaches for interpretability or evaluation.
Theory
To learn more, see
this section in the report.
Authors: Aditi Raghunathan, Sang Michael Xie, Ananya Kumar, Niladri Chatterji, Rohan Taori, Tatsunori Hashimoto, Tengyu Ma
Learning theory provides a broad foundation for the variety of contexts encountered in applied machine learning; theory offers both understanding, principles, and guarantees to complement empirical findings.
At present, the study of foundation models is largely empirical: the theory of standard supervised learning, while relatively mature, is inadequate to fully explain foundation models.
Specifically, the discrepancy between the training phase and the adaptation phase within the foundation model regime pinpoints the insufficiency of existing theory, since these phases correspond to (potentially) completely different tasks and data distributions.
Nevertheless, we endeavor that advances in theory to address this discrepancy, even in simple, limited settings, will provide useful insights.
Interpretability
To learn more, see
this section in the report.
Authors: John Hewitt*, Armin W. Thomas*, Pratyusha Kalluri, Rodrigo Castellon, Christopher D. Manning
Interpretability provides clarity to foundation models: the opacity of the deep neural networks that underpin foundation models, alongside the expected ubiquity of foundation models, heightens the need to understand these models and their capabilities.
Interpretability methods at present generally are designed for interpreting and explaining the behavior of task-specific models; the nature of foundation models (
i.e., the wide array of tasks these models are beneficial for and the unexpected emergent properties they acquire) introduces new challenges for interpretability research.
To frame the discussion of interpretability for foundation models, we propose the
one model-many models paradigm, which aims to determine the extent to which the
one model (the foundation model) and its
many models (its adapted derivatives) share decision-making building blocks.
In addition to interpreting the decision-making components involved, we further discuss
explainability in the context of foundation models (
e.g., the validity of
post hoc explanations generated by models) as well as the
mechanisms that drive model behavior (which may clarify the extent to which understanding foundation models can extend to understanding their adapted derivatives).
Given the critical role we ascribe interpretability in the study of foundation models, we conclude with an assessment of the societal impact of interpretability and non-interpretability.