Recent years have seen rapid advances in AI driven by foundation models, sometimes known as large language models or general purpose AI. These are transformative AI systems trained on large datasets which power a variety of applications, from content generation to interactive conversational interfaces. Already, there is widespread recognition of this technology’s potential for both social benefit and harm. The use of foundation models could enable new forms of creative expression, boost productivity, and accelerate scientific discovery. It could also increase misinformation, negatively impact workers, and automate criminal activity.
Given the potentially far-reaching impacts of foundation models, shared safety principles must be translated into practical guidance for model providers. This requires collective action. To establish effective, collectively-agreed upon practices for responsible model development and deployment, diverse voices across industry, civil society, academia, and government need to work together.
PAI has released Guidance for Safe Foundation Model Deployment, which will continue evolving in collaboration with our global community of civil society, industry, and academic organizations. This is a framework for model providers to responsibly develop and deploy foundation models across a spectrum of current and emerging capabilities, helping anticipate and address risks. The Model Deployment Guidance gives AI developers practical recommendations for operationalizing AI safety principles. We invite more stakeholders to get involved and help shape this truly collective effort.
This guidance assists foundation model providers: organizations developing AI systems trained on broad datasets to power a wide variety of downstream uses and interactive interfaces.
In cases where models do not fit into one category, choose the model type of higher capability.
Models designed for narrowly defined tasks or purposes with limited general capabilities for which there is a lower potential for harm across contexts.
Do any of the following apply to the model, even if it does not satisfy every criteria?
The key difference from Advanced Narrow and General Purpose models is that these have tightly constrained capabilities in terms of input, domain, output complexity, and potential generalizability.
Models with generative capabilities for synthetic content like text, image, audio, video. Can be narrow purpose focused on specific tasks or modalities or general purpose. Also covers some narrow purpose models focused on scientific, biological or other high consequence domains.
Encompasses general purpose models capable across diverse contexts, like chatbots/LLMs and multimodal models.
Do any of the following apply to the model, even if it does not satisfy every criteria?
Cutting edge general purpose models that significantly advance capabilities across modalities compared to the current state of the art.
Do any of the following apply to the model, even if it does not satisfy every criteria?
Choose the intended initial release method. For phased rollouts, select the current stage and revisit this guidance as release plans progress.
Models released publicly with full access to key components, especially model weights. Can also include access to code, and data. Can be free or commercially licensed. Access can be downloadable or via cloud APIs and other hosted services.
Does the release include at least model weights and potentially other components such as code, training data, and architecture?
Models available only through a controlled API, cloud platform, or hosted through a proprietary interface, with limits on use. Does not provide direct possession of the model. Allows restricting access and monitoring usage to reduce potential harms.
Is the model release only accessible through controlled mediums like proprietary APIs, platforms, or interfaces?
Models developed confidentially within an organization first, with highly limited releases for internal evaluation or restricted external testing, before any potential public availability.
Is model access restricted only to internal personnel and limited external third parties for testing, and not to the public?
Models released in a restricted manner to demonstrate research concepts, techniques, demos, fine-tuned versions of existing models. The release is meant to share knowledge and allow others to build upon it and excludes small-scale individual projects.
Do any of the following apply to the model, even if it does not satisfy every criteria?
This category represents restricted utility of research artifacts and experiments, whereas the other categories encompass full production models intended for real-world deployment.
If the release is a significant update to an existing model, you are encouraged to renew governance processes as needed per the guidance for your model and release type.
Models that continue major development post-deployment by significantly expanding capabilities, necessitating renewed governance.
Do any of the following apply to the model, even if it does not satisfy every criteria?
The applicable guidance for your selection will be displayed below.
The Model Deployment Guidance’s guidelines establish a normative baseline and suggest additional practices for responsible development of foundation models, allowing collaborative reassessment as capabilities and uses advance. This accommodates diverse AI models and deployment scenarios. Not intended as a comprehensive set of instructions for implementation, these guidelines provide a framework for ongoing collective research and action.The guidelines aim to inform and catalyze other individual and collaborative efforts to develop specific guidance or tooling in alignment with the guidelines.
To address risks appropriately, the Model Deployment Guidance’s guidelines are tailored to scale oversight and safety practices based on the capabilities and availability of each AI model. The Model Deployment Guidance avoids oversimplification by not solely equating model size or generality with risk.
The Model Deployment Guidance includes guidelines for open access models, offering a starting point into transparency and risk mitigation strategies. This provides guidance for both current and future providers of open source models.
The Model Deployment Guidance applies across the spectrum of foundation models, from existing to frontier.
The Model Deployment Guidance recommends staged releases and restricted access for frontier models initially until adequate safeguards are demonstrated.
The Model Deployment Guidance establishes starting points to address a wide variety of safety risks, including potential harms related to bias, overreliance on AI systems, worker treatment, and malicious activities by bad actors.
There are a total of 22 possible guidelines included in the Model Deployment Guidance. Not all model and release types are treated the same within the paradigm of the Model Deployment Guidance. The suggested guidelines are more extensive for more capable models and more available release types. The full 22 guidelines apply to the “Frontier and Restricted” model and release category. This concept is visualized below:
As PAI continues evolving the Model Deployment Guidance, we welcome additional perspectives and insights to incorporate into future updated versions.
We’ll bring together a collaborative group focused on applying the Framework in practice through yearly case examples or analysis via a public reporting process. This will help us identify challenges and trade-offs that may arise, and we’ll share our findings.
We’ll provide tactical options to put our key guidelines into operation. We aim to support the implementation of these guidelines over time to ensure they are effective.
We’ll explore how responsibility should be shared across the evolving value chain for foundation models.
We’ll continue to update our model and release categorization, ensuring that it remains current and relevant to the evolving landscape.
The Model Deployment Guidance have been tailored specifically for model providers due to:
Model providers have an opportunity to highlight, share, and further develop emerging internal best practices in a way that is beneficial to the ecosystem as a whole.
Sub-categories of risks:
Click here to download a list of the Model Deployment Guidance’s 22 possible guidelines.
The Model Deployment Guidance distinguishes model providers from actors in the broader AI ecosystem (seen below) as those training foundational models that others may build on. There may be overlap, such as when model providers offer their own applications and services integrating their foundation models.
Ecosystem Actor | Role Description |
Compute / Hardware Providers | Providing underlying compute power to train and run models |
Cloud Providers | Providing underlying cloud infrastructure to support training of and deployed models |
Data Providers | Providing training datasets (intentionally or unintentionally) for model providers, may also be model providers |
Model Providers | Training foundational models (proprietary or open-source) that others may build on as well as interfaces to interact with the models. |
Application Developers (or: Service Developers, Model Integrators) | Building applications and services on top of foundational models |
Consumers and/or Affected Users | Consumers (B2C) who are end-users of services built on top of foundational models |
These guidelines are the result of a collaborative effort led by Madhulika Srikumar, Lead of AI Safety at PAI.
The Model Deployment Guidance reflects insights and contributions from individuals from across the PAI community, including Working Group members:
The current version of the Model Deployment Guidance is the result of a participatory, iterative, multistakeholder process and should not be read as representing views from the individual contributors or organizations.
Joelle Pineau
Vice President, AI Research at Meta and Vice-Chair of the PAI Board
Eric Horvitz
Chief Scientific Officer, Microsoft and Board Chair Emeritus of Partnership on AI
Jerremy Holland
Board Chair, Partnership on AI
Jatin Aythora
Director of Research & Development at BBC and Vice Chair of the PAI Board
Francesca Rossi
AI Ethics Global Leader at IBM and Board Member of Partnership on AI
Lama Nachman
Director of Intelligent Systems Research Lab at Intel Labs and Board Member of Partnership on AI
Esha Bhandari
Deputy Director, ACLU Speech, Privacy, and Technology Project and Member, PAI Safety-Critical AI Steering Committee
Markus Anderljung
Head of Policy at Centre for the Governance of AI (GovAI) and Member, PAI Working Group on Model Guidance
Wafa Ben-Hassine
Principal, Responsible Technology, Omidyar Network and Member, PAI Safety-Critical AI Steering Committee
Reena Jana
Head of Content and Partnership Enablement, Responsible Innovation, Google and Member, PAI Working Group on Model Guidance
We want to thank everyone who provided feedback on PAI’s Guidance for Safe Foundation Model Deployment.
In response to the focus areas identified through public comments, we’re pleased to share our new resource: “Risk Mitigation Strategies for the Open Foundation Model Value Chain.” This document provides more detailed guidance for open access models and explores how responsibility should be shared across the evolving value chain for foundation models.
Stay informed about our work in ensuring the safe and responsible deployment of foundation models.