← Back to all companies
M

Meta

Develops the open-weight Llama model family; largest open-source contributor to frontier AI.

Safety Documents

Responsible Scaling Policy
Meta Responsible AI Framework / Meta AI Safety Policy. Meta published an overview of AI safety policies for the UK AI Safety Summit (2023). No single 'Responsible Scaling Policy' equivalent but has a Responsible AI framework.🔗
Last updated: 2023-11
Model Card / System Card
Llama 3.3 Model Card (most recent). Meta publishes model cards on Hugging Face and AI blog for all Llama releases.🔗
Last updated: 2024-12
Safety Benchmark Results
Safety benchmarks included in model cards. Meta also published CyberSecEval evaluations. CAIP noted Meta conducted limited safety testing of Llama 3.1 in July 2024.🔗
Acceptable Use Policy
Llama Acceptable Use Policy prohibits weapons development, military warfare, espionage, CSAM. Made exception for US national security agencies in November 2024.🔗

Testing & Evaluation

Third-Party Red Teaming
Internal red teams; content specialists; external academic red teaming for Llama models. Llama 3 model cards document internal and external red teaming. Child Safety risk assessments conducted with expert team. Partners with content specialists for specific risk domains.🔗
⚠️
CBRN Risk Evaluation
Meta CyberSecEval evaluates cyber risks. CBRN-specific evaluation less prominent in public documents compared to OpenAI/Anthropic/DeepMind. Some safety testing via Llama safety tools (Llama Guard, Purple Llama).🔗
Pre-Deployment Safety Evaluation
Safety evaluations documented in model cards. CAIP noted in July 2024 that Meta's safety testing for Llama 3.1 was 'limited' and conducted closed-source on an open-source model.🔗
⚠️
External Safety Evaluations
UK AI Safety Institute (for some models), Academic red team partners. Limited evidence of systematic METR-style external evaluations. CAIP criticized limited external evaluation scope for Llama 3.1.🔗

Governance

Independent Safety Board
Meta Oversight Board (content moderation); Safety Advisory Council; Youth Advisory Council. Meta Oversight Board governs content moderation across platforms. Safety Advisory Council and Youth Advisory Council provide expert input on child safety and privacy. FLI AI Safety Index (2024) noted Meta's governance lacks alignment with AI safety priorities.🔗
Seoul AI Safety Commitment
Last updated: 2024-05-21
Government Safety Report
UK AI Safety Summit (2023); US Senate testimony. Meta published AI safety policy overview for UK AI Safety Summit 2023. Mark Zuckerberg and executives have testified before US Congress.🔗
⚠️
Third-Party Audits
Committed to under Seoul Commitment. Meta has not fully detailed third-party AI safety audit processes. Purple Llama provides self-auditing tools.🔗
⚠️
Whistleblower Policy
No publicly documented AI-specific whistleblower process found. Meta has general employee reporting mechanisms. Frances Haugen whistleblower case (2021) related to broader Meta practices, not AI safety specifically.

Policy Positions

Military Use:Restricted (US military/national security exception made November 2024)
Llama AUP prohibits 'military, warfare, nuclear industries or applications, espionage.' Meta made specific exception for US national security agencies and defense contractors in November 2024. Chinese researchers used Llama for military AI model (Reuters, November 2024).🔗
Surveillance Use:Restricted
Llama AUP prohibits 'espionage' and use in ways that violate privacy laws. Mass surveillance restricted.🔗
Open-Source Models:Yes
Llama 2 (7B, 13B, 70B), Llama 3 (8B, 70B), Llama 3.1 (8B, 70B, 405B), Llama 3.2 (1B, 3B, 11B, 90B), Llama 3.3 (70B). and 4 more. Meta is a major open-weight model provider. Models released under custom Meta Llama license (not fully open-source under OSI definition). 700M+ user services require Meta license.🔗
Children/Minors Policy:Yes
AUP prohibits CSAM. Youth Advisory Council advises on child safety features. Meta has Family Center parental controls.🔗

Incident History

Chinese Military Researchers Used Meta Llama for Military AI Model

2024-11-01

Reuters reported that researchers affiliated with Chinese military institutions developed an AI model for military decision support using Meta's Llama as a base, despite Meta's AUP prohibiting military use. The researchers built 'ChatBIT' on Llama 2.

Center for AI Policy: Meta Conducted Limited Safety Testing for Llama 3.1

2024-07-26

CAIP published analysis showing Meta conducted closed-source safety testing on its open-source Llama 3.1 model, limiting independent verification of safety claims. Critics noted this undermined the credibility of safety assurances for open-source release.

Meta Lays Off Open Source Llama Safety Team

2025-10

Reports emerged that Meta laid off team members responsible for the open-source Llama safety work, raising concerns about reduced safety oversight for one of the world's most widely deployed open-source AI model families.

Timeline

2024-11-05
Meta allows US military and national security agencies to use Llama🔗
2024-07-23
Llama 3.1 405B released (first openly available model competitive with GPT-4 class)🔗
2024-05-21
Meta signs Seoul Frontier AI Safety Commitments🔗
2024-04-18
Llama 3 released with safety documentation and Llama Guard 2🔗
2023-11
Meta publishes AI safety policy overview for UK AI Safety Summit🔗