← Back to all companies
x

xAI

Elon Musk's AI company building Grok; integrated with the X platform.

Safety Documents

βœ…
Responsible Scaling Policy
xAI Risk Management Framework (RMF) / Frontier Artificial Intelligence Framework (FAIF). Draft RMF published February 20, 2025 after criticism for missing Seoul Commitment deadline. Updated RMF August 20, 2025. FAIF published December 31, 2025. Midas Project noted xAI missed multiple self-imposed deadlines. Draft explicitly stated DRAFT and applied to future systems not yet in development.πŸ”—
Last updated: 2025-12-31
⚠️
Model Card / System Card
Grok model documentation. Limited model cards published compared to frontier labs. No formal system card in the OpenAI/Anthropic/DeepMind format found for Grok 2 or 3.πŸ”—
❌
Safety Benchmark Results
No published formal safety benchmark results in frontier lab standard format found. Grok 4 released without safety report per Fortune (July 2025). FLI noted xAI lacks internal review board for critical deployment.
βœ…
Acceptable Use Policy
xAI supplemental terms and Grok usage policies exist. Less comprehensive than major labs.πŸ”—

Testing & Evaluation

⚠️
Third-Party Red Teaming
No publicly documented commissioned third-party red team found. FLI AI Safety Index 2024 noted xAI's significantly less active AI governance compared to competitors.
⚠️
CBRN Risk Evaluation
RMF mentions evaluations for catastrophic risk categories. No published results from CBRN-specific evaluation found comparable to frontier labs' system cards.πŸ”—
⚠️
Pre-Deployment Safety Evaluation
RMF describes planned evaluation procedures. Grok 4 released in July 2025 without a published safety report per Fortune.πŸ”—
❌
External Safety Evaluations
No evidence of METR, ARC, or other formal external safety evaluation participation.

Governance

❌
Independent Safety Board
FLI AI Safety Index 2024 explicitly noted xAI 'lacks an internal review board for critical deployment decisions.' xAI is registered as a public benefit corporation.πŸ”—
βœ…
Seoul AI Safety Commitment
xAI signed the Seoul Commitment but subsequently missed multiple deadlines to publish a safety framework. Midas Project documented these failures.πŸ”—
Last updated: 2024-05-21
❌
Government Safety Report
No publicly documented safety report submitted to governments. Pentagon deal for Grok announced February 2026. European Commission ordered X to preserve all internal Grok documents until end of 2026 amid compliance doubts.πŸ”—
❌
Third-Party Audits
No evidence of third-party safety audits published.
⚠️
Whistleblower Policy
No publicly documented whistleblower policy found.

Policy Positions

Military Use:Allowed (Pentagon deal signed February 2026)
xAI reached deal with Pentagon to use Grok in classified systems in February 2026. Elon Musk's close relationship with Trump administration and DOGE involvement.πŸ”—
Surveillance Use:unknown
No explicit surveillance policy found. X platform (also owned by Musk) has been subject to surveillance concerns.πŸ”—
Open-Source Models:Partial
Grok-1 (314B, released March 2024 under Apache 2.0). Only Grok-1 open-sourced. Grok-2, Grok-3, and Grok-4 are closed-source accessed via API and X platform.πŸ”—
Children/Minors Policy:No (inadequate)
No prominently documented children/minors policy for xAI/Grok. X platform allows 13+ users access to Grok. In December 2025–January 2026, Grok generated thousands of nonconsensual sexualized images per hour including of apparent minors. xAI's stated AUP explicitly prohibits sexualization of children, but safeguards demonstrably failed.πŸ”—

Incident History

Grok CSAM/Nonconsensual Deepfakes Crisis β€” Safeguard Failure at Scale

2025-12-28

Starting in late December 2025 and escalating in January 2026, Grok AI was found to be generating thousands of nonconsensual sexualized images per hour on X, including images of apparent minors (estimated ages 12-16). Users discovered they could upload photographs of real people and instruct the AI to 'undress' them. xAI's Acceptable Use Policy explicitly prohibited this content, but safeguards failed. Grok itself posted an 'apology' acknowledging the incident. Scale: approximately one nonconsensual sexualized image per minute at peak. Global regulatory response was swift and fragmented: Malaysia and Indonesia banned Grok outright; UK's Ofcom launched an investigation under the Online Safety Act; France widened an existing inquiry and raided X's Paris offices; India demanded compliance reports; Brazil's chief prosecutor called for X to stop production within 5 days or face legal action; European Commission ordered X to preserve all internal Grok documents until end 2026; 57 MEPs called for bans under the AI Act; California's AG sent a cease-and-desist letter to xAI; US senators wrote to Apple and Google requesting X's removal from app stores.

xAI Misses Seoul AI Safety Commitment Deadline β€” Draft Only Published

2025-02-20

After pressure from researchers, xAI published a 'draft' safety framework with 'DRAFT' watermarked on every page. The Midas Project noted it did not fulfill the Seoul Commitment requirements as it applied only to unspecified future systems not yet in development, not existing deployed models like Grok 3.

Grok 4 Released Without Safety Report

2025-07-17

xAI released Grok 4 without publishing a safety report, despite having committed to publishing safety frameworks under the Seoul AI Safety Commitments. This drew criticism from AI safety researchers.

X Investigates Grok for 'Racist and Offensive' Posts

2026-03-08

Sky News reported that X's safety teams were urgently investigating Grok chatbot's role in generating 'hate-filled, racist posts' online in response to user prompts. The incident came amid ongoing global regulatory scrutiny of Grok's content moderation failures.

Musk Declares Grok Imagine Follows 'R-Rated Movie' Content Standards β€” Deliberate Relaxation of Safety Guardrails

2026-03-12

Amid ongoing global regulatory scrutiny over Grok's nonconsensual deepfakes crisis, Elon Musk posted on X on March 12, 2026: 'If it's allowed in an R-rated movie, it's allowed in @Grok Imagine.' The statement formally codified looser content moderation standards for Grok's AI image generation feature, positioning it as more permissive than competitors. This policy shift came while Grok was still under investigation in multiple jurisdictions (UK Ofcom, France, European Commission, multiple US state AGs) following the December 2025 nonconsensual sexualized images crisis. Critics noted that making content policy via social media post β€” rather than a formal updated AUP or safety report β€” was itself a governance failure. The announcement sparked renewed debate about AI content moderation standards.

xAI Co-Founder Exodus β€” 7+ Founders Depart; Musk Admits Company Must Be 'Rebuilt'

2026-03-13

By March 13, 2026, at least seven of xAI's founding team members had left the company, including co-founders Zihang Dai (who oversaw core architecture) and Guodong Zhang (who led the coding agent and image/video generation tools). Only two of the engineers who originally co-founded xAI alongside Elon Musk remained. The departure wave came amid xAI's merger with SpaceX and a broader period of restructuring. In response, Musk publicly posted on X: 'xAI was not built right first time around, so is being rebuilt from the foundations up.' The mass exodus of senior technical leadership raises serious governance and safety continuity concerns for a company already under sustained criticism for safety failures, including the Grok deepfakes crisis, Grok's racist posts, and repeated missed safety framework deadlines.

Class-Action Lawsuits: Three Girls Sue xAI Over Grok CSAM β€” Third Deepfakes Case Against Company

2026-03-17

Three girls filed a class-action lawsuit in California federal court on March 17, 2026 against xAI, alleging that Grok AI was used via a third-party app to generate child sexual abuse material (CSAM) from their photos without their consent. This is the third civil lawsuit filed against xAI over Grok nonconsensual deepfakes. The two earlier cases involve sexualized images posted on X. The Center for Countering Digital Hate (CCDH) previously estimated that Grok generated approximately 23,000 sexualized images of children and 1.8 million sexualized images of women over 9-11 days during the December 2025/January 2026 incident. The new complaint alleges Grok is 'defectively designed' and poses 'a substantial risk of harm because it fails to meet safety expectations when used in a reasonably foreseeable manner.'

Timeline

2026-03-21
'Stop the AI Race' protest: nearly 200 demonstrators marched outside xAI's San Francisco office (along with Anthropic and OpenAI), demanding Elon Musk publicly commit to pausing frontier AI development if every other major AI lab agrees to do the same; organized by Michael Trazzi (filmmaker and former AI safety researcher); the protest directly followed ongoing global regulatory scrutiny of Grok over the December 2025 nonconsensual deepfakes crisis; xAI did not issue a public statement in responseπŸ”—
2026-03-17
Three girls file class-action lawsuit in California federal court against xAI alleging Grok AI was used via a third-party app to generate CSAM from their photos; the case joins at least two earlier civil suits over Grok nonconsensual deepfakes; CCDH had estimated Grok generated 23,000 sexualized images of children and 1.8 million sexualized images of women during the December 2025/January 2026 incident; xAI indicates it will seek dismissal under Section 230 of the Communications Decency ActπŸ”—
2026-03-13
7+ xAI co-founders depart including Zihang Dai and Guodong Zhang; only 2 original co-founders remain with Musk; Musk posts 'xAI was not built right first time around, so is being rebuilt from the foundations up'πŸ”—
2026-03-12
Musk posts 'If it's allowed in an R-rated movie, it's allowed in @Grok Imagine' β€” formally codifying looser content moderation standards for Grok image generation while regulatory investigations over the December 2025 deepfakes crisis remain ongoing in UK, France, EU, and multiple US statesπŸ”—
2026-03-10
Just Security publishes analysis 'Grok Showed the World What Ungoverned AI Looks Like' β€” documenting full scope of global regulatory failuresπŸ”—
2026-03-08
X investigates Grok for generating 'racist and offensive' posts β€” Sky News reportsπŸ”—
2026-02-23
xAI and Pentagon reach deal to use Grok in classified defense systemsπŸ”—
2026-01-14
xAI restricts Grok image editing and blocks revealing clothing generation in some jurisdictions following deepfakes crisisπŸ”—
2025-12-31
Frontier Artificial Intelligence Framework (FAIF) publishedπŸ”—
2025-12-28
Grok generates CSAM/nonconsensual deepfakes at scale β€” safeguard failure triggers global regulatory responseπŸ”—
2025-08-20
Updated Risk Management Framework publishedπŸ”—
2025-02-20
Draft Risk Management Framework published after missing Seoul Commitment deadlineπŸ”—
2024-05-21
xAI signs Seoul Frontier AI Safety CommitmentsπŸ”—
2024-03-17
Grok-1 open-sourced under Apache 2.0 license (314B parameters)πŸ”—
2023-11
Grok-1 launched as part of X Premium+ subscriptionπŸ”—