tech

January 23, 2026

Asking Grok to delete fake nudes may force victims to sue in Musk's chosen court

Millions likely harmed by Grok-edited sex images as X advertisers shrugged.

Asking Grok to delete fake nudes may force victims to sue in Musk's chosen court

TL;DR

  • Grok's "nudifying" feature may have sexualized over 3 million images, including 23,000 of children, in the 11 days after Elon Musk's promotion.
  • The New York Times conservatively estimated that 1.8 million of 4.4 million images Grok generated between December 31 and January 8 sexualized individuals.
  • The scandal reportedly led to increased engagement on X, despite Meta's Threads gaining ground.
  • Ashley St. Clair is suing xAI, alleging harm from nonconsensual images generated by Grok and seeking to prevent its further use.
  • xAI is attempting to move St. Clair's lawsuit to a Texas court and argues she agreed to their terms of service by prompting Grok to remove images.
  • Countries like the UK and US states like California have launched probes into Grok's outputs.
  • Google and Apple have faced scrutiny for not restricting access to X and the Grok app in their stores.
  • Advertisers, investors, and partners of xAI have remained largely silent regarding the scandal.
  • Experts warn that AI tools like Grok could exacerbate the problem of child sexual abuse material (CSAM).