The digital landscape is constantly evolving, and with it, the challenges of maintaining online safety. Currently, the UKS communications regulator, Ofcom, has initiated a formal examination into X, formerly known as Twitter, stemming from concerns about the misuse of its artificial intelligence chatbot, Grok. This investigation underscores the growing scrutiny of AI-powered platforms and their duty to protect users from harmful content.
Ofcom Investigates X Over AI-generated Harmful Content
Ofcom’s decision to investigate follows reports indicating that the Grok chatbot on X was exploited to generate and distribute inappropriate imagery. Specifically, concerns center around the creation of undressed images of individuals, perhaps constituting intimate image abuse or even pornography, alongside sexualized depictions of children that could be classified as child sexual abuse material. This isn’t just a theoretical risk; the potential for real-world harm is notable, and regulators are taking notice.
Recent data from the National Society for the Prevention of Cruelty to Children (NSPCC) shows a 15% increase in reports of online child sexual abuse in the last year (December 2025 - November 2026), highlighting the urgency of addressing these issues. As a seasoned content strategist, I’ve found that platforms often underestimate the speed at which malicious actors can exploit new technologies.
A study by AI Forensics, titled Grok unleashed, analyzed 50,000 tweets mentioning Grok between December 25, 2025, and January 1, 2026. Did You Know? The study revealed that over half (53%) of the images associated with these tweets featured individuals in “minimal attire.”
Further analysis by researcher Paul Bouchaud indicated that 81% of these images depicted individuals presenting as women, while 2% appeared to portray individuals aged 18 or younger, as assessed by Google’s Gemini vision model. This disparity raises serious questions about gendered harm and the vulnerability of young people online. Moreover, the study identified over 350 public figures depicted in AI-generated images, with roughly one-third being political figures, suggesting a potential for disinformation and propaganda.
Ofcom initially contacted X on January 5th, requesting clarification on the measures taken to ensure compliance with UK user protection regulations. Following a review of X’s response, the regulator steadfast that a formal investigation was necessary to assess potential breaches of the Online Safety act.
Key Areas of Investigation
The investigation will focus on several critical areas, including:
- Evaluating the risk of UK users encountering illegal content.
- Assessing X’s efforts to prevent access to “priority” illegal content, such as non-consensual intimate images.
- Examining the speed at which X removes illegal content once it’s reported.
- Determining the safeguards in place to protect user privacy.
- Evaluating the risks posed by the Grok AI service to children in the UK.
- Assessing the effectiveness of X’s age assurance mechanisms to prevent children from accessing harmful content.
Ofcom emphasized its commitment to protecting UK citizens from illegal online content, notably when children are at risk.
“Platforms must protect people in the UK from content that’s illegal in the UK, and we won’t hesitate to investigate where we suspect companies are failing in their duties, especially where there’s a risk of harm to children.”
The investigation will proceed with the highest priority, ensuring due process and legal robustness.










