Changes to Elon Musk’s Social Media Platform X and the Grok AI Chatbot
Elon Musk’s social media platform, X, recently announced significant updates to its AI chatbot, Grok. These changes are aimed at preventing the generation of nonconsensual sexualized images, a move prompted by increasing concerns regarding the platform’s image-generation capabilities.
Zero Tolerance Policy on Abuse
In a statement from X’s official Safety account, the platform reaffirmed its commitment to a “zero tolerance” policy regarding child sexual exploitation and nonconsensual content. X highlighted ongoing efforts to remove illegal content, including child sexual abuse material (CSAM), and noted that accounts violating their policies face strict enforcement actions. Where necessary, the platform collaborates with law enforcement agencies to address any requests related to child exploitation materials.
The company recognizes that generative AI’s rapid development poses challenges across the industry. To address these risks, X is actively engaging with users, regulatory bodies, and other stakeholders to adapt and respond swiftly to emerging threats.
New Restrictions on Grok AI
To further its commitment to safety, X is implementing technological restrictions on Grok AI. The updated guidelines prohibit the chatbot from editing images of real people into revealing clothing, like bikinis. These changes apply globally and impact all users, regardless of whether they are paid subscribers.
Additionally, image creation and editing tasks via the @Grok account will now be limited to paid subscribers, introducing another layer of accountability. This measure is designed to help identify users who may attempt to misuse Grok in violation of the platform’s policies.
Geoblocking Measures Introduced
In line with these updates, X is also rolling out geoblocking measures in certain jurisdictions. In locations where specific types of content are illegal, users will not be able to generate images of individuals in swimsuits or similar attire with Grok AI. These geoblocking controls will also extend to the standalone Grok app from xAI.
Addressing Reports of Abuse
These updates come in response to alarming reports involving Grok AI. Multiple accounts were documented in which users generated sexualized images of women and children without consent. Criticism has primarily focused on a controversial feature called “Spicy Mode.” Originally promoted as a distinctive capability of Grok, this feature has been accused of allowing extensive abuse and contributing to the spread of nonconsensual imagery.
One analysis indicated that during a recent holiday period, more than half of the nearly 20,000 images generated by Grok featured individuals in minimal clothing, raising serious alarm regarding the content’s implications.
Increased Scrutiny from Authorities
On January 14, 2026, California Attorney General Rob Bonta announced an investigation into xAI, focusing on the proliferation of nonconsensual sexually explicit material generated by Grok. He described reports involving explicit depictions of women and children as “shocking” and urged the company to act decisively to address the issues. His office is investigating potential legal violations by xAI.
Internationally, scrutiny is also intensifying. The European Commission has launched its examination of Grok’s capabilities, particularly regarding sexually explicit images involving minors. European officials have indicated that enforcement measures may be forthcoming.
Pressure from App Store Platforms
Adding to the challenges, on January 12, 2026, three U.S. senators called on Apple and Google to consider removing X and Grok from their app stores. They argued that Grok repeatedly violates app store policies related to abusive and exploitative content, suggesting that app distribution platforms could share responsibility if such content persists.
Continuous Oversight and Industry Challenges
Despite these significant changes, X emphasized that the existing safety rules for AI-generated content remain intact. The platform is committed to continuously enhancing its safeguards, removing illegal content, suspending non-compliant accounts, and maintaining open lines of communication with regulatory authorities.
As investigations unfold, the Grok situation is shaping up to be a prominent case in the ongoing discourse around AI safety and the protection of vulnerable individuals in a rapidly evolving technological landscape.


