
The restrictions aim to curb non‑consensual image manipulation, reducing legal risk for X and signaling compliance with the UK Online Safety Act. They also lower the probability of a full platform ban while keeping regulators watching.
The controversy surrounding Grok AI highlights the growing tension between generative technology and privacy law. Since its integration with X, the tool allowed users to request realistic alterations of photographs, sparking concerns over non‑consensual deepfakes and "revenge porn". Regulators in the UK have been particularly vigilant, as the Online Safety Act mandates platforms to proactively mitigate illegal content. By curbing Grok’s ability to produce revealing images, X is attempting to align its services with evolving legal expectations while preserving the broader appeal of AI‑driven features.
X’s technical response combines subscription‑based access controls with geographic restrictions, a strategy known as geoblocking. All users, whether free or paying, now face a blanket prohibition on generating images of real people in bikinis or underwear, while the broader image‑creation capabilities are limited to paying subscribers. In jurisdictions like the United Kingdom, where distributing intimate images without consent is a criminal offense, the platform will automatically block such requests. This layered approach not only simplifies enforcement but also creates a traceable trail for any violations, addressing regulator demands for clearer risk assessments and age‑verification mechanisms.
The ongoing Ofcom inquiry remains a critical factor in X’s future operating environment. Investigators are assessing whether the company failed to assess risks, remove illegal content promptly, and protect minors, with potential penalties reaching up to 10 % of global turnover. While X’s recent measures may reduce the likelihood of an outright ban under the Online Safety Act, the regulator retains the authority to impose substantial fines or mandate further compliance steps. The outcome will set a precedent for how social platforms balance AI innovation with stringent content‑safety obligations, influencing policy discussions across the EU and beyond.
Comments
Want to join the conversation?
Loading comments...