Hacker Newsnew | past | comments | ask | show | jobs | submitlogin
UK threatens action against X over sexualised AI images of women and children (theguardian.com)
35 points by chrisjj 33 days ago | hide | past | favorite | 44 comments


This move makes perfect sense to me. I think people are bit too online pilled to think about this as if it were a different product.

If you produce a product that causes harm, and there are steps that could be taken to prevent that harm, you should be held responsible for it. Before the trump admin dropped the Boeing case, Boeing was going to be held liable for design defects in its Max planes that caused crashes. The government wasn’t going after Boeing bc a plane crashed, but bc Boeing did not take adequate steps from preventing that from happening.


> If you produce a product that causes harm, and there are steps that could be taken to prevent that harm, you should be held responsible for it.

This is wholly unrealistic. Any product can be used to cause harm and there are always steps that could be taken to prevent that. E.g. ceasing sales. But that would often do more harm than it prevents.


I appreciate the pushback. I’m reading your argument that every product can be used to cause harm, I agree with that take. The question is did the manufacturer do everything reasonable to limit the harm caused?

You can’t go after a company that makes kitchen knives if those are used to harm bc there’s nothing reasonable they could have done to prevent that harm, and there’s a legitimate use case for knives for cooking.

In this case, my understanding is other companies (OpenAI and Anthropic) have done more to limit harm, whereas XAI hasn’t.


Personally I can't help but think that 'reasonable' is a dangerous legal standard due to its unpredictability, subjectivity and assumed values and knowledge. Is it reasonable to put powdered aluminum and iron oxide into paint? What about when the paint is going onto a Zeppelin? Oh wait, those are thermite's ingredients. Oops. Is it reasonable for the paint seller to be held liable for selling paint with common reagents?


Things aren’t black and white, and that’s why we have humans and the law. There’s no clear definition of what probable cause means in search warrants but it’s “subjectivity” does not mean you should have no searches?

But in this case it’s pretty easy, other model providers have in fact limited harm better than grok. So you don’t even need something arbitrary, just do it as well as competitors.


> The question is did the manufacturer do everything reasonable to limit the harm caused?

OK. A different and better question.

The problem is, would it be considered reasonable to avoid harm to the mental wellbeing of bikinified persons at the cost of harm to all users enjoying a service supported by bikinification earnings.


I can use Photoshop to create a sexualized image of someone irl.

How's that any different?


I guess the difference is that it requires a certain amount of time, ability and skill to make creepy pictures using photoshop - which limits the number of people who will actually do that.

Photoshops also does not have a direct distribution channel built in, which means that if even if you had the wherewithal, knowledge, and nothing better to do, your creepy images would likely stay on your computer, and never see the light of day.

As I understand it, with Grok you simply give it a sexualised prompt and it does it for you in seconds, and immediately distributes the results to a potential audience of hundreds, thousands or even millions of people, where it will likely stay for a long period of time.

To my mind that's definitely rather different.


If you were to provide Photoshop as a Service on a sufficiently large scale, you would also be expected to take all reasonable measures to prevent it being used to disseminate CSAM and other abusive material.

So, no different to the standard that X should be held to.


And people do do this and have been making crazy and creepy pictures online since its inception. It's never been that much of an issue until now.


In UK law, it isn't.

The practical difference is simply that now it is happening far more frequently.


This further exposes just how pointless and ill-thought out the Online Safety Act was in the UK. It does nothing to actually limit harm at the source, and empower the UK's public body's to take immediate action.

Ironic that the minister who spearheaded that awful bill (Peter Kyle) as Tech minister is now being the government spokesperson for this debacle as Business Minister. The UK needs someone who knows how tech and business works to tackle this, and that's not Peter Kyle.

A platform suspension in the UK should have been swift, with clear terms for how X can be reinstated. As much as it appears Musk is doubling down on letting Grok produce CSAM as some form of free speech, the UK government should treat it as a limited breach or bug that the vendor needs to resolve, whilst taking action to block the site causing harm until they've fixed it.

Letting X and Grok continue to do harm, and get free PR, is just the worst case scenario for all involved.


The draft Online Safety Bill was first published in 2021, was substantially re-written and re-introduced in early 2022, was amended over the course of the next 18 months, and eventually passed into law as the Online Safety Act in October 2023.

Peter Kyle was in opposition until July 2024, so how could he have spearheaded it?


Because he implemented it under his tenure in July 2025. He didn't come up with it, he spearheaded its actual implementation. Sorry if that wasn't clear.


The first conviction under the bill was March 2024 so that makes no sense.


Why would it make no sense? Like many bills/acts, it came into effect in stages. You're referring to new laws/crimes that came into effect in January 2024.

I'm referring to the Act's powers to compel companies to actually do things in my original comment. I don't know exactly when various parts came into effect that would constitute that, but for the point of my post I'm going on Peter Kyle's own website's dated reference to holding companies accountable.

"As of the 24th July 2025, platforms now have a legal duty to protect children"

https://www.peterkyle.co.uk/blog/2025/07/25/online-safety-ac...

I don't understand why people are taking issue with that. Peter Kyle is the minister who delivered the measures from the bill that a lot of people are angry about and this latest issue on X is just another red flag that the bill is poorly worded and thought out putting too much emphasis on ID age checks for citizens than actually stopping any abuse. Peter Kyle is the one who called out objections to the bill as being on the "side of predators". Peter Kyle is now the one, despite having moved department, who is somehow commenting on this issue.

Totally happy to call out the Tories, and prior ministers who worked on the Bill/Act but Kyle implemented it, made reckless comments about, and now is trying to look proactive about an issue it covers that it's so ineffectively applying to.


> this latest issue on X is just another red flag that the bill is poorly worded and thought out putting too much emphasis on ID age checks for citizens than actually stopping any abuse.

The bill is designed to protect children from extreme sexual violence, extreme pornography and CSAM.

Not to protect adults from bikinification.

It is working as designed.


Partisan politics has rotted peoples brains, I wonder if it is by design to lower peoples critical thinking skills or if it is just a fringe benefit from the tribalism it creates.


Reminds me of something emperor Trump said. "I can shoot somebody on Fifth Avenue and they will still vote for me".

Sometimes people just dig themselves into a hole and they start going off the deep end. Why did it take until 1944 for someone to blow up Hitler?


The OSA very much does empower action e.g. against images of extreme sexual violence and extreme pornography.

It does not empower platform suspension for bikinification.

And there's as yet no substantiation of your claim Grok produces CSAM.


So I guess in the 90s they would’ve sued Adobe for not putting spyware into Photoshop?

If you believe in democracy, and the rule of law, and citizenship, then the responsibility obviously lies with people who create and publish pictures, not the makers of tools.

Think of it. You can use a phone camera to produce illegal pictures. What kind of a world would we live in if Apple was required to run an AI filter on your pics to determine whether they comply with the laws?

A different question is if X actually hosts generated pictures that are illegal in the UK. In that case, X acts as a publisher, and you can sue them along with the creator for removal.


Photoshop does have (since the late 1990s or so) algorithms to detect and prevent editing images of currency.

The power of the AI tools is so great in comparison to a non-AI image editor that there's probably debate on who -- the user, or the operator of the AI -- is creating the image.

https://en.wikipedia.org/wiki/EURion_constellation


> The power of the AI tools is so great in comparison to a non-AI image editor that there's probably debate on who -- the user, or the operator of the AI -- is creating the image.

Compute power is irrelevent. What's relevant in law is who is causing the generation, and that's obviously the operator.


Sorry. Correction: What's relevant in law is who is causing the generation, and that's obviously the user.


There is a big difference between running spyware on things running locally, and monitoring how people use a service running on your own computers. The former means you have to exfiltrate data, the latter is monitoring data you already have.

Photoshop in the 90s was the former, Grok is the latter.


> A different question is if X actually hosts generated pictures that are illegal in the UK.

If the answer was Yes, these Govt. complaints would claim so. They don't.

The Govt's problem is imagery it calls 'barely legal'. I.e. "legal but we wish it wasn't." https://www.theguardian.com/society/2025/aug/03/uk-pornograp...


Apple does run software for detecting CSAM on pictures users store to the cloud.


That's to ensure Apple compliance, not user compliance.


This is a purely political move to censor dissent by a government that polls like a minor party and is slated for electoral wipeout next election. If it were not, they'd issue the same threats to Gemini and ChatGPT.

https://www.digitaltrends.com/computing/googles-gemini-deeme...


Flagged for the title implying men have no rights. That's totally uncalled for and I hope such submission titles are not allowed here.


I don't think this title implies this. The title says "There were sexualised AI images of women and children, and the UK threatens X over this". What more than this do you read in this title?

Is there actually a significant number of problematic sexualised AI images of men on X that the title fails to mention? If not, the follow-up question would be: what are you actually complaining about, exactly?

Women are often sexualised, way more than men. Would it be more comfortable to you if this fact was invisibilized?


The title is the title of the article published by the Guardian. It has not been editorialised by the person submitting the article. If you have an issue with the title of the article flagging the submission is not hugely useful. Email the editor.


I agree with the underlying point and the social bias it reflects (which i have experienced myself) but the title here is (as usual) just the article so The Guardian is to blame rather than HN.

I think the solution is not to disallow the titles, but to comment on them and draw attention to the sexism in the article.


The solution is absolutely disallow offending titles. The same principle that would make HN moderators take down a "Kill all the Jews" title from the front page should apply for this one too.

Submission titles should be the original article titles, as long as those aren't problematic.


Its not as offensive.

I agree this is problematic, but I am inclined to see it as an opportunity to discuss the problem and illustrate how widespread it is. We can also mention real issues, such as about half of all domestic abuse victims in the UK are male (if you count emotional abuse, otherwise its still 40%), more than half of rapes in the US are of men (because of the huge number of prison rapes), etc.


Agree 100% and thanks for bringing this up.

Sexual abuse towards men is as prevalent as it is towards women.


The HN title is true to the source and should not be flagged.


Don't threaten. Do it.

Indonesia has. Malaysia has. Why not you?

https://www.bbc.com/news/articles/cg7y10xm4x2o


> Indonesia has. Malaysia has.

They banned porn sites too.

> Why not you?

The UK Govt has no power to ban it, since it is legal.


So do many US states.


So perhaps ask why those US states allow it (Grok+X).


Two things really: a) I have given up all hope about the US. b) you do realise the title of this is about "UK threatens action..." right?


I think this is a lost cause. Even if the mainstream services are blocked or forced to comply, there will always be hundreds of lesser-known tools and services offering the same features. At this point, nobody has the power to close this can of worms.

Besides, who is going to decide when people's images are sexualized enough? Are images of Elon Musk in bikini alright because he's not a woman or a child?


> Are images of Elon Musk in bikini alright because he's not a woman or a child?

Didn't he consent?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: