Hacker News new | past | comments | ask | show | jobs | submit login

8chan is lawful content. Vile, hateful content, but completely lawful. They comply with DMCA and moderate to remove illegal material. If they didn’t, the FBI could already have had them pulled offline just like they do to ISIS websites.

The logical end to deplatforming is arguing that ISPs should be able to block or decline customers based on the content they are hosting. Otherwise a customer can buy business-class internet and host their immoral content themselves on a server farm in their home, which takes away the whole point of deplatforming which is to make the content no longer available on the internet.




> The logical end to deplatforming is arguing that ISPs should be able to block or decline customers based on the content they are hosting.

Not at all. That's an obvious strawman.

You're mischaracterizing what's going on there. If I run a company, it is my right not to make business with radical hate groups and terrorists. They will be someone else's problem then.

That's exactly the reason that Cloudflare has given, not some nebulous talk about "deplatforming".

> which is to make the content no longer available on the internet

LOL. That is decidedly not the purpose of deplatforming, as the word "de-platforming" readily suggests.

> the FBI could already have had them pulled offline just like they do to ISIS websites

It is obvious to me as an outside observer that the FBI applies justice selectively. Domestic terrorism is underrated. Of course, 8chan could be raided and closed for the same reasons as ISIS websites are raided and closed. The laws are there and 8chan could easily be considered aiding and fostering domestic terrorism. The laws are just not applied in this case.

It's also kind of 'reasonable' not to apply them as harshly, since US judges and juries suffer from the same bias. They are unlikely to judge of some deranged gun nut that he was planning or aiding a terrorist attack. They are highly likely to judge of some deranged ISIS sympathizer that he was planning or aiding a terrorist attack. Police authorities make the call on what to pursue and what not to pursue based on the prospects of a successful trial.


>If I run a company, it is my right not to make business with radical hate groups and terrorists. They will be someone else's problem then.

This type of thing is possibly one of the hardest ethical issues to tackle. On the one hand, I don't support racists and fascists at all. But on the other hand, I recognize the potential damage in carving out these exceptions in free speech. As social mores change, the ideas of "acceptable" free speech may change, and we need to be cognizant of the ways that these exceptions could be abused long term. Otherwise, we're just setting up future generations for a collapse of the concept of freedom of speech.

I think the answer to solving hate and bigotry goes much, much deeper than preventing people from speaking their hateful and bigoted views. All that's going to do is sweep the problem under the rug, and eventually that problem will come back out some orders of magnitude worse. Perhaps we could do things like make it illegal to teach kids hate and bigotry? But then you've got the entirety of America mad at you because you're "telling people how to raise their kids". Mere advocacy against bullying and hate doesn't really seem to be working.

I think we'll see better gains in this area when we stop trying to find the first thing we can to "blame" these mass shooting on, and arguing endlessly about what that cause is (guns, video games, unrestricted freedom of speech, etc). We need to dig deep. I think if we understood more on the topic of mental health, we'd have a better chance at understanding these situations.




Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: