More than 200 Substack authors asked the platform to explain why it’s “platforming and monetizing Nazis,” and now they have an answer straight from co-founder Hamish McKenzie:
I just want to make it clear that we don’t like Nazis either—we wish no-one held those views. But some people do hold those and other extreme views. Given that, we don’t think that censorship (including through demonetizing publications) makes the problem go away—in fact, it makes it worse.
While McKenzie offers no evidence to back these ideas, this tracks with the company’s previous stance on taking a hands-off approach to moderation. In April, Substack CEO Chris Best appeared on the Decoder podcast and refused to answer moderation questions. “We’re not going to get into specific ‘would you or won’t you’ content moderation questions” over the issue of overt racism being published on the platform, Best said. McKenzie followed up later with a similar statement to the one today, saying “we don’t like or condone bigotry in any form.”
Freedom of speech doesn't mean that you are obligated to host a platform so shitty people can use it to share shitty ideals. It simply means that you won't get arrested on a federal level.
Websites can do whatever they want, including deciding that they don't want to be a platform for hate speech. If people are seeking a place for this conversation genre to happen, and they want it enough, they can run their own website.
Imagine if you invited a friend of a friend over, and they were sharing nasty ideals at your Christmas party. And they brought their friends. Are you just going to sit there and let them turn your dinner into a political rally? No, you're going to kick them out. It's your dinner, like it is your website. If you don't kick them out, then at some level, you're aligning with them.
I like your example there a lot, I’m going to use that in the future when I’m trying to express that notion. In the past I’ve never been able to articulate that exact concept. So thanks!