- cross-posted to:
- technology@beehaw.org
- cross-posted to:
- technology@beehaw.org
More than 200 Substack authors asked the platform to explain why it’s “platforming and monetizing Nazis,” and now they have an answer straight from co-founder Hamish McKenzie:
I just want to make it clear that we don’t like Nazis either—we wish no-one held those views. But some people do hold those and other extreme views. Given that, we don’t think that censorship (including through demonetizing publications) makes the problem go away—in fact, it makes it worse.
While McKenzie offers no evidence to back these ideas, this tracks with the company’s previous stance on taking a hands-off approach to moderation. In April, Substack CEO Chris Best appeared on the Decoder podcast and refused to answer moderation questions. “We’re not going to get into specific ‘would you or won’t you’ content moderation questions” over the issue of overt racism being published on the platform, Best said. McKenzie followed up later with a similar statement to the one today, saying “we don’t like or condone bigotry in any form.”
IDGAF if it feeds into the narrative. It also shuts down a recruitment pipeline. It reduces their reach. It makes the next generation less likely to continue the ideology. De-platforming is a powerful tool that should be reserved for only the most crucial fights, but the fight against Nazi is one of those fights.
The Nazis were already full-blown conspiracy theorists. EVERYTHING is spun to feed into their narrative. That ship has sailed.
A platform operator needs to AT MINIMUM demonetize the content and censure it, and is likely only being responsible if they ban it outright. If you aren’t prepared to wade into the fraught, complex world of content moderation, don’t run a content platform.