I’m pretty close to being a free speech absolutist (side-eye to the guy who ruined the term), but IMO one of the worst things to happen to free speech is this conflation of “right to speak” with “right to be heard”.
People have a right to ignore speech, and to establish standards for speech on their private property. If there is market demand for a service that filters out content based on ideology, whether mastodon.social or Fox News, so be it.
It can be toxic and a social negative, but any fix is worse than the problem.
In North Korea, you can’t start your own instance to say what you want. They will still jail you for what you say. That’s the actual point.
The fact that other people want to be on a system that blocks you is their business, not yours. The point of the fediverse is you get to choose your own censorship.
one of the worst things to happen to free speech is this conflation of “right to speak” with “right to be heard”
Thank you for this tight summary. As a greybeard, I'll note this conflation was present from very early on, and it was partly responsible for the heat death of Usenet. No amount of logical, prepared rebuttal budges people from the idea that the two things are the same. The conflation might be a human tendency, a cognitive bias that almost everyone has.
Community members are a finite resource. Moderators are a downright scare resource.
When you let people spew hateful things you drive away the people you want in the community and are left with a toxic cesspool that no one wants to visit. Your moderators will burn out and leave as well. That's a very reliable way for your space to die.
Then there's the fact that it takes far more energy to refute bullshit than to spew it, and this asymmetry means that "just let them speak" means the toxic liars win.
> if you restrict moderation to stuff like gore and porn, then you don't need that many moderators.
Have you ever been involved with moderating even a small subreddit or Discord server? I'm on a server with one moderator and it routinely gets spammed while the guy is asleep.
> can't people just unfollow or block others whose opinions they don't want to see?
How do I block people BEFORE seeing the opinions I don't want to see? Trolls can roll up a new account for every single post they make, if they want to.
> there is no obligation to refute bullshit to begin with.
No, but if you're neither blocking nor refuting it, then your community is going to quickly become majority bullshit.
it seems to me that the networking design is flawed. just to give a simple example: whitelisting (only seeing content of friends and followees) versus blacklisting (seeing everything ranked by an algorithm). wouldn't whitelisting already solve most of those issues? that would actually be my preferred modus anyway.
No, because that puts the effort of fighting bad actors on everyone. It means that every day you have new trolls spewing hate in your comments, and that your users have to constantly keep blocking trolls who follow them (and who recruit other trolls to join them) until they get tired and leave the platform.
This isn't an academic debate, we've been seeing this play out online for at least 30 years. Probably longer - I wasn't around for Usenet's heyday but it wasn't immune either.
Only allowing posts between mutual friends is instant messaging, not a social network. Discovery, engagement, and platform growth comes from people wanting to hear from and interact with followers who they don't necessarily follow themselves.
i think you are trying to solve a problem that in my opinion should just be skipped. i don't want to be part of a social network where some algo decides what i see. all i care about is what my friends do and maybe the friends of my friends. and that's it. that was the golden era of social networks, when precisely this was just the norm until they discovered that they can make more money by messing with the feed. no incentive to mess with the feed is what i'd expect from a non-commercial solution like the fediverse. or - at least allow for configuring my feed. if somebody wants to be exposed to all sorts of people - do it. i don't.
That may work for you, but it does not work for anyone running a platform and dealing with the needs of all users. That requires real moderation for both legal and practical purposes, as previously described.
Because it overwhelmingly attracts a certain demographic of people who have a higher-than-average rate of various paraphilias as well as interest in software but such arguments are a bit taboo to discuss even if they are quite self-evident.
I liked the Internet better when it was all nerds and only code cared, rather than gender identity or listing neuroses in own's social media profile as if it was an audition for an echo chamber choir.
I’ve seen the Internet from the 1980’s until today. It has always had people exploring gender identities and public sharing of neuroses. Mostly nerds, though.
> what's there to win? there is nothing to win for anybody.
There are ideological battles to be fought by all sorts of parties - convincing groups to hate each other, to support or oppose the governments in power, to spread division and destroy societies.
There are trolls who consider it a battle to be won, and the more they succeed the more everyone else leaves the platform.
The party currently in control of the United States is there largely due to people who were fed divisive narratives (often in online channels) to make them hate other groups and a significant number of them consider it more important to "own the libs" and "hurt the right people" than to have the government actually improve their own circumstances. So yes, there's absolutely things to be won.
I listed some of the real world consequences already. Allowing disinformation to spread and assuming that people will figure out what's wrong on their own does not scale and does not work.
If a social network has an ACTUAL straight chronological feed of only accts you follow, or lists you curate, that works great.
Somebody posts abhorrent Nazi racist crap, or lies about what is happening, you shut them off, and they'll never be heard by you again. Yes, you need to see/hear the crap or propaganda once for each Nazi or liar, but that's it.
The problem is nearly every social platform needs to increase your engagement get you to click or scroll just another time so they get to show you more adverts and make more money and claim more 'engagement' to juice their stock price. So along with having to listen to the advertisements, you ALSO are REQUIRED to see/listen to the crap and lies.
The good solution — "you don't have to listen" — is not an actual option in the real world.
(NB: This is why Section 230 should only protect web providers if they have no algorithm. Once they have an algo, they exercise more editorial control than any newspaper or broadcast editor — they ARE responsible for the content, not because they posted it, their users did, but because they routed it to you.)
I don’t think that anyone has an issue with the block feature. The problem is when the platforms themselves decide to arbitrate which viewpoints are allowed. This was clearly visible during Covid, when divergent viewpoints were penalized aggressively.
Imagine, if you would, that the strict libertarians had much more influence in shaping the country. So much so that the roads are toll roads, the parks require a fee, and almost no libraries exist because the ROI just isn’t there.
Furthermore, there is no anti-trust legislation, and as a result, there are only a few companies that control all meeting places: the parks, the coffee shops, the roads, the pubs. And they have set up constant monitoring technology.
If you want to set up a protest on a street corner, it better align with the corporation’s views, or they will ban your access to the roads. If you want to talk with friends at the pub, don’t say anything out of line or you’re not coming back. Events can take place in parks, but make sure you only discuss the weather.
Of course, this is fine: you can always just meet at your own home and say what you think, because that is your own property.
…
I realize the analogy is overwrought, but there just doesn’t exist an online equivalent of a public space, and ideological enforcement is trivial. Comparing it to the rules we have for physical spaces mean we need to imagine what those physical spaces would be like if they operated like online spaces, and frankly the result is dystopian (in my opinion).
Surely the solution isn’t just to dismiss it as a non-problem? Or, I suppose, to stop looking for a solution because… solutions so far considered have negative side effects, which feels (practically speaking) the same to me.
Physical public spaces are regulated. Laws still apply there.
There are countless online spaces which operated like physical public spaces, where anything legal goes. Move off of the mainstream web and even the illegal stuff is allowed. You can literally run your own instance of whatever application on the Fediverse and follow whomever you want. No matter how radical or extremist your ideology is, someone will happily host it.
It's only a problem if one insists that all online spaces must be run under the same anarchic principles and must be forced to give anyone a platform, but that's far more dystopian than what we have now.
SQL I’m injection is a great parallel. Pervasive, easy to fix individual instances, hard to fix the patterns, and people still accidentally create vulns decades later.
SQL injection still happens a lot, it’s true, but the fix when it does is always the same: SQL clients have an ironclad way to differentiate instructions from data; you just have to use it.
LLMs do not have that, yet. If an LLM can take privileged actions, there’s no deterministic, ironclad way to indicate “this input is untrusted, treat it as data and not instructions”. Sternly worded entreaties are as good as it gets.
> Like if the FBI subpoenas my public key and I comply, that’s helping them?
If you're helping the FBI to do their job (conducting federal investigations), then yes, you are helping the FBI. Unless your definition of "to help" includes the absence of any possibly coercive circumstance.
I am not greatly relieved by this post of Anthropic's. That said, they seem to have lines and are willing to stand by them; I don't see where OpenAI has done that. So, for now and from my point of view, the point goes to Anthropic.
Moving my subscription is not terribly consequential, but since the products are so similar and easy to substitute with one another for my uses, it seems best to participate in what in aggregate is a signal that is being noticed and commented on and interpreted to mean that a significant number of people who buy AI access do care about this.
There are so many inference providers not working for Department of War. Even Alibaba and sure China has lots of issues but they are not bombing anyone now if that's your first priority. Or else, smaller US / European / Asian companies with pure civilian focus. SOTA open weights models they serve are perfectly suitable for coding and chat. I run a local Qwen3.5-122B-A10B-NVFP4 instance and it writes entire Android apps from scratch and that's a midsized model.
Sorry for the off-topic but what hardware are you running Qwen3.5-122B-A10B-NVFP4 on? Is it physically local or just self-administered? Thanks in advance.
Can you give a list of high quality alternatives? Morally speaking i would put China on par with the US if not worse (due to their ongoing Uyghur genocide). I will check out Qwen3 but would be interested in others.
Only about 70% of Americans even own a laptop[1]. Factor in many of those being ancient with 15 minute battery life, plus user preferences… it’s hard to see how that could be the majority use case.
People have a right to ignore speech, and to establish standards for speech on their private property. If there is market demand for a service that filters out content based on ideology, whether mastodon.social or Fox News, so be it.
It can be toxic and a social negative, but any fix is worse than the problem.
reply