He generally shows most of the signs of the misinformation accounts:
- Wants to repeatedly tell basically the same narrative and nothing else
- Narrative is fundamentally false
- Not interested in any kind of conversation or in learning that what he’s posting is backwards from the values he claims to profess
I also suspect that it’s not a coincidence that this is happening just as the Elon Musks of the world are ramping up attacks on Wikipedia, specially because it is a force for truth in the world that’s less corruptible than a lot of the others, and tends to fight back legally if someone tries to interfere with the free speech or safety of its editors.
Anyway, YSK. I reported him as misinformation, but who knows if that will lead to any result.
Edit: Number of people real salty that I’m talking about this: Lots
You just described the average Tankie around here lmao
Yeah, there’s kind of a Poe’s Law situation.
A lot of the sincere tankies, though, at least want to talk about what they’re into, and have elaborate reasons why it’s all true. The low-effort “I can’t even be bothered to try to mount a defense, I just wanted to say Wikipedia is doxing its users and kowtowing to fascist governments, and now that I’ve said it my task is done” behavior is a little more indicative of a disingenuous propaganda account in my experience.
elaborate reasons why it’s all true
Usually it’s “just read these 10 hundred-year-old books” that they absolutely have not read.
And if you ask them to make a point from those books, they can’t. Apparently they’re only comprehensible as a whole.
That’s now poe’s law, it would be Occam’s razor.
The most likely scenario here is not many puppet accounts spreading sarcasm or parody but rather that there are many actors that all true believers in what they are all saying. They sound the same because they are feeding off the same talking point.
You’re right, I was misremembering Poe’s law. We need a law for “there is no point of view so idiotic that someone won’t be out there passionately proclaiming it, not because they are a propaganda troll, but because they really believe it.”
I am pretty convinced that .ml is legitimately used as a Russian troll training ground before they get promoted to Facebook and reddit.
Meanwhile, at .ml:
Since Pi is infinite and non-repeating, would that mean any finite sequence of non-repeating numbers should appear somewhere in Pi?
That’s actually a really good way to illustrate what is wrong with lemmy.ml.
On math stack exchange:
Let me summarize the things that have been said which are true and add one more thing.
- 𝜋 is not known to have this property, but it is expected to be true.
- This property does not follow from the fact that the decimal expansion of 𝜋 is infinite and does not repeat.
On lemmy.ml:
0.101001000100001000001 . . .
I’m infinite and non-repeating. Can you find a 2 in me?
You can’t prove that there isn’t one somewhere
Why couldn’t you?
Because you’d need to search through an infinite number of digits (unless you have access to the original formula)
And:
Not just any all finite number sequence appear in pi
And:
Yes.
And if you’re thinking of a compression algorithm, nope, pigeonhole principle.
All heavily upvoted.
Even the most extreme extremist of echo chambers will have benign random conversations. Singling out a random blurb of conversation, without even any source link, is just cherry picking.
Could you be anymore vague
Yeah, I was trying to talk about the situation without specifically linking to the comments or starting any kind of brigade situation. I figured being vague was better than being inflammatory, and anyone who cared enough would know what I was taking about, which seems to be accurate.
be anymore vague
-
be any more vague
-
be vague anymore
Don’t split the lanes, man.
I seriously don’t understand this kind of reaction to being corrected about mistakes. All it does is show everyone that you’re even dumber than previously shown via the mistake and makes it obvious you don’t care to learn anything from it so you’ll continue being just as dumb.
Learn from it instead and thank them for teaching you something.
But you’re clearly not mentally grown and/or smart enough to not react like a 5-year-old to someone pointing out you’ve made a mistake. And it’s hilarious that you don’t realize how childish and dumb you make it clear you are by this kind of reaction and choosing to show it with a reply.
It’s Musk, he’s spreading the disinformation about how Wikipedia leaks your data
My source is it’s a joke on how Musk is a misinformation merchant that profits from the attention of greedy and gullible chuds, or the outrage of people that feel disenfranchised by the billionaire class.
Interesting all this WP news I’m hearing today. Last week I downloaded the entirety of Wikipedia. Anyone can do it, the base archive (no pictures) is only about 25G, although the torrent is slow AF, took me… almost 2 weeks to download it.
I did this because I feel like this might be the last chance to get a version of it that has any vestige of the old order in it, the old order being “trying to stick to ideals and express truth rather than rewriting history to the fascists’ specifications.”
I’d love to be wrong, but if I’m not, I feel like it will potentially be a good reference in the future if needed.
This is in the news because Wikipedia is refusing to rewrite history to the fascists’ specifications.
https://www.bbc.com/news/articles/cdrdydkypv7o
It’s possible that India will succeed at eroding by a little bit Wikipedia’s resistance to having things rewritten because of various powerful people demanding it. But, if you’re looking for an organization that’s resistant against those demands, I don’t think you will be able to find one that is anywhere near the equal of Wikipedia in terms of the scale at which it operates combined with the resistance it puts up when people do this.
I donate every year and they made it easier than ever this year if you use Apple Pay or anything equivalent. Like 15 seconds and that includes chhosing amount.
edit: for us with the lazys
That’s interesting and terrifying all at once. If the Indian government is successful, it will basically set the precedent for other powerful entities such as autocrats, oligarchs, and corporations to also force Wikipedia to edit their content to suit their desires. I donate frequently and will keep making sure they can win.
Wow, they really sued the Wikimedia Foundation instead of trying to find a reliable source to refute the article’s claims. I looked up the edits they made. They removed content, citing various Wikipedia policies that govern how the article should be phrased.
In general, so long as the information is presented in a neutral, matter-of-fact manner and cites a reliable source, it can go in the article. Wikipedia’s job is to summarize what reliable sources say about a subject.
So all ANI would’ve needed to do was find a reliable source (preferably more than one) refuting the claims they want to refute. The most they’d likely be able to do is put both points of view in the article rather than removing one point of view entirely from the article, which is what they were trying to do.
Instead, they went to court about it.
Kiwix is a self hostable option for this, and you can get other content databases as well, like wikiHow, iFixit, and Khan Academy.
The downloads are much faster than two weeks too.
Just some context, Hetzner gave the shaft to the Kiwix project and took down their content servers without any apparent notice (Kiwix’s side of the story at least), and they had to rebuild it with another provider.
There are major issues with wikipedia, I say this as someone with thousands of edits. But I know exactly who you are talking about and they spread pure BS.
The last time I saw them their account was called “ihatewikipedia” or “fuckwikipedia” or something like that lol and they were just spreading conspiracies. Or useless drama. Like they were going on about how wikipedia “invades your privacy”, it IP blocks people and tracks IP’s linked to editing.
it IP blocks people and tracks IP’s linked to editing
Unless something changed, this part was at least partially true at one point. But only for anonymous edits iirc. Usually happened for IPs shared by a lot of people like from a campus or some VPNs, probably due to a lot of vandalism from such IPs.
It’s likely this is a bot if it’s wide spread. And Lemmy is INCREDIBLY ill suited to handle even the dumbest of bots from 10+ years ago. Nevermind social media bots today.
Ur a bot. I can tell by the pixels unicode.
Edit: joking aside you bring up a good point and our security through anonymity cultural irrelevance will not last forever. Or maybe it will.
Unfortunately it won’t, assuming Lemmy grows.
Lemmy doesn’t get targeted by bots because it’s obscure, you don’t reach much of an audience and you don’t change many opinions.
It has, conservatively, ~0.005% (Yes, 0.005%, not a typo) of the monthly active users.
To put that into perspective, theoretically, $1 spent on a Reddit has 2,000,000x more return on investment than on Lemmy.
All that needs to happen is that number to become more favorable.
To be fair, it’s virtually impossible to tell whether a text was written by an AI or not. If some motivated actor is willing to spend money to generate quality LLM output, they can post as much as they want on virtually all social media sites.
The internet is in the process of eating itself as we speak.
You don’t analyze the text necessary, you analyze the heuristics, behavioral patterns, sentiment…etc It’s data analysis and signal processing.
You, as a user, probably can’t. Because you lack information that the platform itself is in a position to gather and aggregate that data.
There’s a science to it, and it’s not perfect. Some companies keep their solutions guarded because of the time and money required to mature their systems & ML models to identify artificial behavior.
But it requires mature tooling at the very least, and Lemmy has essentially none of that.
yes of course there are many different data points you can use. along with complex math you can also feed a lot of these data points in machine learning models and get useful systems that can perhaps red flag certain accounts and then have processes with more scrutiny that require more resources (such as a human reviewing)
websites like chess.com do similar things to find cheaters. and they (along with lichess) have put out some interesting material going over some of what their process looks like
here i have two things. one is that lichess, which is mostly developed and maintained by a single individual, is able to maintain an effective anti-cheat system. so I don’t think it’s impossible that lemmy is able to accomplish these types of heuristics and behavioral tracking
the second thing is that these new AIs are really good. it’s not just the text, but the items you mentioned. for example I train a machine learning model and then a separate LLM on all of reddit’s history. the first model is meant to try and emulate all of the “normal” human flags. make it so it posts at hours that would match the trends. vary the sentiments in a natural way. etc. post at not random intervals of time but intervals of time that looks like a natural distribution, etc. the model will find patterns that we can’t imagine and use those to blend in
so you not only spread the content you want (whether it’s subtle product promotion or nation-state propaganda) but you have a separate model trained to disguise that text as something real
that’s the issue it’s not just the text but if you really want to do this right (and people with $$$ have that incentive) as of right now it’s virtually impossible to prevent a motivated actor from doing this. and we are starting to see this with lichess and chess.com.
the next generation of cheaters aren’t just using chess engines like Stockfish, but AIs trained to play like humans. it’s becoming increasingly difficult.
the only reason it hasn’t completely taken over the platform is because it’s expensive. you need a lot of computing power to do this effectively. and most people don’t have the resources or the technical ability to make this happen.
spend money to generate quality LLM output, they can post as much as they want on virtually all social media sites.
$20 for a chatgpt pro account and fractions of pennies to run a bot server. It’s really extremely cheap to do this.
I don’t have an answer to how to solve the “motivated actor” beyond mass tagging/community effort.
$20 for a chatgpt pro account and fractions of pennies to run a bot server. It’s really extremely cheap to do this.
openAI has checks for this type of thing. They limit number of requests per hour with the regular $20 subscription
you’d have to use the API and that comes at a cost per request, depending on which model you are using. it can get expensive very quickly depending on what scale of bot manipulation you are going for
Heuristics, data analysis, signal processing, ML models…etc
It’s about identifying artificial behavior not identifying artificial text, we can’t really identify artificial text, but behavioral patterns are a higher bar for botters to get over.
The community isn’t in a position to do anything about it the platform itself is the only one in a position to gather the necessary data to even start targeting the problem.
I can’t target the problem without first collecting the data and aggregating it. And Lemmy doesn’t do much to enable that currently.
But something like Reddit at least potentially has the resources to throw some money at the problem. They can employ advanced firewalls and other anti-bot/anti-AI thingies. It’s very possible that they’re pioneering some state-of-the-art stuff in that area.
Lemmy is a few commies and their pals. Unless China is bankrolling them, they’re out of their league.