I’ve said this before, but we also need to be cautious about this on lemmy and devise ways to empower mods and the community to fight back against this, I’m not entirely sure how since it’s a very complex problem
It’s bloody difficult.
I used to mod on /r/videos years and years back. We had this one guy who was not very active as a mod in the day to day stuff, but was respected because he’d basically disappear for a few months and then reappear with a huge post in our modding sub basically going “so these are all spammers/malicious actors, here’s their profiles, the accounts were created in these waves, here’s where they’ve copied existing posts / the identical generic comments and things they use to get around our posting requirements, the targets they’ve been promoting, etc”. Just huge pages of thoroughly researched proof.
This was well before we had huge awareness of situations like Russia manipulating social media - it was usually those viral video places that buy up rights to videos and handle licensing and promotion. It’s why for a long time any licensed videos from places like viralhog etc were outright banned - they were constantly trying to manipulate reddit postings in bad faith, and even trying to socially engineer the mod team in modmail, so any videos that mentioned a licensing deal in the description were automatically banned from posting.
If we didn’t have that one guy spotting the patterns, most of it would have gotten by easily. Unfortunately he did eventually disappear for good. No clue what happened to him, hope he just cut out social media or something. But with the spamming and astroturfing stuff… Even after fighting it for years I can’t tell you what to do to counter it besides “have more of that guy”.
I am convinced this is already happening. One example is the endless new accounts posting ibtimes links.
There are also propoganda websites posted regularly by new accounts (especially sowing disinformation about Russia’s war on Ukraine).
Basically be wary of anything posted where it’s their first post. Often they make accounts and don’t use them for months so they look older.
I also think astroturfing is happening but at lower rate than reddit.
Like you, I have no idea how we can counter this at scale.
It might help if a poster’s number of posts and signup date were listed at the top of each post or comment. Would’t be a fix but might help weed out upsprouting autotrolls.
There are a lot of subreddits which routinely award hundreds or thousands of upvotes for repetitive low value posts. … This is a cog in the well-tuned machine of new-accounts being created and matured to look ‘real’ for when they are later used for advertising / manipulation later down the line.
In the early months of a new account, it is easier to spot. Eg. If you see a post on a game subreddit with a title like “Exciting to try this game, any tips get started?”, you might click the profile and see that their entire history is a bunch of low-effort discussion starters. “Name a band from the 80s that everyone has forgotten”; “What’s the most misunderstood concept in maths?”; “What’s the most underrated (movie / band / drug / car / tourist attraction / whatever suits the topic of the subreddit)?”
A heap of threads like that, on a new account with a very generic name (adjective-noun-numbers is a common pattern); posting on a variety of subredits… is highly suspicious. But it gets harder to recognise as the account gets older and has a longer history - at which point it is ready to be sold / used for its next purpose.