I recently made the jump from Reddit for the same immediate reasons as everyone else. But, to be honest, if it was just the Reddit API cost changes I wouldn’t be looking to jump ship. I would just weather the protest and stay off Reddit for a few days. Heck I’d probably be fine paying a few bucks a month if it helped my favorite Reddit app (Joey) stay up and running.
No, the real reason I am taking this opportunity to completely switch platforms is because for a couple years now Reddit has been unbearably swamped by bots. Bot comments are common and bot up/downvotes are so rampant that it’s becoming impossible to judge the genuine community interest in any post or comment. It’s just Reddit (and maybe some other nefarious interests) manufacturing trends and pushing the content of their choice.
So, what does Lemmy do differently? Is there anything in Lemmy code or rules that is designed to prevent this from happening here?
deleted by creator
That’s disappointing. Screening new accounts only forces spammers to create the accounts with a human touch and then turn it over to their AI. What about a system to prevent bots from up/downvoting? Something like websites use to detect bots. Just by clicking in the little box that says “I am not a robot” the website can tell you’re not a bot. What if every single up and down arrow was formulated like that little box?
That’s the thing though - what system? Reddit, YouTube, Twitter, Facebook, you name it, nobody managed to prevent bots. How would Lemmy be more successful at this? It’s an extremely challenging battle, unfortunately.
Do those for-profit social media companies want to drive-down traffic that makes them seem more valuable to advertisers? I get that it’s still insanely difficult, and we can’t actually implement a captcha on every up-vote, but it seems like there’s a conflict of interest between moderators and site owners when it comes to bot activity.
Arguably, some of the platforms I mentioned have even more of an interest on preventing bots. If I want to place ads on your website, but you can’t tell me if out of 100 impressions 10 are bots or 90 are bots… I’m not wasting my money, or at the very least, I’ll expect rates significantly lower than other competitors.
I don’t know. Wouldn’t their motivation be to know exactly how many bots there are (so they could disclose the number if/when asked) but continue to let them proliferate?
Social media companies generally benefit from high traffic for advertiser appeal, but combating bots is crucial for maintaining user trust and engagement. Implementing CAPTCHAs for every upvote may not be feasible, but addressing bot activity is generally in the long-term interest of social media companies.
This message was generated by ChatGPT.
Not sure if you bought that, but if I was applying for an account on Beehaw using a LLM assistant, I bet the odds of passing a human review is better than 50%.
Oh god. Could you imagine doing a captcha every time you upvoted? Please DO NOT do this, Ernest.