Reddit CEO Steve Huffman is standing by Reddit’s decision to block companies from scraping the site without an AI agreement.

Last week, 404 Media noticed that search engines that weren’t Google were no longer listing recent Reddit posts in results. This was because Reddit updated its Robots Exclusion Protocol (txt file) to block bots from scraping the site. The file reads: “Reddit believes in an open Internet, but not the misuse of public content.” Since the news broke, OpenAI announced SearchGPT, which can show recent Reddit results.

The change came a year after Reddit began its efforts to stop free scraping, which Huffman initially framed as an attempt to stop AI companies from making money off of Reddit content for free. This endeavor also led Reddit to begin charging for API access (the high pricing led to many third-party Reddit apps closing).

In an interview with The Verge today, Huffman stood by the changes that led to Google temporarily being the only search engine able to show recent discussions from Reddit. Reddit and Google signed an AI training deal in February said to be worth $60 million a year. It’s unclear how much Reddit’s OpenAI deal is worth.

Huffman said:

Without these agreements, we don’t have any say or knowledge of how our data is displayed and what it’s used for, which has put us in a position now of blocking folks who haven’t been willing to come to terms with how we’d like our data to be used or not used.

“[It’s been] a real pain in the ass to block these companies,” Huffman told The Verge.

422 points
*

Honestly, any platforms hosting user-generated content who use the legal argument that they only provide hosting and aren’t responsible for what their user post shouldn’t also be able to sell the same data and claim owning any of it.

Otherwise, take away their legal immunity. Nazis or pedophiles post something awful? You get in front of the judge.

edit: typo

permalink
report
reply
179 points

Exactly this. You can claim that their scraping is abusing your servers, but the moment you claim copyright for the content of the site, then you give up your Section 230 rights.

permalink
report
parent
reply
35 points

You’d also probably lose a whole lot more processing power trying to stop the crawlers vs just letting them have API access with some sort of limit to queries.

permalink
report
parent
reply
29 points
*

Eh, not really.

I block bot user agents to my Lemmy instance, and the overhead is pretty negligible for that (it’s all handled in my web firewall/load balancer).

Granted, those are bots that correctly identify themselves via user agent and don’t spoof a browser’s.

It’s also cheaper and easier to add another load balancer than size up or scale out my DB server to handle the bot traffic.

permalink
report
parent
reply
6 points

I don’t think they actually block malicious bots, the change they’ve made is just to the robots.txt, they don’t have to do anything.

permalink
report
parent
reply
38 points

Can’t sell something you don’t own.

So if they’re selling the parts people want, they need to own the parts no one wants.

permalink
report
parent
reply
3 points
*

Well, you can give money to Reddit for a piece of paper, but unless Reddit is claiming copyright to the content posted there, then they can’t sue anyone for not paying. It would be very interesting to see the text of these “licensing agreements”.

permalink
report
parent
reply
4 points

They’re not claiming copyright. They have a perpetual, non-revokable license to the content, granted by the people who use their site when they post the content.

permalink
report
parent
reply
4 points

Good point!

permalink
report
parent
reply
102 points
*

Robots.txt isn’t a binding agreement, this isn’t stopping anyone for whom their drive for profit outweighs their ethics.

Also, Fuck Spez.

permalink
report
reply
66 points

The enshittification cycle:

Phase one, attract users by providing a good service.
Phase two, once the users are locked in, squeeze them for all they’re worth by selling them to business customers (advertisers and/or data buyers).
Phase three, once the business customers are locked in, squeeze them for all they’re worth by threatening to deny them access to the users on whom they now depend.

Spez seems to think Reddit has the pull to make phase 3 happen. I rather doubt it, but we’ll see.

permalink
report
reply
28 points
*

https://catvalente.substack.com/p/stop-talking-to-each-other-and-start

Blog post (?) from Catherynne Valente about this exact topic

Stop benefitting from the internet, it’s not for you to enjoy, it’s for us to use to extract money from you. Stop finding beauty and connection in the world, loneliness is more profitable and easier to control.

permalink
report
parent
reply
14 points

If he really had balls he’d restrict access to the site and improve the built-in search engine.

If reddit’s own search worked well nobody would care. Engines like DDG even have bang codes that send you to a site’s own engine. So instead of having to add “site:reddit.com” to the search on DDG I’d just add “r!” and it would end up being the same thing. IF the internal search didn’t suck.

permalink
report
parent
reply
12 points

Spez is tracing Elon’s steps with X really closely.

Also fuck spez.

permalink
report
parent
reply
9 points
*

Yeah, as soon as the API thing happened I switched to Lemmy for mobile browsing and like it more than Reddit (Connect is pretty good, but even the mobile browser site is solid).

The more they squeeze, the more popular alternatives like Lemmy, Kbin/Mbin, Tildes, etc. will become.

permalink
report
parent
reply
6 points

My guess is that phase three will work for a while. But I think you’re right that eventually they are going to drive that thing into the ground. Because it’s never enough pure profit for rent-seeking scum, and there is no lower limit to the abuse they’ll inflict on their content creators (who they call users but think of as products).

permalink
report
parent
reply
57 points

Fuck Spez. He’s probably editing the comments anyway, he literally can’t help himself.

permalink
report
reply
46 points

Honestly, my biggest issue with LLMs is how they source their training data to create “their own” stuff. A meme calling it a plagiarism machine struck a chord with me. Almost anyone else I’d sympathize with, but fuck Spez.

permalink
report
reply
6 points

What resonated with me is people calling LLMs and Stable Diffusion “copyright laundering”. If copyright ever swung in AI’s favor it would be super easy to train an AI on stuff you want to steal, add in some generic training, and now you have a “new” piece of art.

LLMs and Stable Diffusion are just compression algorithms for abstract patterns, only one level above data.

permalink
report
parent
reply
6 points
*

The real takeaway of all of this is that copyright law is massively out of date and not fit for purpose in the 21st century or frankly the late 20th.

The current state of copyright law cannot deal with the internet, let alone AI

permalink
report
parent
reply
5 points

Yep they now get paid for the data we have them. I have no sympathy lol. At least these models can’t actually store it all losslessly by any stretch of the imagination. The compression factors would have to be like 100-200X+ anything we’ve ever been able to achieve before. The numbers don’t work out. The models do encode a lot though and some of it is going to include actual full text data etc but it’ll still be kinda fuzzy.

I think we do need ALL OPEN SOURCE. Not just for AI, but I know on that point I’m preaching to the choir here lol

permalink
report
parent
reply

Technology

!technology@lemmy.world

Create post

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


Community stats

  • 18K

    Monthly active users

  • 5.1K

    Posts

  • 91K

    Comments