I suspect that this is the direct result of AI generated content just overwhelming any real content.

I tried ddg, google, bing, quant, and none of them really help me find information I want these days.

Perplexity seems to work but I don’t like the idea of AI giving me “facts” since they are mostly based on other AI posts

ETA: someone suggested SearXNG and after using it a bit it seems to be much better compared to ddg and the rest.

You are viewing a single thread.
View all comments
-17 points
*

I think it’s just you. Differential Transformers are pretty good at regurgitating information that’s widely talked about. They fall short when it comes to specific information on niche subjects, but generally that’s only a matter of understanding the jargon needed to plug into a search engine to find what you’re looking for. Paired with uBlock Origin, it’s all typically pretty straight forward, so long as you know which to use in which circumstance.

Almost always, I can plug some error for an OS into a LLM and get specific instructions on how to resolve it.

Additionally if you understand and learn how to use a model that can parse your own set of user-data, it’s easy to feed in documentation to make it subject-specific and get better results.

Honestly, I think the older generation who fail to embrace and learn how to use this tool will be left in the dust, as confused as the pensioners who don’t know how to write an email.

permalink
report
reply
13 points

Stable Diffusors are pretty good at regurgitating information that’s widely talked about.

Stable Diffusion is an image generator. You probably meant a language model.

And no, it’s not just OP. This shit has been going on for a while well before LLMs were deployed. Cue to the old “reddit” trick that some people used.

permalink
report
parent
reply
6 points

Also, they’re pretty good at regurgitating bullshit. Like the famous ‘glue on pizza’ answer.

permalink
report
parent
reply
2 points

Or, in a deeper aspect: they’re pretty good at regurgitating what we interpret as bullshit. They simply don’t care about truth value of the statements at all.

That’s part of the problem - you can’t prevent them from doing it, it’s like trying to drain the ocean with a small bucket. They shouldn’t be used as a direct source of info for anything that you won’t check afterwards; at least in kitnaht’s use case if the LLM is bullshitting it should be obvious, but go past that and you’ll have a hard time.

permalink
report
parent
reply
5 points

I’m not eating pizza at your house, that’s for sure.

permalink
report
parent
reply

Technology

!technology@lemmy.world

Create post

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


Community stats

  • 14K

    Monthly active users

  • 6.8K

    Posts

  • 158K

    Comments