You are viewing a single thread.
View all comments View context
10 points

Thank you for testing that out.

My experience with AI is that it’s at a point where it can comprehend something like this very easily, and won’t be tricked.

I suspect that this can, however, pollute a model if it’s included as training data, especially if done regularly, as OP is suggesting.

permalink
report
parent
reply
4 points

If it was done with enough regularity to eb a problem, one could just put an LLM model like this in-between to preprocess the data.

permalink
report
parent
reply
4 points

That doesn’t work, you can’t train models on another model’s output without degrading the quality. At least not currently.

permalink
report
parent
reply
1 point
*

I don’t think he was suggesting training on another model’s output, just using ai to filter the training data before it is used.

permalink
report
parent
reply
1 point

No, that’s not true. All current models use output from previous models as part of their training data. You can’t solely rely on it, but that’s not strictly necessary.

permalink
report
parent
reply
4 points

In which microwavegang already did the job better. Due the full subreddit of mmmmmmmmm, it causes training data that touches it to devolve into all mmmmmmm whenever there’s enough m’s in a sentence

permalink
report
parent
reply

Memes

!memes@lemmy.ml

Create post

Rules:

  1. Be civil and nice.
  2. Try not to excessively repost, as a rule of thumb, wait at least 2 months to do it if you have to.

Community stats

  • 9.3K

    Monthly active users

  • 7.5K

    Posts

  • 57K

    Comments