Nice job! This is a fairly common trick with AI. In traditional programming, there’s a clear separation between code and data. That’s not the case for GenAI, so these kinds of hacks have worked all over the place.
I don’t want to have to make legal threats to an LLM in all data not intended for LLM consumption, especially since the LLM might just end up ignoring it anyway, since there is no defined behavior with them.
@bitofhope Absolutely agree, but this is where technology is evolving and we have to learn to adapt or not. Since it’s not going away, I’m not sure that not adapting is the best strategy.
And I say the above with full awareness that it’s a rubbish response.
have you ever run into the term “learned helplessness”? it may provide some interesting reading material for you
(just because samai and friends all pinky promise that this is totally 170% the future doesn’t actually mean they’re right. this is trivially argued too: their shit has consistently failed to deliver on promises for years, and has demonstrated no viable path to reaching that delivery. thus: their promises are as worthless as the flashy demos)