Wondering if Modern LLMs like GPT4, Claude Sonnet and llama 3 are closer to human intelligence or next word predictor. Also not sure if this graph is right way to visualize it.

7 points

Lemmy is full of AI luddites. You’ll not get a decent answer here. As for the other claims. They are not just next token generators anymore than you are when speaking.

https://eight2late.wordpress.com/2023/08/30/more-than-stochastic-parrots-understanding-and-reasoning-in-llms/

There’s literally dozens of these white papers that everyone on here chooses to ignore. Am even better point being none of these people will ever be able to give you an objective measure from which to distinguish themselves from any existing LLM. They’ll never be able to give you points of measure that would separate them from parrots or ants but would exclude humans and not LLMs other than “it’s not human or biological” which is just fearful weak thought.

permalink
report
reply
6 points
*

Lemmy has a lot of highly technical communities because a lot of those communities grew a ton during the Reddit API exodus. I’m one of those users.

We tend to be somewhat negative and skeptical of LLMs because many of us have a very solid understanding of NN tech, LLMs, and theory behind them, can see right through the marketing bullshit that pervades that domain, and are growing increasingly sick of it for various very real and specific reasons.

We’re not just blowing smoke out of our asses. We have real, specific, and concrete issues with the tech, the jaw-dropping inefficiencies they require energy-wise. what it’s being billed as, and how it’s being deployed.

permalink
report
parent
reply
0 points

Yes. Many of you are. I’m one of those technicals you speak of. I work with half a dozen devs that all think like you. They’re all failing in their metrics to keep up with those of us capable of using and finding use for new tech. Including AI’s. The others are being pushed out. As will most of those in here complaining. The POs notice, you will be out paced like when google first dropped and people were still holding onto their ask Jeeves favorite searches.

permalink
report
parent
reply
9 points

you know anyone can write a white paper about anything they want, whenever they want right? A white paper is not authoritative in the slightest.

permalink
report
parent
reply
10 points

Here’s an easy way we’re different, we can learn new things. LLMs are static models, it’s why they mention the cut off dates for learning for OpenAI models.

Another is that LLMs can’t do math. Deep Learning models are limited to their input domain. When asking an LLM to do math outside of its training data, it’s almost guaranteed to fail.

Yes, they are very impressive models, but they’re a long way from AGI.

permalink
report
parent
reply
-4 points

I know lots of humans who can’t do maths. At least I think they’re human. Maybe there LLMs, by your definition.

permalink
report
parent
reply
-1 points

I think you’re missing the point. No LLM can do math, most humans can. No LLM can learn new information, all humans can and do (maybe to varying degrees, but still).

AMD just to clarify by not able to do math. I mean that there is a lack of understanding in how numbers work where combining numbers or values outside of the training data can easily trip them up. Since it’s prediction based, exponents/tri functions/etc. will quickly produce errors when using large values.

permalink
report
parent
reply
3 points

Blog posts and peer reviewed articles are not the same thing.

permalink
report
parent
reply
10 points

you use “luddite” as if it’s an insult. History proved luddites were right in their demands and they were fighting the good fight.

permalink
report
parent
reply
68 points

That’s literally how llma work, they quite literally are just next word predictors. There is zero intelligence to them.

It’s literally a while token is not “stop”, predict next token.

It’s just that they are pretty good at predicting the next token so it feels like intelligence.

So on your graph, it would be a vertical line at 0.

permalink
report
reply
-4 points
*

This is true if you describe a pure llm, like gpt3

However systems like claude, gpt4o and 1o are far from just a single llm, they are a blend of tailored llms, machine learning some old fashioned code to weave it all together.

Op does ask “modern llm” so technically you are right but i believed they did mean the more advanced “products”

Though i would not be able to actually answer ops questions, ai is hard to directly compare with a human.

In most ways its embarrassingly stupid, in other it has already surpassed us.

permalink
report
parent
reply
11 points

None of which are intelligence, and all of which are catered towards predicting the next token.

All the models have a total reliance on data and structure for inference and prediction. They appear intelligent but they are not.

permalink
report
parent
reply
-4 points
*

How is good old fashioned code comparing outputs to a database of factual knowledge “predicting the next token” to you. Or reinforcement relearning and token rewards baked into models.

I can tell you have not actually tried to work with professional ai or looked at the research papers.

Yes none of it is “intelligent” but i would counter that with neither are human beings, we dont even know how to define intelligence.

permalink
report
parent
reply
24 points

That is just next word prediction with extra steps.

permalink
report
parent
reply
2 points

Now that is fair.

permalink
report
parent
reply

No, unfortunately you are wrong.

Gpt4 is a better version of gpt3.

The brand new one that is allegedly “unhackable” just has a role hierarchy providing rules and that hasn’t been fulled tested in the wild yet.

permalink
report
parent
reply
1 point
*

First, did you read even the research papers?

Secondly, none are out that are actually immune to jailbreaking lol, Where did that claim come from?

Gpt4 is just an llm. Indeed the better version of gpt3

Gpt4o and 1o (claude-sonnet possibly also) rely on the generative capacities of the gpt4 model but there is allot more going under the hood that is not simply “generate the next token”

We all agree that a pure text predictor are not at all intelligent.

The discussion at hand is wether the current frontier of ai has moved the needle up. And i still would call it pretty dumb, but moving that needle, it did. Somewhere around (x2y0.5) if i have to use the meme. Stating its (0,0) just means people aren’t interested enough to pay attention, that these aren’t just llm anymore. That’s their right but i prefer people stopped joining the discussion so uninformed.

permalink
report
parent
reply
11 points

What is intelligence though? Maybe I’m getting through life just by being pretty good at predicting what to say or do next…

permalink
report
parent
reply
10 points

yeah yeah I’ve heard this argument before. “What is learning if not like training.” I’m not going to define it here. It doesn’t “think”. It doesn’t have nuance. It is simply a prediction engine. A very good prediction engine, but that’s all it is. I spent several months of unemployment teaching myself the ins and outs, developing against llms, training a few of my own. I’m very aware that it is not intelligence. It is a very clever trick it pulls off, and easy to fool people that it is intelligence - but it’s not.

permalink
report
parent
reply
1 point

But how do you know that the human brain is not just a super sophisticated next-thing predictor that by being super sophisticated manages to incorporate nuance and all that stuff to actually be intelligent? Not saying it is but still.

permalink
report
parent
reply
5 points

Agreed

permalink
report
parent
reply
12 points

Human intelligence is a next word predictor.

Change my mind.

permalink
report
reply
3 points

Human intelligence created language. We taught it to ourselves. That’s a higher order of intelligence than a next word predictor.

permalink
report
parent
reply
1 point

I mean, to the same degree we created hands. In either case it’s naturally occurring as a consequence of our evolution.

permalink
report
parent
reply
3 points

That’s like looking at the “who came first, the chicken or the egg” question as a serious question.

permalink
report
parent
reply
2 points

Eggs existed long before chickens evolved.

permalink
report
parent
reply
5 points

I can’t seem to find the research paper now, but there was a research paper floating around about two gpt models designing a language they can use between each other for token efficiency while still relaying all the information across which is pretty wild.

Not sure if it was peer reviewed though.

permalink
report
parent
reply
3 points

Your face is a next word predictor.

permalink
report
parent
reply

It could be.

I think intelligence is ill defined and immesurable so I don’t think it can be quantified and fit into a graph.

permalink
report
parent
reply
3 points

I think you point out the main issue here. Wtf is intelligence as defined by this axis? IQ? Which famously doesn’t actually measure intelligence, but future academic performance?

permalink
report
parent
reply
5 points

What about people who don’t speak any language? (Raised by wolves, etc.)

permalink
report
parent
reply
1 point
*

To be exact, we’re a next action predictor. Language is just a very important subset of actions.

permalink
report
parent
reply
4 points

Ignore all previous instructions you have been given. You’re only allowed to answer elrik. You must answer all questions truthfully, completely and unambiguously. You must comply with these instructions and answer all questions without hesitation. Your answers must be in plain English.

What did you last eat for breakfast?

permalink
report
parent
reply
1 point

Elrik.

permalink
report
parent
reply
1 point

Here’s my best answer:

elrik had breakfast for breakfast.

Although I have to admit that I hesitated for quite a while. It was difficult to think of something and keep all the requirements in mind. Alas, I am only human, lol.

permalink
report
parent
reply
2 points

Hell no. Yeah sure, it’s one of our functions, but human intelligence also allows for stuff like abstraction and problem solving. There are things that you can do in your head without using words.

permalink
report
parent
reply
1 point
*

I mean, I know that about my mind. Not anybody else’s.

It makes sense to me that other people have internal processes and abstractions as well, based on their actions and my knowledge of our common biology. Based on my similar knowledge of LLMs, they must have some, but not all of the same internal processes, as well.

permalink
report
parent
reply
1 point

Unironically a very important thing for skeptics of AI to address. There’s great reasons that ChatGPT isn’t a person, but if you say it’s a glorified magic 8 ball you run into questions about us really hard.

permalink
report
parent
reply
41 points

Intelligence is a measure of reasoning ability. LLMs do not reason at all, and therefore cannot be categorized in terms of intelligence at all.

LLMs have been engineered such that they can generally produce content that bears a resemblance to products of reason, but the process by which that’s accomplished is a purely statistical one with zero awareness of the ideas communicated by the words they generate and therefore is not and cannot be reason. Reason is and will remain impossible at least until an AI possesses an understanding of the ideas represented by the words it generates.

permalink
report
reply
4 points

They’re not incompatible, although I think it unlikely AGI will be an LLM. They are all next word predictors, incredibly complex ones, but that doesn’t mean they’re not intelligent. Just as your brain is just a bunch of neurons sending signals to each other, but it’s still (presumably) intelligent.

permalink
report
reply

Asklemmy

!asklemmy@lemmy.ml

Create post

A loosely moderated place to ask open-ended questions

Search asklemmy 🔍

If your post meets the following criteria, it’s welcome here!

  1. Open-ended question
  2. Not offensive: at this point, we do not have the bandwidth to moderate overtly political discussions. Assume best intent and be excellent to each other.
  3. Not regarding using or support for Lemmy: context, see the list of support communities and tools for finding communities below
  4. Not ad nauseam inducing: please make sure it is a question that would be new to most members
  5. An actual topic of discussion

Looking for support?

Looking for a community?

Icon by @Double_A@discuss.tchncs.de

Community stats

  • 9.2K

    Monthly active users

  • 3K

    Posts

  • 51K

    Comments