Wondering if Modern LLMs like GPT4, Claude Sonnet and llama 3 are closer to human intelligence or next word predictor. Also not sure if this graph is right way to visualize it.

68 points

That’s literally how llma work, they quite literally are just next word predictors. There is zero intelligence to them.

It’s literally a while token is not “stop”, predict next token.

It’s just that they are pretty good at predicting the next token so it feels like intelligence.

So on your graph, it would be a vertical line at 0.

permalink
report
reply
11 points

What is intelligence though? Maybe I’m getting through life just by being pretty good at predicting what to say or do next…

permalink
report
parent
reply
10 points

yeah yeah I’ve heard this argument before. “What is learning if not like training.” I’m not going to define it here. It doesn’t “think”. It doesn’t have nuance. It is simply a prediction engine. A very good prediction engine, but that’s all it is. I spent several months of unemployment teaching myself the ins and outs, developing against llms, training a few of my own. I’m very aware that it is not intelligence. It is a very clever trick it pulls off, and easy to fool people that it is intelligence - but it’s not.

permalink
report
parent
reply
1 point

But how do you know that the human brain is not just a super sophisticated next-thing predictor that by being super sophisticated manages to incorporate nuance and all that stuff to actually be intelligent? Not saying it is but still.

permalink
report
parent
reply
5 points

Agreed

permalink
report
parent
reply
-4 points
*

This is true if you describe a pure llm, like gpt3

However systems like claude, gpt4o and 1o are far from just a single llm, they are a blend of tailored llms, machine learning some old fashioned code to weave it all together.

Op does ask “modern llm” so technically you are right but i believed they did mean the more advanced “products”

Though i would not be able to actually answer ops questions, ai is hard to directly compare with a human.

In most ways its embarrassingly stupid, in other it has already surpassed us.

permalink
report
parent
reply
24 points

That is just next word prediction with extra steps.

permalink
report
parent
reply
2 points

Now that is fair.

permalink
report
parent
reply
11 points

None of which are intelligence, and all of which are catered towards predicting the next token.

All the models have a total reliance on data and structure for inference and prediction. They appear intelligent but they are not.

permalink
report
parent
reply
-4 points
*

How is good old fashioned code comparing outputs to a database of factual knowledge “predicting the next token” to you. Or reinforcement relearning and token rewards baked into models.

I can tell you have not actually tried to work with professional ai or looked at the research papers.

Yes none of it is “intelligent” but i would counter that with neither are human beings, we dont even know how to define intelligence.

permalink
report
parent
reply

No, unfortunately you are wrong.

Gpt4 is a better version of gpt3.

The brand new one that is allegedly “unhackable” just has a role hierarchy providing rules and that hasn’t been fulled tested in the wild yet.

permalink
report
parent
reply
1 point
*

First, did you read even the research papers?

Secondly, none are out that are actually immune to jailbreaking lol, Where did that claim come from?

Gpt4 is just an llm. Indeed the better version of gpt3

Gpt4o and 1o (claude-sonnet possibly also) rely on the generative capacities of the gpt4 model but there is allot more going under the hood that is not simply “generate the next token”

We all agree that a pure text predictor are not at all intelligent.

The discussion at hand is wether the current frontier of ai has moved the needle up. And i still would call it pretty dumb, but moving that needle, it did. Somewhere around (x2y0.5) if i have to use the meme. Stating its (0,0) just means people aren’t interested enough to pay attention, that these aren’t just llm anymore. That’s their right but i prefer people stopped joining the discussion so uninformed.

permalink
report
parent
reply
41 points

They’re still word predictors. That is literally how the technology works

permalink
report
reply
6 points

Yeah, the only question is whether human brains are also just that.

permalink
report
parent
reply
0 points

no, they are not. try showing an ai a huge number of pictures of cars from the front. Then show them one car from the side, and ask them what it is.

Show a human one picture of a car from the front, then the one from the side and ask them what it is.

permalink
report
parent
reply
2 points

What if the human had never seen or heard of anything similar to cars?

I bet it’d be confused as much as the llm.

permalink
report
parent
reply
1 point

lol, you got me, i definitely hadn’t thought of that.

permalink
report
parent
reply
41 points

Intelligence is a measure of reasoning ability. LLMs do not reason at all, and therefore cannot be categorized in terms of intelligence at all.

LLMs have been engineered such that they can generally produce content that bears a resemblance to products of reason, but the process by which that’s accomplished is a purely statistical one with zero awareness of the ideas communicated by the words they generate and therefore is not and cannot be reason. Reason is and will remain impossible at least until an AI possesses an understanding of the ideas represented by the words it generates.

permalink
report
reply
25 points
*

There’s a preprint paper out that claims to prove that the technology used in LLMs will never be able to be extended to AGI, due to the exponentially increasing demand for resources they’d require. I don’t know enough formal CS to evaluate their methods, but to the extent I understand their argument, it is compelling.

permalink
report
reply
21 points

i think the first question to ask of this graph is, if “human intelligence” is 10, what is 9? how you even begin to approach the problem of reducing the concept of intelligence to a one-dimensional line?

the same applies to the y-axis here. how is something “more” or “less” of a word predictor? LLMs are word predictors. that is their entire point. so are markov chains. are LLMs better word predictors than markov chains? yes, undoubtedly. are they more of a word predictor? um…


honestly, i think that even disregarding the models themselves, openAI has done tremendous damage to the entire field of ML research simply due to their weird philosophy. the e/acc stuff makes them look like a cult, but it matches with the normie understanding of what AI is “supposed” to be and so it makes it really hard to talk about the actual capabilities of ML systems. i prefer to use the term “applied statistics” when giving intros to AI now because the mind-well is already well and truly poisoned.

permalink
report
reply
3 points

what is 9?

exactly! trying to plot this is in 2D is hella confusing.

plus the y-axis doesn’t really make sense to me. are we only comparing humans and LLMs? where do turtles lie on this scale? what about parrots?

the e/acc stuff makes them look like a cult

unsure what that acronym means. in what sense are they like a cult?

permalink
report
parent
reply
3 points
*

Effective Accelerationism. an AI-focused offshoot from the already culty effective altruism movement.

basically, it works from the assumption that AGI is real, inevitable, and will save the world, and argues that any action that slows the progress towards AGI is deeply immoral as it prolongs human suffering. this is the leading philosophy at openai.

their main philosophical sparring partners are not, as you might think, people who disagree on the existence or usefulness of AGI. instead, they take on the other big philosophy at openai, the old-school effective altruists, or “ai doomers”. these people believe that AGI is real, inevitable, and will save the world, but only if we’re nice to it. they believe that any action that slows the progress toward AGI is deeply immoral because when the AGI comes online it will see that we were slow and therefore kill us all because we prolonged human suffering.

permalink
report
parent
reply
4 points

That just seems like someone read about Roko’s basilisk and decided to rebrand that nightmare as the mission/vision of a company.

What a time to be alive!

permalink
report
parent
reply

Asklemmy

!asklemmy@lemmy.ml

Create post

A loosely moderated place to ask open-ended questions

Search asklemmy 🔍

If your post meets the following criteria, it’s welcome here!

  1. Open-ended question
  2. Not offensive: at this point, we do not have the bandwidth to moderate overtly political discussions. Assume best intent and be excellent to each other.
  3. Not regarding using or support for Lemmy: context, see the list of support communities and tools for finding communities below
  4. Not ad nauseam inducing: please make sure it is a question that would be new to most members
  5. An actual topic of discussion

Looking for support?

Looking for a community?

Icon by @Double_A@discuss.tchncs.de

Community stats

  • 9.2K

    Monthly active users

  • 3K

    Posts

  • 51K

    Comments