You are viewing a single thread.
View all comments View context
3 points
*

A better mathematical system of storing words does not mean the LLM understands any of them. It just has a model that represents the relation between words that it uses.

If I put 10 minus 8 into my calculator I get 2. The calculator doesn’t actually understand what 2 means, or what subtracting represents, it just runs the commands that gives the appropriate output.

permalink
report
parent
reply
0 points
*

That’s a bad analogy, because the calculator wasn’t trained using an artificial neural network literally designed by studying biological brains (aka biological neutral networks).

And “understand” doesn’t equate to consciousness or sapience. For example, it is entirely and factually correct to state that an LLM is capable of reasoning. That’s not even up for debate. The accuracy of an LLM’s reasoning capability is one of the fundamental benchmarks used for evaluating its quality.

But that doesn’t mean it’s “thinking” in the way most people consider.

Edit: anyone up voting this CileTheSane clown is in the same boat of not comprehending how LLMs work.

permalink
report
parent
reply
2 points

it is entirely and factually correct to state that an LLM is capable of reasoning

Citation needed.

If you’re going to tell me LLMs are modeled after biological brains and capable of reasoning then I call bullshit on your claims that you actually work in AI.

Imagine you put a man in an enclosed room. There is a slot in the wall where messages get passed through written in Chinese. The man does not speak Chinese or even recognize the written language, he just thinks they’re weird symbols.
First the man is shown examples of sequences of symbols to train him. Then he is shown incomplete sequences and asked which symbol comes next. If incorrect he is corrected, if correct he gets cookie. Eventually this man is able to carry on “conversations” with people in Chinese through continued practice.
This man still does not speak Chinese, he is not having reasoned, rational arguments with the people he is conversing with, and if you told him it was a language he’s look at you like your crazy. “There’s no language here, just if I have these symbols and I next put the one that looks like a man wearing a hat they give me a cookie.”

Thinking LLMs are capable of reasoning is the digital equivalent of putting eyes on a pencil then feeling bad when it gets broken in half.

permalink
report
parent
reply
0 points

Citation needed.

Certainly!

In machine learning, a neural network (also artificial neural network or neural net, abbreviated ANN or NN) is a model inspired by the structure and function of biological neural networks in animal brains

Source

A neural network is a method in artificial intelligence that teaches computers to process data in a way that is inspired by the human brain.

Source

A neural network is a machine learning program, or model, that makes decisions in a manner similar to the human brain

Source

*A neural network, or artificial neural network, is a type of computing architecture that is based on a model of how a human brain functions *

Source

Would you like some more citations?

Thinking LLMs are capable of reasoning is the digital equivalent of putting eyes on a pencil then feeling bad when it gets broken in half.

In this paper, we present Reasoning via Planning (RAP), a novel LLM reasoning framework that equips LLMs with an ability to reason akin to human-like strategic planning

Source - Reasoning with Language Model is Planning with World Model

Motivated by the observation that adding more concise CoT examples in the prompt can improve LLM reasoning performance

Source - Microsoft Research

LegalBench - a tool to evaluate the reasoning performance of an LLM in the legal domain.

A paper on benchmarking an LLMs temporal reasoning.

Shall I provide some more?

permalink
report
parent
reply

Technology

!technology@lemmy.world

Create post

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


Community stats

  • 17K

    Monthly active users

  • 6.1K

    Posts

  • 132K

    Comments