10
submitted 9 months ago* (last edited 9 months ago) by kromem@lemmy.world to c/technology@lemmy.world

I've been saying this for about a year since seeing the Othello GPT research, but it's nice to see more minds changing as the research builds up.

Edit: Because people aren't actually reading and just commenting based on the headline, a relevant part of the article:

New research may have intimations of an answer. A theory developed by Sanjeev Arora of Princeton University and Anirudh Goyal, a research scientist at Google DeepMind, suggests that the largest of today’s LLMs are not stochastic parrots. The authors argue that as these models get bigger and are trained on more data, they improve on individual language-related abilities and also develop new ones by combining skills in a manner that hints at understanding — combinations that were unlikely to exist in the training data.

This theoretical approach, which provides a mathematically provable argument for how and why an LLM can develop so many abilities, has convinced experts like Hinton, and others. And when Arora and his team tested some of its predictions, they found that these models behaved almost exactly as expected. From all accounts, they’ve made a strong case that the largest LLMs are not just parroting what they’ve seen before.

“[They] cannot be just mimicking what has been seen in the training data,” said Sébastien Bubeck, a mathematician and computer scientist at Microsoft Research who was not part of the work. “That’s the basic insight.”

you are viewing a single comment's thread
view the rest of the comments
[-] Redacted@lemmy.world 8 points 9 months ago* (last edited 9 months ago)

I have a theory... They are sophisticated auto-complete.

[-] kromem@lemmy.world 2 points 9 months ago* (last edited 9 months ago)

You are making the common mistake of confusing how they are trained with how they operate.

For example, in the MIT/Harvard Othello-GPT paper I mentioned, feeding in only millions of legal Othello moves into a GPT model (i.e. trained to autocomplete moves) resulted in the neural network internally building a world model of an Othello board - even though it wasn't explicitly told anything about the board outside of being fed legal moves.

Later, a researcher at DeepMind replicated the work and found it was encoded as a linear representation, which has then since been shown to be how models encode a number of other world models developed from their training corpus (Max Tegmark coauthored two interesting studies in particular about this regarding modeling space and time and modeling truthiness).

[-] Redacted@lemmy.world 0 points 9 months ago

They operate by weighting connections between patterns they identify in their training data. They then use statistics to predict outcomes.

I am not particularly surprised that the Othello models built up an internal model of the game as their training data were grid moves. Without loooking into it I'd assume the most efficient way of storing that information was in a grid format with specific nodes weighted to the successful moves. To me that's less impressive than the LLMs.

[-] kromem@lemmy.world 0 points 9 months ago

They operate by weighting connections between patterns they identify in their training data. They then use statistics to predict outcomes.

Again, this isn't quite correct. They can do this, but it isn't the only way they can achieve completion of tokens.

Without loooking into it I'd assume the most efficient way of storing that information was in a grid format with specific nodes weighted to the successful moves.

(It also developed representations of what constituted legal vs non-legal moves.)

You are getting closer to the point. Think about a model asked to complete Pythagorean theorem sequences based on a, b inputs to arrive at c inputs.

What's the most efficient way to represent that data for successfully completing sequences?

[-] Redacted@lemmy.world 0 points 9 months ago

So somewhere in there I'd expect nodes connected to represent the Othello grid. They wouldn't necessarily be in a grid, just topologically the same graph.

Then I'd expect millions of other weighted connections to represent the moves within the grid including some weightings to prevent illegal moves. All based on mathematics and clever statistical analysis of the training data. If you want to refer to things as tokens then be my guest but it's all graphs.

If you think I'm getting closer to your point can you just explain it properly? I don't understand what you think a neural network model is or what you are trying to teach me with Pythag.

[-] kromem@lemmy.world 0 points 9 months ago

If you think I'm getting closer to your point can you just explain it properly?

The most efficient way for a neural network to predict Pythagorean results given inputs would be to reverse engineer a Pythagorean function within itself rather than simply trying to model statistical relationships between inputs and results. To effectively build a world model of Pythagorean calculation.

Training to autocomplete doesn't mean that the way it achieves this is limited to any one approach or solution, and it would be useful to keep in mind that a neural network of unbounded size can model any possible function.

[-] Redacted@lemmy.world 0 points 9 months ago

It wouldn't reverse engineer anything. It would start by weighting neurons based on it's training set of Pythagorean triples. Over time this would get tuned to represent Pythag in the form of mathematical graphs.

This is not "understanding" as most people would know it. More like a set of encoded rules.

[-] kromem@lemmy.world 0 points 9 months ago* (last edited 9 months ago)

Furthermore, we presented a method for adapting a symbolic function learner to find mathematical relationships between measured quantities in an unsupervised way. This method does not explicitly label any variate as being an output value, bypassing the assumptions made in standard regression problems about causal relationships. We demonstrated how this method was able to recover rules of geometry from raw data alone. This included the law of sines and the Pythagorean theorem, two relationships on measurements relating to triangles where no variable is necessarily considered an output of a function of the others.

  • Panju, Automated Knowledge Discovery Using Neural Networks (2021)
[-] Redacted@lemmy.world 0 points 9 months ago* (last edited 9 months ago)

Seems to me you are attempting to understand machine learning mathematics through articles.

That quote is not a retort to anything I said.

Look up Category Theory. It demonstrates how the laws of mathematics can be derived by forming logical categories. From that you should be able to imagine how a neural network could perform a similar task within its structure.

It is not understanding, just encoding to arrive at correct results.

[-] kromem@lemmy.world 0 points 9 months ago

What I quoted isn't an article, it was a mathematics dissertation.

And you disputed that a NN could arrive at the theorem before being corrected about it.

[-] Redacted@lemmy.world 0 points 9 months ago* (last edited 9 months ago)

There you go arguing in bad faith again by putting words in my mouth and reducing the nuance of what was said.

You do know dissertations are articles and don't constitute any form or rigorous proof in and of themselves? Seems like you have a very rudimentary understanding of English, which might explain why you keep struggling with semantics. If that is so, I apologise because definitions are difficult when it comes to language, let alone ESL.

I didn't dispute that NNs can arrive at a theorem. I debate whether they truly understand the theorem they have encoded in their graphs as you claim.

This is a philosophical/semantical debate as to what "understanding" actually is because there's not really any evidence that they are any more than clever pattern recognition algorithms driven by mathematics.

[-] kromem@lemmy.world 0 points 9 months ago

I debate whether they truly understand the theorem they have encoded in their graphs as you claim.

Where did I claim that? Cite the exact phrase.

I said reverse engineer. Not deduce or prove.

[-] Redacted@lemmy.world 0 points 9 months ago* (last edited 9 months ago)

Title of your post is literally "New Theory Suggests Chatbots Can Understand Text".

You also hinted at it with your Pythag analogy.

[-] kromem@lemmy.world 0 points 9 months ago

Title of your post is literally "New Theory Suggests Chatbots Can Understand Text".

I didn't write the headline, and I happen to interpret it the same way I interpreted it in "Bees understand the concept of zero." Language can have more than one narrowly scoped meaning, and the article body makes it clear it isn't saying anything about human consciousness or introspective understanding.

You also hinted at it with your Pythag analogy.

No, I correctly stated that a model happening upon the Pythagorean function would outperform ones approximating it by statistical correlations. That, as Hinton has said in the past, "predicting the next thing takes knowledge." It makes sense that the development of world models and abstractions from the training data and not simply surface statistics would correlate with both increased next token prediction and network complexity increases.

You interpreted what I was saying as implying the network has some woo woo interpretation of 'understanding' because you seem to be more committed to debating a straw man using inaccurate and overly narrow semantics than actually discussing the topic at hand in good faith.

[-] Redacted@lemmy.world 0 points 9 months ago* (last edited 9 months ago)

You posted the article rather than the research paper and had every chance of altering the headline before you posted it but didn't.

You questioned why you were downvoted so I offered an explanation.

Your attempts to form your own arguments often boil down to "no you".

So as I've said all along we just differ on our definitions of the term "understanding" and have devolved into a semantic exchange. You are now using a bee analogy but for a start that is a living thing not a mathematical model, another indication that you don't understand nuance. Secondly, again, it's about definitions. Bees don't understand the number zero in the middle of the number line but I'd agree they understand the concept of nothing as in "There is no food."

As you can clearly see from the other comments, most people interpret the word "understanding" differently from yourself and AI proponents. So I infer you are either not a native English speaker or are trying very hard to shoehorn your oversimplified definition in to support your worldview. I'm not sure which but your reductionist way of arguing is ridiculous as others have pointed out and full of logical fallacies which you don't seem to comprehend either.

Regarding what you said about Pythag, I agree and would expect it to outperform statistical analysis. That is due to the fact that it has arrived at and encoded the theorem within its graphs but I and many others do not define this as knowledge or understanding because they have other connotations to the majority of humans. It wouldn't for instance be able to tell you what a triangle is using that model alone.

I spot another apeal to authority... "Hinton said so and so..." It matters not. If Hinton said the sky is green you'd believe it as you barely think for yourself when others you consider more knowledgeable have stated something which may or may not be true. Might explain why you have such an affinity for AI...

[-] kromem@lemmy.world -1 points 9 months ago
[-] Redacted@lemmy.world 1 points 9 months ago

Lol indeed, just seen you moderate a Simulation Theory sub.

Congratulations, you have completed the tech evangelist starter pack.

Next thing you'll be telling me we don't have to worry about climate change because we'll just use carbon capture tech and failing that all board Daddy Elon's spaceship to teraform Mars.

[-] FaceDeer@kbin.social 2 points 9 months ago

I have a theory... so are you and I.

[-] Redacted@lemmy.world 1 points 9 months ago

Orders of magnitude of differece between the most complex known object in the universe and some clever statistical analysis.

We understand very little about the human brain. For example, we don't know if it leverages quantum interactions or whether it can be decoupled from its substrate.

LLMs are pattern matching models loosly based on the structure of neurons that work well for deriving predictions from a vast body of data but are not anywhere near human brain level of understanding. I personally don't think they will ever be until we have solved the hard problem of conciousness.

[-] FaceDeer@kbin.social -2 points 9 months ago
[-] Redacted@lemmy.world 2 points 9 months ago

Welp looks like we both know the arguments and fall on different sides of the debate then.

Much better than being confidently wrong like most LLMs...

[-] Phanatik@kbin.social -1 points 9 months ago

I don't need a theory for this, you're being highly reductive by focusing on a few features of human communication.

[-] Redacted@lemmy.world 1 points 9 months ago

Thank you, much more succinctly put than my attempt.

[-] Phanatik@kbin.social 1 points 9 months ago

I've just done the dance already and I'm tired of their watered-down attempts at bringing human complexity down to a level that makes their chat bots seem smart.

this post was submitted on 24 Jan 2024
10 points (56.1% liked)

Technology

59374 readers
3467 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS