There have been multiple accounts created with the sole purpose of posting advertisement posts or replies containing unsolicited advertising.

Accounts which solely post advertisements, or persistently post them may be terminated.

LLMs develop their own understanding of reality as their language abilities improve

“An intriguing open question is whether the LLM is actually using its internal model of reality to reason about that reality as it solves the robot navigation problem,” says Rinard. “While our results are consistent with the LLM using the model in this way, our experiments are not designed to answer this next question.”

The paper, “Emergent Representations of Program Semantics in Language Models Trained on Programs” can be found here.

Abstract

We present evidence that language models (LMs) of code can learn to represent the formal semantics of programs, despite being trained only to perform next-token prediction. Specifically, we train a Transformer model on a synthetic corpus of programs written in a domain-specific language for navigating 2D grid world environments. Each program in the corpus is preceded by a (partial) specification in the form of several input-output grid world states. Despite providing no further inductive biases, we find that a probing classifier is able to extract increasingly accurate representations of the unobserved, intermediate grid world states from the LM hidden states over the course of training, suggesting the LM acquires an emergent ability to interpret programs in the formal sense. We also develop a novel interventional baseline that enables us to disambiguate what is represented by the LM as opposed to learned by the probe. We anticipate that this technique may be generally applicable to a broad range of semantic probing experiments. In summary, this paper does not propose any new techniques for training LMs of code, but develops an experimental framework for and provides insights into the acquisition and representation of formal semantics in statistical models of code.

A_A ,
@A_A@lemmy.world avatar

i feel that many of us, when confronted to this, are doing like Z.B. (president of the Galaxy) in The hitchhiker guide… when he says :
… “whenever I stop and think why did I want to do something? – how did I work out how to do it? – I get a very strong desire to just to stop thinking about it” …
We don’t want to be surpassed by machines … and this explains the large amount of downvotes

atrielienz ,

I’m actually pretty sure the downvotes are because LLM’s don’t think. They don’t even process. They pick the highest number and spit out the information attached to it.

Deceptichum ,
@Deceptichum@quokk.au avatar

Do I “think” or does my brain pick the closest neuron and spit out a function based on that input?

If we could recreate the universe, would I do the exact same thing in the exact same situation?

A_A ,
@A_A@lemmy.world avatar

Science cannot say much about what it is to think since it doesn’t understand the brain well enough … and the day we can fully explained it, we will also be able to replicated it on computers.

wizardbeard ,

Wow, that headline is significantly overstated. This has nothing to do with “understanding” or “reality”. Like holy shit this is an absurd leap from what they actually state in the abstract. Thanks OP for including it in the “preview” here.

Given a number of “starting states” and “ending states” of 2D grid based “environments”, an LLM is able to return increasing accurate “between states” as it is trained.

Neat! Not an understanding of reality. At the biggest stretch it is an “understanding” of an incredibly constrained problem space.

Hackworth OP ,

I did some source digging to hopefully best address your observations. Science journalism (even when internal and likely done in concert with the authors) is fundamentally a game of telephone. But looking at the source papers:

They say it in an incredibly formal way, but they do seem to come to the conclusion that the LLM develops understanding. The paper makes that case within an incredibly narrow context, but it does include:

We anticipate that this technique may be generally applicable to a broad range of semantic probing experiments. We argue that the observed semantic content cannot be fully attributed to a retrieval-like process, and instead requires the LM to perform some degree of generalization over the semantics. More broadly, we see programs and their precise formal semantics as a promising direction for working toward a deeper understanding of the behavior of LMs, such as whether or how LMs acquire and use semantic representations of the underlying domain more generally.

With it now clear that the generalized case is not shown: the specific type of understanding that they have shown is non-trivial.

Conclusion: This paper presents empirical evidence that LMs of code can acquire the formal semantics of programs from next token prediction.

A foundational topic in the theory of programming languages, formal semantics (Winskel, 1993) is the study of how to formally specify the meaning of programs.

From Winskel: The Formal Semantics of Programming Languages provides the basic mathematical techniques necessary for those who are beginning a study of the semantics and logics of programming languages. These techniques will allow students to invent, formalize, and justify rules with which to reason about a variety of programming languages.

Also notable but unrelated: Jin and Rinard’s paper was supported, in part, by grants from the U.S. Defense Advanced Research Projects Agency (DARPA).

MagicShel ,

I mostly get what you’re saying, though I don’t have the requisite understanding to follow formal proofs, but if there is one thing I do know for certain, it’s that “understanding” is anthropomorphizing and shorthand for something that is very much not understanding in a human context at all.

I get that it can be hard to find the right words to explain a some of these emergent phenomena, but I think it’s misleading to use words that make AI appear to have a thought process akin to anything we could understand as such—at least in settings where folks might not understand the shorthand as such.

And maybe everyone here is aware of that, but it makes me uneasy, hence this comment to hopefully make that point.

Hackworth OP , (edited )

The paper is kind of saying that as well. I added a quote to the post to help set the context a bit more. As I understand it, they’ve shown that an LLM contains a model of its “world” (training data) and that this model becomes a more meaningful map of that “world” the longer the model is trained. Notably, they haven’t shown that this model is actively employed when the LLM is generating text (robot commands in this case), only that it exists within the neural network and can be probed. And to be clear - its world is so dissimilar from ours, the form its understanding takes is likely to seem alien.

Reverendender ,

We’re all sick of LLM, but the article is actually a really interesting read. How any of these systems can “understand” anything remains to be sufficiently explained. The abstract is not very indicative of the article content.

Hackworth OP ,

From MIT again - our exploration of how LLMs can do the things they can is pointing us in some interesting directions re: our exploration of how our own brains understand.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • [email protected]
  • random
  • lifeLocal
  • goranko
  • All magazines