I found the aeticle in a post on the fediverse, and I can’t find it anymore.

The reaserchers asked a simple mathematical question to an LLM ( like 7+4) and then could see how internally it worked by finding similar paths, but nothing like performing mathematical reasoning, even if the final answer was correct.

Then they asked the LLM to explain how it found the result, what was it’s internal reasoning. The answer was detailed step by step mathematical logic, like a human explaining how to perform an addition.

This showed 2 things:

  • LLM don’t “know” how they work

  • the second answer was a rephrasing of original text used for training that explain how math works, so LLM just used that as an explanation

I think it was a very interesting an meaningful analysis

Can anyone help me find this?

EDIT: thanks to @theunknownmuncher @lemmy.world https://www.anthropic.com/research/tracing-thoughts-language-model its this one

EDIT2: I’m aware LLM dont “know” anything and don’t reason, and it’s exactly why I wanted to find the article. Some more details here: https://feddit.it/post/18191686/13815095

    • Voldemort@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      2 days ago

      Formatting might be off on some of these, had to convert some papers to text as some were only scanned and I couldn’t be bothered writing it all out by hand:

      “The PDP models are inspired by the structure and function of the brain. In particular, they are based on networks of neuron-like units whose interactions resemble those among neurons in the cerebral cortex.” Parallel Distributed Processing: Explorations in the Microstructure of Cognition McClelland, Rumelhart, & the PDP Research Group

      “The design of artificial neural networks was inspired by knowledge of the brain, in particular the way biological neurons are interconnected and the way they communicate through synapses.” Deep Learning LeCun, Bengio, Hinton

      “The design of deep learning architectures owes much to our understanding of the hierarchical structure of the visual cortex, particularly as revealed by Hubel and Wiesel’s work on simple and complex cells.” Neuroscience-Inspired Artificial Intelligence Hassabis et al.

      “The relationship between biological and artificial neural networks has now become a central issue in both neuroscience and AI. Deep networks trained with backpropagation may offer a plausible model of some aspects of human cognition.” Cognitive computational neuroscience Kriegeskorte & Douglas (2018)

      “Goal-driven deep learning models, when trained to solve behavioral tasks, can develop internal representations that match those found in the brain.” Using goal-driven deep learning models to understand sensory cortex Yamins & DiCarlo

      • Voldemort@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        2 days ago

        And longer excepts on the similarities of AI neural networks to biological brains, more specifically human children, in the pursuit of study with improving learning and education development. Super interesting papers that are easily accessible to anyone:

        “Humans are imperfect reasoners. We reason most effectively about entities and situations that are consistent with our understanding of the world. Our experiments show that language models mirror these patterns of behavior. Language models perform imperfectly on logical reasoning tasks, but this performance depends on content and context. Most notably, such models often fail in situations where humans fail — when stimuli become too abstract or conflict with prior understanding of the world. Beyond these parallels, we also observed reasoning effects in language models that to our knowledge have not been previously investigated in the human literature. For example, the patterns of errors on the ‘violate realistic’ rules, or the relative ease of ‘shuffled realistic’ rules in the Wason tasks. Likewise, language model performance on the Wason tasks increases most when they are demonstrated with realistic examples; benefits of concrete examples have been found in cognitive and educational contexts (Sweller et al., 1998; Fyfe et al., 2014), but remain to be explored in the Wason problems. Investigating whether humans show similar effects is a promising direction for future research.” 5.9-10 Language models show human-like content effects on reasoning Ishita Dasgupta*,1, Andrew K. Lampinen*,1, Stephanie C. Y. Chan1, Antonia Creswell1, Dharshan Kumaran1, James L. McClelland1,2 and Felix Hill1 *Equal contributions, listed alphabetically, 1DeepMind, 2Stanford University

        “In this article we will point out several characteristics of human cognitive processes that conventional computer architectures do not capture well. Then we will note that connectionist models {neural networks} are much better able to capture these aspects of human processing. After that we will mention three recent applications in connectionist artificial intelligence which exploit these characteristics. Thus, we shall see that connectionist models offer hope of overcoming the limitations of conventional AI. The paper ends with an example illustrating how connectionist models can change our basic conceptions of the nature of intelligent processing.”

        “The framework for building connectionist models is laid out in detail in Rumelhart, McClelland and the PDP Group (1986), and many examples of models constructed in that framework are described. Two examples of connectionist models of human processing abilities that capture these characteristics are the interactive activation model of visual word recognition from McClelland and Rumelhart (1981), and the model of past tense learning from Rumelhart and McClelland (1986). These models were motivated by psychological experiments, and were constructed to capture the data found in these studies. We describe them here to illustrate some of the roots of the connectionist approach in an attempt to understand detailed aspects of human cognition.”

        “The models just reviewed capture important aspects of data from psychological experiments, and illustrate how the characteristics of human processing capabilities enumerated above can be captured in an explicit comptutational framework. Recently connectionist models that capture these same characteristics have begun to give rise to a new kind of Artificial Intelligence, which we will call connectionist AI. Connectionist AI is beginning to address several topics that have not been easily solved using other approaches. We will consider three cases of this. In each case we will describe recent progress that illustrates the ability of connectionist networks to capture the characteristics of human performance mentioned above.”

        “This paper began with the idea that humans exploit graded information, and that computational mechanisms that aim to emulate the natural processing capabilities of humans should exploit this kind of information as well. Connectionist models do exploit graded information, and this gives them many of their attractive characteristics.” Parallel Distributed Processing: Bridging the Gap Between Human and Machine Intelligence James L. McClelland, Axel Cleeremans, and David Servan-Schreiber Carnegie Mellon University

        “Artificial neural networks have come and gone and come again- and there are several good reasons to think that this time they will be around for quite a while. Cheng and Titterington have done an excellent job describing that nature of neural network models and their relations to statistical methods, and they have overviewed several applications. They have also suggested why neuroscientists interested in modeling the human brain are interested in such models. In this note, I will point out some additional motivations for the investigation of neural networks. These are motivations arising from the effort to capture key aspects of human cognition and learning that have thus far eluded cognitive science. A central goal of congnitive science is to understand the full range of human cognitive function”…“{there are} good reasons for thinking that artificial neural networks, or at least computationally explicit models that capture key properties of such networks, will play an important role in the effort to capture some of the aspects of human cognitive function that have eluded symbolic approaches.” Neural Networks: A Review from Statistical Perspective]: Comment: Neural Networks and Cognitive Science: Motivations and Applications James L. McClelland Statistical Science, Vol. 9, No. 1. (Feb., 1994), pp. 42-45.

        “The idea has arisen that as the scale of experience and computation begins to approach the scale of experience and computation available to a young child—who sees millions of images and hears millions of words per year, and whose brain contains 10–100 billion neuron-like processing units each updating their state on a time scale of milliseconds—the full power and utility of neural networks to capture natural computation is finally beginning to become a reality, allowing artificially intelligent systems to capture more fully the capabilities of the natural intelligence present in real biological networks in the brain.”

        “One major development in the last 25 years has been the explosive growth of computational cognitive neuroscience. The idea that computer simulations of neural mechanisms might yield insight into cognitive phenomena no longer requires, at least in most quarters, vigorous defense—there now exist whole fields, journals, and conferences dedicated to this pursuit. One consequence is the elaboration of a variety of different computationally rigorous approaches to neuroscience and cognition that capture neural information processing mechanisms at varying degrees of abstraction and complexity. These include the dynamic field theory, in which the core representational elements are fields of neurons whose activity and interactions can be expressed as a series of coupled equations (Johnson, Spencer, & Sch€oner, 2008); the neural engineering framework, which seeks to understand how spiking neurons might implement tensor-product approaches to symbolic representations (Eliasmith & Anderson, 2003; Rasmussen & Eliasmith, 2011); and approaches to neural representation based on ideal-observer models and probabilistic inference (Deneve, Latham, & Pouget, 1999; Knill & Pouget, 2004). Though these perspectives differ from PDP in many respects, all of these efforts share the idea that cognition emerges from interactions among populations of neurons whose function can be studied in simplified, abstract form.” Parallel Distributed Processing at 25: Further Explorations in the Microstructure of Cognition T. T. Rogers, J. L. McClelland / Cognitive Science 38 (2014) p1062-1063

        • Voldemort@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          2 days ago

          “In this paper, a simple neural network model will be used to simulate the development of children’s ability to solve equivalence problems. The model treats algebraic problem solving as an acquired skill, emerging slowly from practice solving example problems… In summary, a recurrent neural network can serve as a useful model of how we learn mathematical equivalence… We investigated the strategies of the model, and found that it adopted an add-all strategy like many children do before mastering equivalence problems” A neural network model of learning mathematical equivalence Kevin W. Mickey, James L. McClelland

          “We explore a recurrent neural network model of counting based on the differentiable recurrent attentional model of Gregor et al. (2015). Our results reveal that the model can learn to count the number of items in a display, pointing to each of the items in turn and producing the next item in the count sequence at each step, then saying ‘done’ when there are no more blobs to count. The model thus demonstrates that the ability to learn to count does not depend on special knowledge relevant to the counting task. We find that the model’s ability to count depends on how well it has learned to point to each successive item in the array, underscoring the importance of coordination of the visuospatial act of pointing with the recitation of the count list… Yet the errors that it makes have similarities with the patterns seen in human children’s counting errors, consistent with idea that children rely on graded and somewhat variable mechanisms similar to our neural networks.” Can a Recurrent Neural Network Learn to Count Things? Mengting Fang, Zhenglong Zhou, Sharon Y. Chen, James L. McClelland

          “Over the course of development, humans learn myriad facts about items in the world, and naturally group these items into useful categories and structures. This semantic knowledge is essential for diverse behaviors and inferences in adulthood. How is this richly structured semantic knowledge acquired, organized, deployed, and represented by neuronal networks in the brain? We address this question by studying how the nonlinear learning dynamics of deep linear networks acquires information about complex environmental structures. Our results show that this deep learning dynamics can self-organize emergent hidden representations in a manner that recapitulates many empirical phenomena in human semantic development. Such deep networks thus provide a mathematically tractable window into the development of internal neural representations through experience.” “In addition to matching neural similarity patterns across subjects, experiments using fMRI and single-unit responses have also documented a correspondence between neural similarity patterns and behavioral similarity patterns (21).” A mathematical theory of semantic development in deep neural networks Andrew M. Saxe, James L. McClelland, and Surya Ganguli

          I personally think there are plenty of examples out there in neuroscience and computer science papers let alone what other fields are starting to discover with the use of AI. In my opinion it should be of no surprise and quite clear how replicating a mechanism of self-adapting logic would create behaviours that we can find directly within ourselves.

          Let me know if this is enough to prove my point, but I think I’m tired of reading papers for a bit.