• progress_activity cloud_sync

    Reconnection to the server…

    Movim cannot talk with the server, please try again later

  • back_to_tab fullscreen tile_small dialpad mic videocam switch_camera screen_share

    mic_none No sound detected from your microphone


    • ArsTechnica


      article 31785 posts • people 689 subscribers
      assignment_ind Only publishers can publish


    • Public subscriptions

    • chevron_right

      coopr8

    • chevron_right

      gabagoo

    • chevron_right

      kenu_demon

  • Register Login

    Movim

    movim.chatterboxtown.us


  • rss_feed
    add Follow

    ArsTechnica

    people 689 subscribers • news.movim.eu

    • chevron_right

      Researchers isolate memorization from reasoning in AI neural networks

      news.movim.eu / ArsTechnica • 10 November 2025

    When engineers build AI language models like GPT-5 from training data, at least two major processing features emerge: memorization (reciting exact text they’ve seen before, like famous quotes or passages from books) and reasoning (solving new problems using general principles). New research from AI startup Goodfire.ai provides the first potentially clear evidence that these different functions actually work through completely separate neural pathways in the model’s architecture.

    The researchers discovered that this separation proves remarkably clean. In a preprint paper released in late October, they described that when they removed the memorization pathways, models lost 97 percent of their ability to recite training data verbatim but kept nearly all their “logical reasoning” ability intact.

    For example, at layer 22 in Allen Institute for AI’s OLMo-7B language model, the bottom 50 percent of weight components showed 23 percent higher activation on memorized data, while the top 10 percent showed 26 percent higher activation on general, non-memorized text. This mechanistic split enabled the researchers to surgically remove memorization while preserving other capabilities.

    Read full article

    Comments

    • tagbiz & it tagai tagmachine learning tagcopyright tagai safety tagai alignment tagai research tagai behavior tagai architecture tagai memorization tagallen institute for ai tagcomputational neuroscience taggeneralization taggoodfire taggradient descent tagk-fac tagloss curvature tagmechanistic interpretability tagmodel editing tagneural networks tagolmo tagoverfitting tagtransformer models tagweight matrices

    • Pictures 1 image

    • visibility
  • cloud_queue

    Powered by Movim