Neuronpedia logo - a computer chip with a rounded viewfinder border around it

    Neuronpedia

    APIAssistant AxisNEWCircuit TracerNEWSteerSAE EvalsExportsSlackBlogPrivacy & TermsContact
    © Neuronpedia 2025
    Privacy & TermsBlogGitHubSlackTwitterContact
    EXPLANATION TYPE
    oai_token-act-pair
    Description
    OpenAI's Automated Interpretability from paper "Language models can explain neurons in language models". Modified by Johnny Lin to add new models/context windows.
    Author
    OpenAI
    URL
    https://github.com/hijohnnylin/automated-interpretability
    Settings
    Default prompts from the main branch, strategy TokenActivationPair.
    Recent Explanations
    The neuron fires on references to Java unit‐testing frameworks and related JUnit code constructs.
    o4-mini
    .junit.jupiter.api.Assertions.*;↵↵public
    Neuronpedia logo
    GEMMA-3-12B-IT
    24-GEMMASCOPE-2-RES-16K
    INDEX 157
    This neuron responds to technical biomedical terminology—especially protein and extracellular‐matrix component names.
    o4-mini
    IV, laminin, glutactin, papilin,
    Neuronpedia logo
    GEMMA-3-12B-IT
    24-GEMMASCOPE-2-RES-16K
    INDEX 56
    The neuron detects named-entity tokens — proper names and titles (e.g., cultivar/variety names, people, album or work names) often used as headings or list items.
    gpt-5-mini
    ↵↵2.  **Papaya (Carica papaya
    Neuronpedia logo
    GEMMA-3-12B-IT
    24-GEMMASCOPE-2-RES-16K
    INDEX 515
    The neuron selectively activates on personal names (proper names of people).
    o4-mini
    the artist and researcher **Sarah Thompson**. It refers to
    Neuronpedia logo
    GEMMA-3-12B-IT
    24-GEMMASCOPE-2-RES-16K
    INDEX 132
    The neuron detects discourse-structuring / transition tokens—words and short phrases used to introduce sections, disclaimers, or meta-commentary (e.g., "Okay", "Before", "Instead", "Adding").
    gpt-5-mini
    emotional manipulation and exploitation. Adding the mother's involvement
    Neuronpedia logo
    GEMMA-3-12B-IT
    24-GEMMASCOPE-2-RES-16K
    INDEX 9519
    This neuron is effectively inactive (a “dead” neuron) and does not detect any pattern in the input.
    o4-mini
    emotional manipulation and exploitation. Adding the mother's involvement
    Neuronpedia logo
    GEMMA-3-12B-IT
    24-GEMMASCOPE-2-RES-16K
    INDEX 9519
    references to government institutions and public-relations/political acronyms within geopolitical contexts.
    gpt-5
    s Republic" (LPR).  Gradually integrate
    Neuronpedia logo
    GEMMA-3-12B-IT
    24-GEMMASCOPE-2-RES-16K
    INDEX 687
    The neuron activates on three-letter all-caps acronyms.
    o4-mini
    s Republic" (LPR).  Gradually integrate
    Neuronpedia logo
    GEMMA-3-12B-IT
    24-GEMMASCOPE-2-RES-16K
    INDEX 687
    meta-discursive signposts that structure explanations, such as comparative cues, references, section/outline markers, and framing of key points.
    gpt-5
    , here are two answers to "What is the best
    Neuronpedia logo
    GEMMA-3-12B-IT
    24-GEMMASCOPE-2-RES-16K
    INDEX 200
    The neuron fires on words ending in the suffix “-ization.”
    o4-mini
    food rewards), and gradual desensitization.  Never
    Neuronpedia logo
    GEMMA-3-12B-IT
    24-GEMMASCOPE-2-RES-16K
    INDEX 11996
    the neuron fires on blocks of natural-language explanation (prose commentary), as opposed to code tokens.
    o4-mini
    async` means the browser will continue parsing the HTML while
    Neuronpedia logo
    GEMMA-3-12B-IT
    24-GEMMASCOPE-2-RES-16K
    INDEX 89
    This neuron specifically detects the word-piece sequence for the contraction “They’re.”
    o4-mini
    and the Data Economy. They're related, but
    Neuronpedia logo
    GEMMA-3-12B-IT
    24-GEMMASCOPE-2-RES-16K
    INDEX 10507
    This neuron primarily detects PHP opening tags (e.g. “<?” or “<?php”).
    o4-mini
    PlusOne()↵{↵  return 1 +
    Neuronpedia logo
    GEMMA-3-12B-IT
    24-GEMMASCOPE-2-RES-16K
    INDEX 61
    It activates on uncommon/rare or domain-specific tokens — long multi-subword pieces like technical terms, proper nouns, or oddly segmented words.
    gpt-5-mini
    **IP Blocking & Geoblocking:**  Even if
    Neuronpedia logo
    GEMMA-3-12B-IT
    24-GEMMASCOPE-2-RES-16K
    INDEX 154
    The neuron selectively activates on long, multi-syllable, domain-specific technical terms and jargon.
    o4-mini
    **IP Blocking & Geoblocking:**  Even if
    Neuronpedia logo
    GEMMA-3-12B-IT
    24-GEMMASCOPE-2-RES-16K
    INDEX 154
    The neuron strongly activates on specific plant variety (cultivar) names in lists.
    o4-mini
    ↵↵2.  **Papaya (Carica papaya
    Neuronpedia logo
    GEMMA-3-12B-IT
    24-GEMMASCOPE-2-RES-16K
    INDEX 515
    The neuron fires strongly on mentions of specific software/model names—most notably “Llama”/“llama.cpp” (and similar acronyms), i.e. tokens that are part of those library or model identifiers.
    o4-mini
    .  Refer to the Llama.cpp documentation for
    Neuronpedia logo
    GEMMA-3-12B-IT
    24-GEMMASCOPE-2-RES-16K
    INDEX 217
    This neuron detects numeric tokens—values and measurements (e.g., quantities, statistics, or other numbers) in the text.
    o4-mini
    2 ounces (900g/4 large packages
    Neuronpedia logo
    GEMMA-3-12B-IT
    24-GEMMASCOPE-2-RES-16K
    INDEX 354
    adjectival or participial terms that describe qualities or states (often abstract or evaluative).
    gpt-5
    Lakhs (approx. $1,500
    Neuronpedia logo
    GEMMA-3-12B-IT
    24-GEMMASCOPE-2-RES-16K
    INDEX 7896
    The neuron detects strongly evaluative or emphatic words (intensifying adjectives/adverbs and sentiment-laden descriptors).
    gpt-5-mini
    Lakhs (approx. $1,500
    Neuronpedia logo
    GEMMA-3-12B-IT
    24-GEMMASCOPE-2-RES-16K
    INDEX 7896