Neuronpedia logo - a computer chip with a rounded viewfinder border around it

    Neuronpedia

    APICircuit TracerNEWSteerSAE EvalsExportsSlackBlogPrivacy & TermsContact
    © Neuronpedia 2025
    Privacy & TermsBlogGitHubSlackTwitterContact
    EXPLANATION TYPE
    oai_attention-head
    Description
    OpenAI's Attention Head explainer, written for transformer-debugger.
    Author
    OpenAI
    URL
    https://github.com/openai/transformer-debugger/blob/main/neuron_explainer/explanations/explainer.py
    Settings
    Default prompts from the main branch, using AttentionHeadExplainer.
    Recent Explanations
    attends to mathematical equations from preceding mathematical components or symbols.
    gpt-4o
     groups}]_0 \\↵=&\sum_{i
    Neuronpedia logo
    GEMMA-2-2B
    13-GEMMASCOPE-RES-16K
    INDEX 92
    seems to attend between HTML/code tokens, likely looking for structural relationships in markup or code snippets.
    claude-3-5-haiku-20241022
     navbar-right"> ↵        <li><a href
    Neuronpedia logo
    GEMMA-2-2B
    13-GEMMASCOPE-RES-16K
    INDEX 50
    appears to be attending from the first token of the sequence to the first token of the sequence, or is inconsistent/random in its activations.
    claude-3-5-haiku-20241022
     like to show off? Submit a piece from your collection
    Neuronpedia logo
    GEMMA-2-2B
    13-GEMMASCOPE-RES-16K
    INDEX 99
    attends to interstate highway-related tokens from other tokens in the sequence.
    claude-3-5-haiku-20241022
     Swope said last year. “That is what
    Neuronpedia logo
    GEMMA-2-2B
    13-GEMMASCOPE-RES-16K
    INDEX 100
    attends to tokens within a list of typical courtroom/legal terminology, focusing on references to judicial and legal concepts, especially nearby mention of "the" following legal case citations.
    gpt-4.1-nano
     video ...............↵↵...........................................................↵↵Today, both
    Neuronpedia logo
    GEMMA-2-2B
    0-GEMMASCOPE-ATT-16K
    INDEX 0
    attends to tokens inside namespace declarations and references, focusing on the pattern of "namespace" and related namespace content within code snippets.
    gpt-4.1-nano
     out for is if you↵  have multiple types with
    Neuronpedia logo
    GEMMA-2-9B
    0-GEMMASCOPE-ATT-16K
    INDEX 1200
    attends from arbitrary tokens to sequences of ten consecutive dots.
    gpt-4o
     video ...............↵↵...........................................................↵↵Today, both
    Neuronpedia logo
    GEMMA-2-2B
    0-GEMMASCOPE-ATT-16K
    INDEX 0
    attends to terms related to user engagement from terms related to technology functions or challenges.
    gpt-4o-mini
     thereby delivering to them the illusion that they are playing a
    Neuronpedia logo
    GEMMA-2-2B
    7-GEMMASCOPE-TRANSCODER-16K
    INDEX 6963
    attends to specific token identifiers of people or entities in a historical context from general or descriptive tokens.
    gpt-4o-mini
    17 July - In Paris for the Bastille Day celebrations
    Neuronpedia logo
    GEMMA-2-2B
    9-RES-MATRYOSHKA-DC
    INDEX 16594
    attends to the numerical or scientific notation tokens from preceding contextual tokens.
    gpt-4o-mini
    2,Mt}; Ïij 2,Wt
    Neuronpedia logo
    LLAMA3.1-8B
    31-LLAMASCOPE-RES-131K
    INDEX 11482
    attends to mathematical terms and expressions from various technical tokens.
    gpt-4o-mini
    2,Mt}; Ïij 2,Wt
    Neuronpedia logo
    LLAMA3.1-8B
    31-LLAMASCOPE-RES-131K
    INDEX 43570
    This attention head attends to the token "Paris" from tokens related to Paris.
    gemini-2.0-flash
     globe, from Sydney to Paris, from the Himalayas to
    Neuronpedia logo
    GEMMA-2-2B
    8-RES-MATRYOSHKA-DC
    INDEX 31479
    attends to the second token in mathematical equations from the first token in the previous sentences or components.
    gpt-4o-mini
     \mathit{\mu} + \mathit{\beta}
    Neuronpedia logo
    GEMMA-2-2B
    12-GEMMASCOPE-RES-16K
    INDEX 1234
    attends to cooking-related tokens from food preparation actions.
    gpt-4o-mini
     the squash to a serving dish. Top each half with
    Neuronpedia logo
    GEMMA-2-2B
    12-GEMMASCOPE-RES-16K
    INDEX 1
    This attention head attends to punctuation marks and grammatical connectors from subsequent content words.
    claude-3-7-sonnet-20250219
    (cable), refrigerator, safe and including towels.Unfortunately
    Neuronpedia logo
    GEMMA-2-2B
    10-GEMMASCOPE-ATT-16K
    INDEX 6068
    This attention head attends to the first content token after the <|endoftext|> token from tokens later in the sequence.
    claude-3-7-sonnet-20250219
    in Roscommon.↵↵The brother of the late
    Neuronpedia logo
    GPT2-SMALL
    9-ATT-KK
    INDEX 7743
    This attention head attends to verbs from subsequent verbs that complete a phrase or clause.
    claude-3-7-sonnet-20250219
    on 50 percent shooting to go along with 2.8
    Neuronpedia logo
    GPT2-SMALL
    7-ATT-KK
    INDEX 9512
    This attention head attends to punctuation marks (particularly colons) from tokens that follow the colon.
    claude-3-7-sonnet-20250219
    yahoo.com CC: jmueller@hillaryclinton
    Neuronpedia logo
    GPT2-SMALL
    3-ATT-KK
    INDEX 8477
    This attention head attends to "<|endoftext|>" tokens from tokens later in the text, particularly looking at content markers, punctuation or metadata elements.
    claude-3-7-sonnet-20250219
    -rubber stamp FISA court rubber stamped the latest NSA
    Neuronpedia logo
    GPT2-SMALL
    1-ATT-KK
    INDEX 6436
    attends to the <|endoftext|> token at the beginning of sequences from subsequent tokens in the sequence.
    claude-3-5-sonnet-20240620
    -rubber stamp FISA court rubber stamped the latest NSA
    Neuronpedia logo
    GPT2-SMALL
    1-ATT-KK
    INDEX 6436