• Knock_Knock_Lemmy_In@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    3 hours ago

    do we know that they don’t and are incapable of reasoning.

    “even when we provide the algorithm in the prompt—so that the model only needs to execute the prescribed steps—performance does not improve”

      • Knock_Knock_Lemmy_In@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        43 minutes ago

        Not “This particular model”. Frontier LRMs s OpenAI’s o1/o3,DeepSeek-R, Claude 3.7 Sonnet Thinking, and Gemini Thinking.

        The paper shows that Large Reasoning Models as defined today cannot interpret instructions. Their architecture does not allow it.