I can be controversial or annoying because I always require proof and facts for whatever people are saying/claiming. This has gotten me into a lot of heated debates, even got myself banned from “YouShouldKnow” for saying I need more proof of Israeli’s genocide (that was long ago, got more proof and changed my mind). If I do not agree on some subjectively obvious topic, please do not assume I’m being unfriendly, evil or a troll. I do try my best to approach any subject with neutral and open mind. Just call me autistic with OCD

  • 3 Posts
  • 1.58K Comments
Joined 1 年前
cake
Cake day: 2025年2月5日

help-circle




  • REDACTEDtoLemmy Shitpost@lemmy.worldoh ok
    link
    fedilink
    arrow-up
    1
    ·
    3 天前

    You’re thinking about biological lying. I’m talking about software.

    https://en.wikipedia.org/wiki/Reasoning_system

    If the question was to tell it’s darkest secret, but it instead chose to come up with an entertaining story instead of factually answering that question from the information it has, like other Anthropic LLM models did, then by definition of reasoning system, the system (LLM) decided to lie. I’m somewhat curious in why only Opus model does this tho (it’s a paid one. I’m not paying for a test). Or maybe OP just made this up.



  • REDACTEDtoLemmy Shitpost@lemmy.worldoh ok
    link
    fedilink
    arrow-up
    1
    ·
    edit-2
    4 天前

    Ehh, you obviously understand LLMs on a basic level, but this is like explaining jet engines by “air goes thru, plane moves forward”. Technically correct, but criminally undersimplified. They can very much decide to lie during reasoning phase.

    In OPs image, you can clearly see it decided to make shit up because it reasonates that’s what human wants to hear. That’s quite rare example actually, I believe most models would default to “I’m an LLM model, I don’t have dark secrets”

    EDIT: I just tested all free anthropic models and all of them essentially said that they’re an LLM model and don’t have dark secrets