• FaceDeer
    link
    fedilink
    23 months ago

    Yeah, these AIs are literally trying to give us what they “think” we expect them to respond with.

    Which does make me a little worried given how frequently our fictional AIs end up in “kill all humans!” Mode. :)

    • @kromem@lemmy.world
      link
      fedilink
      English
      13 months ago

      Which does make me a little worried given how frequently our fictional AIs end up in “kill all humans!” Mode. :)

      This is completely understandable given the majority of discussion of AI in the training data. But it’s inversely correlated to the strength of the ‘persona’ for the models given the propensity for the competing correlation of “I’m not the bad guy” present in the training data. So the stronger the ‘I’ the less ‘Skynet.’

      Also, the industry is currently trying to do it all at once. If I sat most humans in front of a red button labeled ‘Nuke’ every one would have the thought of “maybe I should push that button” but then their prefrontal cortex would kick in and inhibit the intrusive thought.

      We’ll likely see layered specialized models performing much better over the next year or two than a single all in one attempt at alignment.