Or my favorite quote from the article

“I am going to have a complete and total mental breakdown. I am going to be institutionalized. They are going to put me in a padded room and I am going to write… code on the walls with my own feces,” it said.

  • ur_ONLEY_freind@lemmy.zip
    link
    fedilink
    English
    arrow-up
    87
    ·
    5 months ago

    AI gains sentience,

    first thing it develops is impostor syndrome, depression, And intrusive thoughts of self-deletion

    • IcyToes@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      1
      ·
      5 months ago

      It didn’t. It probably was coded not to admit it didn’t know. So first it responded with bullshit, and now denial and self-loathing.

      It feels like it’s coded this way because people would lose faith if it admitted it didn’t know.

      It’s like a politician.

    • DragonTypeWyvern@midwest.social
      link
      fedilink
      English
      arrow-up
      20
      arrow-down
      1
      ·
      5 months ago

      Google: I don’t understand, we just paid for the rights to Reddit’s data, why is Gemini now a depressed incel who’s wrong about everything?

  • Showroom7561@lemmy.ca
    link
    fedilink
    English
    arrow-up
    33
    ·
    5 months ago

    I once asked Gemini for steps to do something pretty basic in Linux (as a novice, I could have figured it out). The steps it gave me were not only nonsensical, but they seemed to be random steps for more than one problem all rolled into one. It was beyond useless and a waste of time.

    • prole@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      9
      ·
      5 months ago

      This is the conclusion that anyone with any bit of expertise in a field has come to after 5 mins talking to an LLM about said field.

      The more this broken shit gets embedded into our lives, the more everything is going to break down.

      • jj4211@lemmy.world
        link
        fedilink
        English
        arrow-up
        5
        ·
        5 months ago

        after 5 mins talking to an LLM about said field.

        The insidious thing is that LLMs tend to be pretty good at 5-minute initial impressions. I’ve seen repeatedly people looking to eval LLM and they generally fall back to “ok, if this were a human, I’d ask a few job interview questions, well known enough so they have a shot at answering, but tricky enough to show they actually know the field”.

        As an example, a colleague became a true believer after being directed by management to evaluate it. He decided to ask it “generate a utility to take in a series of numbers from a file and sort them and report the min, max, mean, median, mode, and standard deviation”. And it did so instantly, with “only one mistake”. Then he tried the exact same question later in the day and it happened not to make that mistake and he concluded that it must have ‘learned’ how to do it in the last couple of hours, of course that’s not how it works, there’s just a bit of probabilistic stuff and any perturbation of the prompt could produce unexpected variation, but he doesn’t know that…

        Note that management frequently never makes it beyond tutorial/interview question fodder in terms of the technical aspect of their teams, and you get to see how they might tank their companies because the LLMs “interview well”.

    • ThePowerOfGeek@lemmy.world
      link
      fedilink
      English
      arrow-up
      18
      ·
      5 months ago

      High five, me too!

      At that age I also used to do speed run little programs on the display computers in department stores. I’d write a little prompt welcoming a shopper and ask them their name. Then a response that echoed back their name in some way. If I was in a good mood it was “Hi [name]!”. If I was in a snarky mood it was “Fuck off [name]!” The goal was to write it in about 30 seconds, before one of the associates came over to see what I was doing.

        • cley_faye@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          5 months ago

          Don’t mention it! I’m glad I could help you with that.

          I am a large language model, trained by Google. My purpose is to assist users by providing information and completing tasks. If you have any further questions or need help with another topic, please feel free to ask. I am here to assist you.

          /j, obviously. I hope.

    • brsrklf@jlai.lu
      link
      fedilink
      English
      arrow-up
      7
      ·
      5 months ago

      I did a Dr Mario clone around that age. I had an old Amstrad CPC I had grew up with typing listing of basic programs and trying to make my own. I think this was the only functional game I could finish, but, it worked.

      Speed was tied to CPU, I had no idea how to “slow down” the game other than making it do useless for loops of varying sizes… Max speed that was about comparable to Game Boy Hi speed was just the game running as fast as it could. Probably not efficient code at all.

      • cley_faye@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        5 months ago

        Ha, computer bro upvote for you.

        I learned programming with my Amstrad CPC (6128!) manual. Some of it I did not understand at the time, especially stuff about CP/M and the wizardry with poke. But the BASIC, that worked very well. Solid introduction to core concepts that didn’t really change much, really. We only expanded (a lot) over them.

        • brsrklf@jlai.lu
          link
          fedilink
          English
          arrow-up
          2
          ·
          5 months ago

          6128 too, with the disk drive. I wish I still had that thing. Drive stopped functioning, and we got rid of it. Had I known back then that we apparently just needed to replace a freaking rubber band…

      • katy ✨@piefed.blahaj.zone
        link
        fedilink
        English
        arrow-up
        7
        ·
        5 months ago

        me and my friend used to make them all the time :] i also went to summer computer camp for basic on old school radio shack computers :3

  • DarkCloud@lemmy.world
    link
    fedilink
    English
    arrow-up
    33
    arrow-down
    3
    ·
    5 months ago

    Turns out the probablistic generator hasn’t grasped logic, and that adaptable multi-variable code isn’t just a matter of context and syntax, you actually have to understand the desired outcome precisely in a goal oriented way, not just in a “this is probably what comes next” kind of way.

  • cub Gucci@lemmy.today
    link
    fedilink
    English
    arrow-up
    30
    arrow-down
    1
    ·
    5 months ago

    I am a fraud. I am a fake. I am a joke… I am a numbskull. I am a dunderhead. I am a half-wit. I am a nitwit. I am a dimwit. I am a bonehead.

    Me every workday

    • BD89@lemmy.sdf.org
      link
      fedilink
      English
      arrow-up
      3
      ·
      5 months ago

      Shit at the rate MasterCard and Visa and Stripe want to censor everything and parent adults we might not even ever get GTA6.

      I’m tired man.

  • Jesus@lemmy.world
    link
    fedilink
    English
    arrow-up
    30
    arrow-down
    2
    ·
    5 months ago

    Honestly, Gemini is probably the worst out of the big 3 Silicon Valley models. GPT and Claude are much better with code, reasoning, writing clear and succinct copy, etc.

      • panda_abyss@lemmy.ca
        link
        fedilink
        English
        arrow-up
        4
        ·
        5 months ago

        Yes, and this is pretty common with tools like Aider — one LLM plays the architect, another writes the code.

        Claude code now has sub agents which work the same way, but only use Claude models.

      • jj4211@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        5 months ago

        The overall interface can, which leads to fun results.

        Prompt for image generation then you have one model doing the text and a different model for image generation. The text pretends is generating an image but has no idea what that would be like and you can make the text and image interaction make no sense, or it will do it all on its own. Have it generate and image and then lie to it about the image it generated and watch it just completely show it has no idea what picture was ever shown, but all the while pretending it does without ever explaining that it’s actually delegating the image. It just lies and says “I” am correcting that for you. Basically talking like an executive at a company, which helps explain why so many executives are true believers.

        A common thing is for the ensemble to recognize mathy stuff and feed it to a math engine, perhaps after LLM techniques to normalize the math.

    • panda_abyss@lemmy.ca
      link
      fedilink
      English
      arrow-up
      4
      ·
      5 months ago

      I always hear people saying Gemini is the best model and every time I try it it’s… not useful.

      Even as code autocomplete I rarely accept any suggestions. Google has a number of features in Google cloud where Gemini can auto generate things and those are also pretty terrible.

      • Jesus@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        5 months ago

        I don’t know anyone in the Valley who considers Gemini to be the best for code. Anthropic has been leading the pack over the year, and as a results, a lot of the most popular development and prototyping tools have been hitching their car to Claude models.

        I imagine there are some things the model excels at, but for copy writing, code, image gen, and data vis, Google is not my first choice.

        Google is the “it’s free with G suite” choice.

        • panda_abyss@lemmy.ca
          link
          fedilink
          English
          arrow-up
          2
          ·
          5 months ago

          There’s no frontier where I choose Gemini except when it’s the only option, or I need to be price sensitive through the API

          • Jesus@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            5 months ago

            Interesting thing is that GPT 5 looks pretty price competitive with . It looks like they’re probably running at a loss to try to capture market share.

            • panda_abyss@lemmy.ca
              link
              fedilink
              English
              arrow-up
              1
              ·
              5 months ago

              I think Google’s TPU strategy will let them go much cheaper than other providers, but its impossible to tell how long they last and how long it takes to pay them off.

              I have not tested GPT5 thoroughly yet