I think AI is neat.

  • Dr. JenkemA
    link
    fedilink
    English
    arrow-up
    12
    arrow-down
    3
    ·
    5 months ago

    I mean of you have two boxes, one of which is actually intelligent and the other is “just” a very advanced parrot - it doesn’t matter, given they produce the same output.

    You’re making a huge assumption; that an advanced parrot produces the same output as something with general intelligence. And I reject that assumption. Something with general intelligence can produce something novel. An advanced parrot can only repeat things it’s already heard.

      • rambaroo@lemmy.world
        link
        fedilink
        arrow-up
        2
        arrow-down
        2
        ·
        5 months ago

        LLMs can’t produce anything without being prompted by a human. There’s nothing intelligent about them. Imo it’s an abuse of the word intelligence since they have exactly zero autonomy.

    • Meowoem@sh.itjust.works
      link
      fedilink
      arrow-up
      2
      arrow-down
      5
      ·
      5 months ago

      I use LLMs to create things no human has likely ever said and it’s great at it, for example

      ‘while juggling chainsaws atop a unicycle made of marshmallows, I pondered the existential implications of the colour blue on a pineapples dream of becoming a unicorn’

      When I ask it to do the same using neologisms the output is even better, one of the words was exquimodal which I then asked for it to invent an etymology and it came up with one that combined excuistus and modial to define it as something beyond traditional measures which fits perfectly into the sentence it created.

      You can’t ask a parrot to invent words with meaning and use them in context, that’s a step beyond repetition - of course it’s not full dynamic self aware reasoning but it’s certainly not being a parrot

      • rambaroo@lemmy.world
        link
        fedilink
        arrow-up
        4
        arrow-down
        1
        ·
        edit-2
        5 months ago

        Producing word salad really isn’t that impressive. At least the art LLMs are somewhat impressive.

        • Meowoem@sh.itjust.works
          link
          fedilink
          arrow-up
          2
          ·
          5 months ago

          If you ask it to make up nonsense and it does it then you can’t get angry lol. I normally use it to help analyse code or write sections of code, sometimes to teach me how certain functions or principles work - it’s incredibly good at that, I do need to verify it’s doing the right thing but I do that with my code too and I’m not always right either.

          As a research tool it’s great at taking a basic dumb description and pointing me to the right things to look for, especially for things with a lot of technical terms and obscure areas.

          And yes they can occasionally make mistakes or invent things but if you ask properly and verify what you’re told then it’s pretty reliable, far more so than a lot of humans I know.