Yes! This is a brilliant explanation of why language use is not the same as intelligence, and why LLMs like chatGPT are not intelligence. At all.

  • FaceDeer@kbin.social
    link
    fedilink
    arrow-up
    18
    arrow-down
    8
    ·
    1 year ago

    A lot of people are deeply invested in the notion that human intelligence is unique and special and impossible to replicate. Either their personal sense of worth is bound up in that notion (see for example many of the artists who get very angry when people call AI generated images “art”) or it’s simply a threat to their jobs and economic wellbeing. The result is a powerful need to convince themselves that there’s a special something that’s missing from ChatGPT and its ilk that will “never” be replicated by machines.

    It’s true that ChatGPT isn’t intelligent in the same way that human brains are intelligent. But it is intelligent, in ways that are useful. And “never” is a bad bet to make for the rest of those capabilities.

    • rob64@startrek.website
      link
      fedilink
      arrow-up
      4
      arrow-down
      1
      ·
      edit-2
      1 year ago

      My sense in reading the article was not that the author thinks artificial general intelligence is impossible, but that we’re a lot farther away from it than recent events might lead you to believe. The whole article is about the human tendency to conflate language ability and intelligence, and the author is making the argument both that natural language does not imply understanding of meaning and that those financially invested in current “AI” benefit from the popular assumption that it does. The appearance or perception of intelligence increases the market value of AIs, even if what they’re doing is more analogous to the actions of a very sophisticated parrot.

      Edit all of which is to say, I don’t think the article is asserting that true AI is impossible, just that there’s a lot more to it than smooth language usage. I don’t think she’d say never, but probably that there’s a lot more to figure out—a good deal more than some seem to think—before we get Skynet.

      • Spzi
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        The whole article is about the human tendency to conflate language ability and intelligence, and the author is making the argument both that natural language does not imply understanding of meaning and that those financially invested in current “AI” benefit from the popular assumption that it does.

        That’s a good summary.

    • kaffiene@lemmy.world
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      6
      ·
      1 year ago

      Chatgpt is not intelligent. Not in the sense where we use that word anywhere else, including the animal kingdom. Transformer is an extraordinary clever and sophisticated algorithm, thou

      • FaceDeer@kbin.social
        link
        fedilink
        arrow-up
        8
        arrow-down
        3
        ·
        1 year ago

        As I said:

        It’s true that ChatGPT isn’t intelligent in the same way that human brains are intelligent.

        There isn’t just one kind of intelligence.

      • FaceDeer@kbin.social
        link
        fedilink
        arrow-up
        10
        arrow-down
        4
        ·
        1 year ago

        Based on its abilities. Based on studies that researchers have done on it. It has learned to do more than just regurgitate bits of its training material. It has learned from the patterns in it and can extrapolate new information.

        Do you think this is not possible?