cross-posted from: https://lemmy.ml/post/20858435

Will AI soon surpass the human brain? If you ask employees at OpenAI, Google DeepMind and other large tech companies, it is inevitable. However, researchers at Radboud University and other institutes show new proof that those claims are overblown and unlikely to ever come to fruition. Their findings are published in Computational Brain & Behavior today.

  • JackOverlord@beehaw.org
    link
    fedilink
    arrow-up
    9
    ·
    12 hours ago

    Whenever I hear someone say that something is impossible with current technology, I think about my grandma. When she was a kid, only some important people had telephones. Doctors, police, etc.

    In her lifetime we went from that to today, and, since she’s still alive, even further into the future.

    Whenever someone calls something impossible, I think about how far technology will progress in my own lifetime and I know that they’ve got no idea what they’re talking about. (Unless, like you said, it’s against the laws of physics. But sometimes even then I’m not so sure, cause it’s not like we understand those entirely. )

    • DdCno1@beehaw.org
      link
      fedilink
      arrow-up
      5
      ·
      9 hours ago

      The thing is, we have no idea where technological progress is taking us. So far, most predictions have been wrong. 50 to 60 years ago, people thought we would already be colonizing other planets by now. Barely anyone was able to predict the Internet, smartphones, social media, etc. - the kind of technology that is actually shaping our civilization’s future right now.

      Another aspect that I feel is often neglected is the assumption that technological progress will continue forever or at least continue at this current rapid pace. This wasn’t true in the past and we might simply be experiencing a historical anomaly right now, one that could correct itself very soon in the future, either towards stagnation or even regression.

      • d0ntpan1c@lemmy.blahaj.zone
        link
        fedilink
        arrow-up
        1
        ·
        1 hour ago

        The space example is extremely apt. Its possible we could have had tons of space stations, a moon colony, maybe even some other stuff going on around the solar system, asteroid mining, etc. But thay would have at least required the space race to continue longer and for spending to grow to create a big enoigh industry to ensure thay outcome, assuming no capacity or time issue. Alas, we took another path.

        Something that seems important to us might not matter in even 10 years, or at least, not have a monetary and/or societal incentive to keep advancing.

        • DdCno1@beehaw.org
          link
          fedilink
          arrow-up
          1
          ·
          edit-2
          59 minutes ago

          I was also based on the assumption that the rapid progress of aerospace technology that happened in the 1920s to 1960s would continue onward at the same pace, whereas what actually happened was that barriers emerged that nobody was able to circumvent, like for example engineering things to withstand incredibly abrasive Moon dust (or really do anything productive on that lifeless rock), how to deal with the endless pitfalls of a long Mars journey, how to bring down the cost of launch vehicles so that grand projects like giant space stations would even be remotely possible (von Braun was already thinking about huge space stations all the way back in 1945). Many of these issues couldn’t simply be solved by throwing more money at them, which is important. Deciders, both in Washington and Moscow, were smart enough to realize this in the 1970s, for the most part at least (the Space Shuttle and its Soviet clone, each a gigantic waste of money, are major counter example from this era).

          The point I’m making here is that everyone assumed linear progress in this area, just like there are people currently making many billion dollar bets on linear progress in regards to computer technology in general and AI in particular, but at least, with the benefit of hindsight given past examples, there’s a reasonable amount of doubt this time around.

      • sydneybrokeit@beehaw.org
        link
        fedilink
        arrow-up
        1
        ·
        6 hours ago

        This wasn’t true in the past and we might simply be experiencing a historical anomaly right now

        While our exact pacing might be slightly different from the pure extrapolation, human history has been a long, steady increase in the rate of invention. Access to education has meant that more people are making things, and then the next generations build on top of their work to make even bigger things.

    • Saganaki@lemmy.one
      link
      fedilink
      arrow-up
      8
      arrow-down
      1
      ·
      11 hours ago

      That’s not an apt comparison.

      More like “we’ll have flying cars 50 years from now.”

      • megopie@beehaw.org
        link
        fedilink
        arrow-up
        3
        ·
        7 hours ago

        I love the flying car example because it reveals a huge issue with the whole “tech will get better” idea. People are still trying to make flying cars happen but it’s running in to the same fundamental issues; large things that are mechanically complex, energy intensive, and moving at high speeds in a crowded urban environments are just too expensive and dangerous.

        There is no way around the physical realities, no clever trick or efficiency that will push it over some threshold of practicality.

    • massive_bereavement@fedia.io
      link
      fedilink
      arrow-up
      6
      ·
      12 hours ago

      Let’s put it this way: If in our lifetime we can simulate the intelligence of a vinegar fly as general intelligence, that would be a monumental landmark in AGI. And we’re far, far, far away from it.

      As far as the iron age was from the metal alloys used in the Space Shuttle.

      Talking about AGI simulating higher intelligence at the level of a dog or a cat, dear I say a pigeon or a crow is as far fetched as expecting ancient Egyptians to harness the power of the atom.

      • JackOverlord@beehaw.org
        link
        fedilink
        arrow-up
        3
        ·
        edit-2
        11 hours ago

        Let’s put it this way: If in our lifetime we can simulate the intelligence of a vinegar fly as general intelligence, that would be a monumental landmark in AGI. And we’re far, far, far away from it.

        I get what you mean here and I agree with it, if we’re talking about current “AI”, which isn’t anywhere close. I know, because I’ve programmed some simple “AIs” (Mainly ML models) myself.

        But your comparison to ancient egypt is somewhat lacking, considering we had the aptly named dark ages between then and now.

        Lot’s of knowledge got lost all the time during humanity’s history, but ever since the printing press, and more recently the internet, came into existence, this problem has all but disappeared. As long as humanity doesn’t nuke itself back to said dark ages, I recon we aren’t that far away from AGI, or at least something close to it. Maybe not in my lifetime, but another ~2000 years seems a little extreme.

    • TranquilTurbulence@lemmy.zip
      link
      fedilink
      arrow-up
      2
      ·
      edit-2
      10 hours ago

      Could take a while, but how long? Progress tends to be non-linear, so things can slow down and speed up suddenly. I’m pretty sure we’ll get there sooner or later unless we nuke ourselves to oblivion before that.

      If AI development isn’t prioritized, it could take centuries. Maybe we’re still missing some crucial corner stores we haven’t even thought of yet. Just imagine what it was like to build an airplane in an age when the internal combustion engine hadn’t been invented yet. Maybe we’re still missing something that big. On the other hand, it could also be just around the corner, but I find it unlikely.