My wife and I started talking about this after she had to help an old lady at the DMV figure out how to use her iPhone to scan a QR code. We’re in our early 40s.

  • nednobbins
    link
    fedilink
    arrow-up
    21
    ·
    11 months ago

    My wife and I regularly joke that one day we’ll harass our kids to help us with our neural interfaces but I don’t think that sort of thing will happen any time soon.

    When I was a kid in the 80’s a lot of people could already afford computers. They weren’t so cheap that everyone had them but they were affordable to a fair number of people if they really wanted one. A C64 cost $595 at launch, that’s under $2,000 in today’s dollars.

    The biggest barrier to computers were that they weren’t “user friendly”. If you wanted to play a simple video game you needed to know some basic command line instructions. When I wanted to set up my first mouse for my 8086 it involved installing drivers and editing config.sys and autoexec.bat. You couldn’t really do anything with a computer those days unless you were willing to nerd out.

    At the same time, nerding out on a computer could easily get you deep into the guts of your computer in a functional way. I learned that the only way I could play video games at night was if I opened up the computer and disconnected the speaker wire so it wouldn’t alert my parents. I also learned that I could “hack” Bards Tale by opening up the main file with debug and editing it so that the store would sell an infinite number of “Crystal Swords”.

    Today there are 2 cell phones for every human on earth. Kids walk around with supercomputers in their pockets. But they’ve become so “user friendly” that you barely even need to be literate to operate one. That’s generally a good thing but it removes an incentive to figuring out how the stuff works. Most people only bother with that if they’re having some trouble getting it working in the first place.

    At the same time it’s gotten much harder to make changes to your computer. The first Apple was a pile of circuits you needed to solder together. You can’t even remove the battery on a modern one (without jumping through a lot of hoops). If you edit some of your games it’s more likely to trigger some piracy or cheat protection than to let you actually change it.

    There are still large communities of computer nerds but your average person today basically treats computers like magic boxes.

    I’d expect that kind of gap in other areas. I’d take 3d printing as an example. You can get one now for a few hundred bucks. They’re already used in industry but, at this point, they’re still very fiddly. The people who have them at home are comfortable doing stuff like troubleshooting, flashing ROMs, wading through bad documentation and even printing custom upgrades for their printer.

    • drphungky@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      11 months ago

      My wife and I regularly joke that one day we’ll harass our kids to help us with our neural interfaces

      This but unironically.

      Seriously, there will be new interfaces in the next 20 years. People always underestimate tech change and growth, and we already see VR and AR in their infancy. Do you remember what it was like watching your parents or grandparents hunt and peck type, or struggle to double click something, or double clicking things that don’t need double clicked? Did you struggle with helping them Google their problem (back when Google was useful)? There will always be luddites and people who don’t adopt new tech, but even among those who do, they’re often slower or just have a less intuitive understanding of newer tech. This will happen to us. Even simple motion controls while in AR will likely be hard for people to pick up and develop new muscle memory. Neural interfaces will likely require you to “think a certain way” to best interact with things, and I don’t doubt many of us will be bad at it. And most likely of all, we’ll be bad at something we don’t even predict - and many people won’t care that much. I’d argue in some cases it actuallystarts with “what’s the use?” which tons of millennials have already done with Twitter, Instagram, Tiktok etc. Even if you recognize the value, you’re comfortable and happy not using the new thing, and that’s a double whammy when combined with the effort it takes to learn new tech.

      • nednobbins
        link
        fedilink
        arrow-up
        1
        ·
        11 months ago

        I think you’re sort of right but it will depend heavily on how radical a shift the new technology is. In order for there to be this kind of divide there needs to be a steep learning curve to the technology. People are only willing to put up with those learning curves if there’s a significant advantage. That means that manufacturers can only successfully market “difficult” technologies if they provide a big advantage.

        I’m not aware of any old people having difficulty transitioning from quills to, fountain pens to ball point pens. They all basically did the same thing and you only had to make minor adjustments. Nobody bothered learning how to use the Writer since it didn’t actually let you do anything better. They were willing to go through the significant curve of learning how to use typewriters because, once they did, they could write significantly faster.

        Computers and cell phones are a whole different way of interacting with people and information than “hardcopy” was. You didn’t just swap some objects that did the same thing with a different approach. It wasn’t even just a slightly different way of doing the same thing. Those technologies allowed us to interact with the world in a totally new way. It was worth learning a bunch of weird computer stuff that older generations had never heard of because we could do things they never dreamed of. (eg I used to get rushed when talking with my grandmother to save on long distance bills, now I don’t even think about long distance costs other than latency.

        I’m sure that sort of thing will happen again but it would require a far more disruptive technology than AR. That’s a small iteration that we’ve already been primed for. When Terminator 1 came out, nobody was confused when it switched to “terminator vision” and you saw the AR display. That’s why I joke about neural interfaces. In theory, that could give a person significantly higher throughput rates to their computer. There are all kinds of potential benefits to. It would be worth it for people to put up with steep learning curves, unintuitive interfaces and lots of troubleshooting if it meant they could suddenly “read” at 10,000 words a minute or control complex robots. Not everyone would go through that effort and it would create the kinds of divides that we saw with computers.

        When I look at current technologies as an old(ish) person, it’s a very different view than my parents and grandparents had. They didn’t understand the new technologies. I have no trouble understanding them, I just think a lot of them are a waste of my time (unlike screwing around on Lemmy, which is totally productive /s).