view the rest of the comments
Ask Lemmy
A Fediverse community for open-ended, thought provoking questions
Please don't post about US Politics. If you need to do this, try !politicaldiscussion@lemmy.world
Rules: (interactive)
1) Be nice and; have fun
Doxxing, trolling, sealioning, racism, and toxicity are not welcomed in AskLemmy. Remember what your mother said: if you can't say something nice, don't say anything at all. In addition, the site-wide Lemmy.world terms of service also apply here. Please familiarize yourself with them
2) All posts must end with a '?'
This is sort of like Jeopardy. Please phrase all post titles in the form of a proper question ending with ?
3) No spam
Please do not flood the community with nonsense. Actual suspected spammers will be banned on site. No astroturfing.
4) NSFW is okay, within reason
Just remember to tag posts with either a content warning or a [NSFW] tag. Overtly sexual posts are not allowed, please direct them to either !asklemmyafterdark@lemmy.world or !asklemmynsfw@lemmynsfw.com.
NSFW comments should be restricted to posts tagged [NSFW].
5) This is not a support community.
It is not a place for 'how do I?', type questions.
If you have any questions regarding the site itself or would like to report a community, please direct them to Lemmy.world Support or email info@lemmy.world. For other questions check our partnered communities list, or use the search function.
Reminder: The terms of service apply here too.
Partnered Communities:
Logo design credit goes to: tubbadu
This but unironically.
Seriously, there will be new interfaces in the next 20 years. People always underestimate tech change and growth, and we already see VR and AR in their infancy. Do you remember what it was like watching your parents or grandparents hunt and peck type, or struggle to double click something, or double clicking things that don't need double clicked? Did you struggle with helping them Google their problem (back when Google was useful)? There will always be luddites and people who don't adopt new tech, but even among those who do, they're often slower or just have a less intuitive understanding of newer tech. This will happen to us. Even simple motion controls while in AR will likely be hard for people to pick up and develop new muscle memory. Neural interfaces will likely require you to "think a certain way" to best interact with things, and I don't doubt many of us will be bad at it. And most likely of all, we'll be bad at something we don't even predict - and many people won't care that much. I'd argue in some cases it actuallystarts with "what's the use?" which tons of millennials have already done with Twitter, Instagram, Tiktok etc. Even if you recognize the value, you're comfortable and happy not using the new thing, and that's a double whammy when combined with the effort it takes to learn new tech.
I think you're sort of right but it will depend heavily on how radical a shift the new technology is. In order for there to be this kind of divide there needs to be a steep learning curve to the technology. People are only willing to put up with those learning curves if there's a significant advantage. That means that manufacturers can only successfully market "difficult" technologies if they provide a big advantage.
I'm not aware of any old people having difficulty transitioning from quills to, fountain pens to ball point pens. They all basically did the same thing and you only had to make minor adjustments. Nobody bothered learning how to use the Writer since it didn't actually let you do anything better. They were willing to go through the significant curve of learning how to use typewriters because, once they did, they could write significantly faster.
Computers and cell phones are a whole different way of interacting with people and information than "hardcopy" was. You didn't just swap some objects that did the same thing with a different approach. It wasn't even just a slightly different way of doing the same thing. Those technologies allowed us to interact with the world in a totally new way. It was worth learning a bunch of weird computer stuff that older generations had never heard of because we could do things they never dreamed of. (eg I used to get rushed when talking with my grandmother to save on long distance bills, now I don't even think about long distance costs other than latency.
I'm sure that sort of thing will happen again but it would require a far more disruptive technology than AR. That's a small iteration that we've already been primed for. When Terminator 1 came out, nobody was confused when it switched to "terminator vision" and you saw the AR display. That's why I joke about neural interfaces. In theory, that could give a person significantly higher throughput rates to their computer. There are all kinds of potential benefits to. It would be worth it for people to put up with steep learning curves, unintuitive interfaces and lots of troubleshooting if it meant they could suddenly "read" at 10,000 words a minute or control complex robots. Not everyone would go through that effort and it would create the kinds of divides that we saw with computers.
When I look at current technologies as an old(ish) person, it's a very different view than my parents and grandparents had. They didn't understand the new technologies. I have no trouble understanding them, I just think a lot of them are a waste of my time (unlike screwing around on Lemmy, which is totally productive /s).