Speaking as a designer, it’s important to separate the style/trend of a UI from its function. I think what you’re looking for is actually UX design.
As a discipline, User Experience uses evidence-based research to understand how and why users behave they do. This leads to specific design patterns and principles that underlie all the good UI design seen from the giants like Apple, Google, Microsoft, etc. It gives you the language to evaluate designs. This is the foundation of your UI and the rest is just style — fonts, colors, imagery and icons which is subjective and less important. I lost ambition to be a trendy UI designer, so every design looks the same, but usability will shines through. Clean, simple and accessible is timeless.
Study the articles from nngroup.com. They pretty much established the field of UX Design, with content talking about user behavior in the 1990s. https://lawsofux.com is a more attractive and consumable option, also heavily influenced by NN Group. Finally, accessible design is good design for all, not just those with disabilities. Understand the guidelines set by the W3C for accessibility, like minimum font sizes or contrast ratios for colors.
First, we don’t understand our own neurons enough to model them.
AI’s “neuron” or node is a math equation that takes a numeric input with a variable “weight” that affects the output. An actual neuron a cell with something like 6000 synaptic connections each and 600 trillion synapses total. How do you simulate that? I’d argue the magic of AI is how much more efficient it is comparatively with only 176 billion parameters in GPT4.
They’re two fundamentally different systems and so is the resulting knowledge. AI doesn’t need to learn like a baby, because the model is the brain. The magic of our neurons is their plasticity and our ability to freely move around in this world and be creative. AI is just a model of what it’s been fed, so how do you get new ideas? But it seems that with LLMs, the more data and parameters, the more emergent abilities. So we just need to scale it up and eventually we can raise the.
AI does pretty amazing and bizarre things today we don’t understand, and they are already using giant expensive server farms to do it. AI is super compute heavy and require a ton of energy to run. So, the cost is a rate limiting the scale of AI.
There are also issues related to how to get more data. Generative AI is already everywhere and what good s is it to train on its own shit? Also, how do you ethically or legally get that data? Does that data violate our right to privacy?
Finally, I think AI actually possess an intelligence with an ability to reason, like us. But it’s fundamentally a different form of intelligence.