"Even within the coding, it's not working well," said Smiley. "I'll give you an example. Code can look right and pass the unit tests and still be wrong. The way you measure that is typically in benchmark tests. So a lot of these companies haven't engaged in a proper feedback loop to see what the impact of AI coding is on the outcomes they care about. Lines of code, number of [pull requests], these are liabilities. These are not measures of engineering excellence."
Measures of engineering excellence, said Smiley, include metrics like deployment frequency, lead time to production, change failure rate, mean time to restore, and incident severity. And we need a new set of metrics, he insists, to measure how AI affects engineering performance.
"We don't know what those are yet," he said.
One metric that might be helpful, he said, is measuring tokens burned to get to an approved pull request – a formally accepted change in software. That's the kind of thing that needs to be assessed to determine whether AI helps an organization's engineering practice.
To underscore the consequences of not having that kind of data, Smiley pointed to a recent attempt to rewrite SQLite in Rust using AI.
"It passed all the unit tests, the shape of the code looks right," he said. It's 3.7x more lines of code that performs 2,000 times worse than the actual SQLite. Two thousand times worse for a database is a non-viable product. It's a dumpster fire. Throw it away. All that money you spent on it is worthless."
All the optimism about using AI for coding, Smiley argues, comes from measuring the wrong things.
"Coding works if you measure lines of code and pull requests," he said. "Coding does not work if you measure quality and team performance. There's no evidence to suggest that that's moving in a positive direction."
This isn't the "very beginning", that was either 70 or 120 years ago, depending on whether you're counting from the formalization of "AI" as an academic discipline with the advent of the Markov Decision Process or the earlier foundational work on Markov Chains.
Chatbots are old-hat, I was playing around with Eliza back in the 90's. Hell, even Large Language Models aren't new, the transformer architecture they're based on is almost 10 years old and itself merely a minor evolution of earlier statistical and recurrent neural network language processing models. By the time big tech started ramping up the "AI" bubble in 2024, I had already been bored with LLMs for two years.
There's no "early adaptation" here, just a rushed and wildly excessive implementation of a very interesting but fundamentally untrustworthy tech with no practical value proposition for the people it is nevertheless being sold to.
It’s the beginning of AI in terms of where it will be.
Could you try rephrasing that in a way that makes sense?
You understand it.
No, I'm afraid I don't.
The beginning of the development of "AI" is temporal, not spatial, unless you are referring to the path of development which, for no obvious reason, you refuse to trace backwards as well as forwards.
︋︆︆︅︌︈︄︂︆︄︃︃︈︄︄︊︎︃︆︀︆︌︉︌︈︍︋︈︇︊︁︄︆Y︄︄︀︇︈︁︀︈︅︍︂︂︄︉︎︊︌︌︀︂︋︃о︆︆︄︍︄︀︇︈︎︇︆︁︍︉︍︌︎︌︅︈︋︁︅︆u︄︃︅︎︎︅︁︋︃︆︈︃︈︄︋︇︅︃︎︂︎︄︊︆︂︇︋’︇︄︀︃︂︊︁︉︅︁︃︁︎︀︇︁︁︇︅︂︂︊︋︇︄︁l︁︍︄︋︈︌︄︌︅︋︉︊︍︍︃︉︈︇︇︎︈︉︁︍︈︋︉l︌︀︄︊︊︅︈︈︍︉︊︋︅︁︉︋︉︅︋︉︇︎︋︄︆︌︄︁︈ ︈︃︋︈︌︀︈︎︀︂︉︄︅︊︋︈︈︀︈︆︇︎︊︁g︍︇︀︀︎︂︍︀︂︋︀︉︉︃︆︊︄︌︉︈︈︎︎︈︍︉︃︂︊︂︁︃︃︈︎︋е︁︂︆︁︃︆︄︍︃︄︅︉︎︍︇︈︌︄︅︄t︃︇︈︁︈︋︆︄︈︅︁︊︀︄︄︌︃︈︄︇︍︁ ︌︌︁︂︁︂︈︍︄︅︀︊︍︁︊︎︉︎︊︂︆︎︋︄︂︋︂︂︈︃i︁︊︃︁︌︇︇︊︉︈︋︅︀︂︅︁︌︄︉︊︎︅︊︀︆︂︋︆︍︅︆︋︆︂︃︈︌︂︋t︌︅︉︍︅︋︆︊︃︋︆︂︎︅︎︍︄︋︆︎︋︀︆ ︀︉︍︍︆︃︈︋︀︋︍︂︈︁︀︂︄︌︁︉︍︄︊е︎︌︂︆︊︊︌︍︄︈︄︉︄︌︎︌︅︋︀︆︄︉︃︁︇︌︊v︇︀︍︆︁︁︌︆︇︌︊︃︆︍︇︉︈︁︋︈︁︂︁︊︁︁︎︆︎︎︉︆е︌︄︉︈︄︌︉︈︀︃︆︎︈︉︀︎︍︌︁︄︄︅︁︌︋︇︊︃︇︋︃︉︉n︌︇︆︇︉︋︉︄︄︌︎︁︃︅︁︆︋︉︁︅︀︉︎︎︇︋︌︉t︄︈︅︎︋︊︋︋︊︉︄︍︂︅︌︊︆︅︁︅︋︇︃︍u︀︌︈︌︉︃︋︇︈︇︊︀︎︈︈︇︍︊︃︄︀︉︍︅︍а︀︁︄︁︌︍︅︉︅︁︇︃︍︉︀︂︋︍︌︆︍︎︌︀︀︇︉︆︉︇l︉︌︀︋︇︄︅︅︈︊︌︍︊︍︀︉︎︃︎︁︃︌︇l︆︈︍︎︌︁︂︃︂︄︈︍︀︎︊︀︀︉︉︄︂︍︃︋у︄︅︈︌︀︅︅︀︁︍︎︋︁︋︌︋︄︅︅︅︉︈︍︄︈︎︃︂︂︌︇︅︉︌︀︀
If I'm not getting it immediately then you're communicating your point ineffectively.
What, precisely, do you mean when you assert that the last three to six generations of work on "AI" don't count?
I’m not here to talk in kindergarten sentences.
And yet, you just posted one.
And hey it deserves some praise for operating at a level far above its native ability. Let's be honest, Org is something of a dichotomy, both a rambling idiot, and an example of the highest capabilities of an AI chud. Its parents must be so proud.