1298
"prompt engineering"
(lemmy.world)
Welcome to Programmer Humor!
This is a place where you can post jokes, memes, humor, etc. related to programming!
For sharing awful code theres also Programming Horror.
I don't necessarily disagree that we may figure out AGI, and even that LLM research may help us get there, but frankly, I don't think an LLM will actually be any part of an AGI system.
Because fundamentally it doesn't understand the words it's writing. The more I play with and learn about it, the more it feels like a glorified autocomplete/autocorrect. I suspect issues like hallucination and "Waluigis" or "jailbreaks" are fundamental issues for a language model trying to complete a story, compared to an actual intelligence with a purpose.