Well, I think 1) is correct, while 2) 3) and 4) are mostly misleading. The Reasoning Models are still LLMs. Just that they emulate some inner dialogue before writing the final text. I wouldn't use the word "accuracy" in context with AI. It's far from accurate these days, no matter what we do. And in 2), it uses intermediate reasoning. But the Mechanism is still predicting tokens based on statistical likelihood. And regarding 4) also the patterns still come from data, so do the intermediary "logical" steps. And if we look for example at the "thinking" steps of Deepseek R1, it's often unalike human reasoning. I had it reason very illogical steps and outright false things, and despite that, it sometimes arrives at a correct conclusion... Make of this what you will, but I don't think "logical" or "accuracy" are the right words here. Table rows 2, 3 and 4 are oversimplified to the point where it's not correct any more. IMO
I tested out a Deepseek model the other day. It took one minute to generate text and used up all my context space in one message. Local consumer models and "small" server hosted models are probably different classes because for my home pc it was a big performance downgrade.
this post was submitted on 20 Feb 2025
4 points (83.3% liked)
LocalLLaMA
2590 readers
3 users here now
Community to discuss about LLaMA, the large language model created by Meta AI.
This is intended to be a replacement for r/LocalLLaMA on Reddit.
founded 2 years ago
MODERATORS