198
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
this post was submitted on 10 Aug 2025
198 points (100.0% liked)
Technology
74251 readers
816 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
founded 2 years ago
MODERATORS
I doubt it, LLMs have already become significantly more efficient and powerful in just the last couple months.
In a year or two we will be able to run something like Gemini 2.5 Pro on a gaming PC which right now requires a server farm.
Current gen models got less accurate and hallucinated at a higher rate compared to the last ones, from experience and from openai. I think it's either because they're trying to see how far they can squeeze the models, or because it's starting to eat its own slop found while crawling.
https://cdn.openai.com/pdf/2221c875-02dc-4789-800b-e7758f3722c1/o3-and-o4-mini-system-card.pdf