218

We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.

Then retrain on that.

Far too much garbage in any foundation model trained on uncorrected data.

Source.

More Context

Source.

Source.

top 50 comments
sorted by: hot top controversial new old
[-] MummifiedClient5000@feddit.dk 81 points 1 month ago

Isn't everyone just sick of his bullshit though?

[-] mannycalavera@feddit.uk 23 points 1 month ago

US tax payers clearly aren't since they're subsidising his drug habit.

[-] MagicShel@lemmy.zip 21 points 1 month ago* (last edited 1 month ago)

If we had direct control over how our tax dollars were spent, things would be different pretty fast. Might not be better, but different.

load more comments (2 replies)
[-] finitebanjo@lemmy.world 64 points 1 month ago

"If we take this 0.84 accuracy model and train another 0.84 accuracy model on it that will make it a 1.68 accuracy model!"

~Fucking Dumbass

[-] Hupf@feddit.org 11 points 1 month ago

1.68 IQ move

[-] brucethemoose@lemmy.world 47 points 1 month ago* (last edited 1 month ago)

I elaborated below, but basically Musk has no idea WTF he’s talking about.

If I had his “f you” money, I’d at least try a diffusion or bitnet model (and open the weights for others to improve on), and probably 100 other papers I consider low hanging fruit, before this absolutely dumb boomer take.

He’s such an idiot know it all. It’s so painful whenever he ventures into a field you sorta know.

But he might just be shouting nonsense on Twitter while X employees actually do something different. Because if they take his orders verbatim they’re going to get crap models, even with all the stupid brute force they have.

[-] rottingleaf@lemmy.world 39 points 1 month ago

So where will Musk find that missing information and how will he detect "errors"?

[-] 18107@aussie.zone 7 points 1 month ago

I expect he'll ask Grok and believe the answer.

[-] Deflated0ne@lemmy.world 39 points 1 month ago

Dude is gonna spend Manhattan Project level money making another stupid fucking shitbot. Trained on regurgitated AI Slop.

Glorious.

[-] ViatorOmnium@piefed.social 29 points 1 month ago

Because neural networks aren't known to suffer from model collapse when using their output as training data. /s

Most billionaires are mediocre sociopaths but Elon Musk takes it to the "Emperors New Clothes" levels of intellectual destitution.

[-] JackbyDev@programming.dev 25 points 1 month ago

Training an AI model on AI output? Isn't that like the one big no-no?

[-] breecher@sh.itjust.works 11 points 1 month ago

We have seen from his many other comments about this, that he just wants a propaganda bot that regurgitates all of the right wing talking points. So that will definitely be easier to achieve if he does it that way.

load more comments (1 replies)
[-] oliver@lemmy.1984.network 17 points 1 month ago

So "Deleting errors" meaning rewriting history, further fuckin' up facts and definitely sowing hatred and misinformation. Just call it like it is, techbro‘s new reality. 🖕🏻

[-] Elgenzay@lemmy.ml 14 points 1 month ago

Aren't you not supposed to train LLMs on LLM-generated content?

Also he should call it Grok 5; so powerful that it skips over 4. That would be very characteristic of him

[-] Voroxpete@sh.itjust.works 19 points 1 month ago* (last edited 1 month ago)

There are, as I understand it, ways that you can train on AI generated material without inviting model collapse, but that's more to do with distilling the output of a model. What Musk is describing is absolutely wholesale confabulation being fed back into the next generation of their model, which would be very bad. It's also a total pipe dream. Getting an AI to rewrite something like the total training data set to your exact requirements, and verifying that it had done so satisfactorily would be an absolutely monumental undertaking. The compute time alone would be staggering and the human labour (to check the output) many times higher than that.

But the whiny little piss baby is mad that his own AI keeps fact checking him, and his engineers have already explained that coding it to lie doesn't really work because the training data tends to outweigh the initial prompt, so this is the best theory he can come up with for how he can "fix" his AI expressing reality's well known liberal bias.

load more comments (1 replies)
[-] hansolo@lemmy.today 5 points 1 month ago* (last edited 1 month ago)

Musk probably heard about "synthetic data" training, which is where you use machine learning to create thousands of things that are typical-enough to be good training data. Microsoft uses it to take documents users upload to Office365, train the ML model, and then use that ML output to train an LLM so they can technically say "no, your data wasn't used to train an LLM." Because it trained the thing that trained the LLM.

However, you can't do that with LLM output and stuff like... History. WTF evidence and documents are the basis for the crap he wants to add? The hallucinations will just compound because who's going to cross-check this other than Grok anyway?

[-] brucethemoose@lemmy.world 4 points 1 month ago* (last edited 1 month ago)

There’s some nuance.

Using LLMs to augment data, especially for fine tuning (not training the base model), is a sound method. The Deepseek paper using, for instance, generated reasoning traces is famous for it.

Another is using LLMs to generate logprobs of text, and train not just on the text itself but on the *probability a frontier LLM sees in every ‘word.’ This is called distillation, though there’s some variation and complication. This is also great because it’s more power/time efficient. Look up Arcee models and their distillation training kit for more on this, and code to see how it works.

There are some papers on “self play” that can indeed help LLMs.

But yes, the “dumb” way, aka putting data into a text box and asking an LLM to correct it, is dumb and dumber, because:

  • You introduce some combination of sampling errors and repetition/overused word issues, depending on the sampling settings. There’s no way around this with old autoregressive LLMs.

  • You possibly pollute your dataset with “filler”

  • In Musk's specific proposition, it doesn’t even fill knowledge gaps the old Grok has.

In other words, Musk has no idea WTF he’s talking about. It’s the most boomer, AI Bro, not techy ChatGPT user thing he could propose.

load more comments (1 replies)
[-] Lumidaub@feddit.org 13 points 1 month ago* (last edited 1 month ago)

adding missing information

Did you mean: hallucinate on purpose?

Wasn't he going to lay off the ketamine for a while?

Edit: ... i hadnt seen the More Context and now i need a fucking beer or twnety fffffffffu-

[-] Carmakazi@lemmy.world 8 points 1 month ago

He means rewrite every narrative to his liking, like the benevolent god-sage he thinks he is.

[-] ahornsirup@feddit.org 4 points 1 month ago

Let's not beat around the bush here, he wants it to sprout fascist propaganda.

load more comments (1 replies)
[-] namingthingsiseasy@programming.dev 13 points 1 month ago

Whatever. The next generation will have to learn to trust whether the material is true or not by using sources like Wikipedia or books by well-regarded authors.

The other thing that he doesn't understand (and most "AI" advocates don't either) is that LLMs have nothing to do with facts or information. They're just probabilistic models that pick the next word(s) based on context. Anyone trying to address the facts and information produced by these models is completely missing the point.

[-] Kyrgizion@lemmy.world 5 points 1 month ago

Thinking wikipedia or other unbiased sources will still be available in a decade or so is wishful thinking. Once the digital stranglehold kicks in, it'll be mandatory sign-in with gov vetted identity provider and your sources will be limited to what that gov allows you to see. MMW.

Wikipedia is quite resilient - you can even put it on a USB drive. As long as you have a free operating system, there will always be ways to access it.

load more comments (11 replies)
[-] FireWire400@lemmy.world 13 points 1 month ago* (last edited 1 month ago)

How high on ketamine is he?

3.5 (maybe we should call it 4)

I think calling it 3.5 might already be too optimistic

[-] Antaeus@lemmy.world 12 points 1 month ago

Elon should seriously see a medical professional.

[-] LordWiggle@lemmy.world 5 points 1 month ago

He should be locked up in a mental institute. Indefinitely.

[-] NigelFrobisher@aussie.zone 11 points 1 month ago

I figure the whole point of this stuff is to trick people into replacing their own thoughts with these models, and effectively replace consensus reality with nonsense. Meanwhile, the oligarchy will utilise mass data collection via Palantir and ML to power the police state.

load more comments (1 replies)
[-] antihumanitarian@lemmy.world 11 points 1 month ago

Most if not all leading models use synthetic data extensively to do exactly this. However, the synthetic data needs to be well defined and essentially programmed by the data scientists. If you don't define the data very carefully, ideally math or programs you can verify as correct automatically, it's worse than useless. The scope is usually very narrow, no hitchhikers guide to the galaxy rewrite.

But in any case he's probably just parroting whatever his engineers pitched him to look smart and in charge.

[-] Flukas88@feddit.it 10 points 1 month ago

When you think he can't be more of a wanker with an ameba brain.... He surprises you

[-] Saleh@feddit.org 10 points 1 month ago

We have never been at war with Eurasia. We have always been at war with East Asia

[-] MolecularCactus1324@lemmy.world 9 points 1 month ago

iamverysmart

[-] squaresinger@lemmy.world 9 points 1 month ago

First error to correct:

We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing ~~information~~ errors and deleting ~~errors~~ information.

[-] Kurious84@eviltoast.org 8 points 1 month ago

Leme guess. The holocaust was a myth is first on his list.

He should just goto hell early.

load more comments (3 replies)
[-] nieceandtows@programming.dev 7 points 1 month ago

What a loser. He'll keep rewriting it until it fits his world-view

[-] minorkeys@lemmy.world 6 points 1 month ago

We will take the entire library of human knowledge, cleans it, and ensure our version is the only record available.

The only comfort I have is knowing anything that is true can be relearned by observing reality through the lense of science, which is itself reproducible from observing how we observe reality.

[-] ImmersiveMatthew@sh.itjust.works 5 points 1 month ago

Delusional and grasping for attention.

[-] Auli@lemmy.ca 5 points 1 month ago
load more comments (1 replies)
[-] ZoteTheMighty@midwest.social 5 points 1 month ago

I wonder how many papers he's read since ChatGPT released about how bad it is to train AI on AI output.

[-] RizzRustbolt@lemmy.world 5 points 1 month ago

Good thing he's not trying to rewrite the Human Animus, because that's how you end up losing Deimos.

[-] hansolo@lemmy.today 4 points 1 month ago

Prepare for Grokipedia to only have one article about white genocide, then every other article links to "Did you mean White Genocide?"

[-] AI_toothbrush@lemmy.zip 4 points 1 month ago

Lol turns out elon has no fucking idea about how llms work

[-] Knock_Knock_Lemmy_In@lemmy.world 4 points 1 month ago

It's pretty obvious where the white genocide "bug" came from.

[-] sentient_loom@sh.itjust.works 4 points 1 month ago

adding missing information

From where?

Spoiler: He's gonna fix the "missing" information with MISinformation.

[-] ThePowerOfGeek@lemmy.world 4 points 1 month ago

That's not how knowledge works. You can't just have an LLM hallucinate in missing gaps in knowledge and call it good.

load more comments (1 replies)
[-] xenomor@lemmy.world 4 points 1 month ago

Every single endeavor that musk is involved in, is toxic af. He, and all of this businesses, are cancers metastasizing within our society. We really should remove them.

[-] Archangel1313@lemmy.ca 4 points 1 month ago

How does anyone consider him a "genius"? This guy is just so stupid.

load more comments (1 replies)
[-] tonytins@pawb.social 3 points 1 month ago

That's a hell of way to say that he wants to rewrite history.

load more comments
view more: next ›
this post was submitted on 22 Jun 2025
218 points (95.4% liked)

Technology

73290 readers
1280 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS