713
submitted 3 months ago by lseif@sopuli.xyz to c/programmerhumor@lemmy.ml
you are viewing a single comment's thread
view the rest of the comments
[-] stingpie@lemmy.world 7 points 3 months ago

I completely agree that it's a stupid way of doing things, but it is how openai reduced the vocab size of gpt-2 & gpt-3. As far as I know–I have only read the comments in the source code– the conversion is done as a preprocessing step. Here's the code to gpt-2: https://github.com/openai/gpt-2/blob/master/src/encoder.py I did apparently make a mistake, as the vocab reduction is done through a lut instead of a simple mod.

this post was submitted on 12 Jun 2024
713 points (98.0% liked)

Programmer Humor

32045 readers
1336 users here now

Post funny things about programming here! (Or just rant about your favourite programming language.)

Rules:

founded 5 years ago
MODERATORS