438
submitted 2 weeks ago by misk@sopuli.xyz to c/technology@lemmy.world
you are viewing a single comment's thread
view the rest of the comments
[-] Blackmist@feddit.uk 61 points 2 weeks ago

Not really sure who this is for. With soldered RAM is less upgradeable than a regular PC.

AI nerds maybe? Sure got a lot of RAM in there potentially attached to a GPU.

But how capable is that really when compared to a 5090 or similar?

[-] brucethemoose@lemmy.world 39 points 2 weeks ago* (last edited 2 weeks ago)

The 5090 is basically useless for AI dev/testing because it only has 32GB. Mind as well get an array of 3090s.

The AI Max is slower and finicky, but it will run things you'd normally need an A100 the price of a car to run.

But that aside, there are tons of workstations apps gated by nothing but VRAM capacity that this will blow open.

[-] KingRandomGuy@lemmy.world 16 points 2 weeks ago

Useless is a strong term. I do a fair amount of research on a single 4090. Lots of problems can fit in <32 GB of VRAM. Even my 3060 is good enough to run small scale tests locally.

I'm in CV, and even with enterprise grade hardware, most folks I know are limited to 48GB (A40 and L40S, substantially cheaper and more accessible than A100/H100/H200). My advisor would always say that you should really try to set up a problem where you can iterate in a few days worth of time on a single GPU, and lots of problems are still approachable that way. Of course you're not going to make the next SOTA VLM on a 5090, but not every problem is that big.

[-] brucethemoose@lemmy.world 1 points 1 week ago

Fair. True.

If your workload/test fits in 24GB, that's already a "solved" problem. If it fits in 48GB, it's possibly solved with your institution's workstation or whatever.

But if it takes an 80GB A100, you are likely using very expensive cloud GPU time. I really love the idea of being able to tinker with a "full" 80GB+ workload (even having to deal with ROCM) without having to pay per hour.

[-] KeenFlame@feddit.nu 1 points 2 weeks ago

Exactly, 32 is plenty to develop on, and why would you need to upgrade ram? It was years ago I did that in any computer let alone a tensor workstation. I feel like they made pretty good choices for what it's for

[-] felixwhynot@lemmy.world 2 points 2 weeks ago

… but only OpenCL workloads, right?

[-] amon@lemmy.world 2 points 2 weeks ago

No, it runs off integrated graphics, which is a good thing because you can have a large capacity of ram dedicated to GPU loads

[-] MonkderVierte@lemmy.ml 1 points 2 weeks ago

Not really sure who this is for.

Second sentence in the linked article.

this post was submitted on 25 Feb 2025
438 points (98.7% liked)

Technology

65882 readers
1287 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS