255
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
this post was submitted on 07 Feb 2024
255 points (96.4% liked)
Games
16822 readers
1119 users here now
Video game news oriented community. No NanoUFO is not a bot :)
Posts.
- News oriented content (general reviews, previews or retrospectives allowed).
- Broad discussion posts (preferably not only about a specific game).
- No humor/memes etc..
- No affiliate links
- No advertising.
- No clickbait, editorialized, sensational titles. State the game in question in the title. No all caps.
- No self promotion.
- No duplicate posts, newer post will be deleted unless there is more discussion in one of the posts.
- No politics.
Comments.
- No personal attacks.
- Obey instance rules.
- No low effort comments(one or two words, emoji etc..)
- Please use spoiler tags for spoilers.
My goal is just to have a community where people can go and see what new game news is out for the day and comment on it.
Other communities:
founded 1 year ago
MODERATORS
I love that the game is such a CPU hogging mess that LTT used it to test over clocking a brand new AMD thread ripper and the game still ran like garbage even on one of the fastest and most multithreaded CPUs that exist.
I love Cities Skylines but whatever is happening in 2 is a three alarm fire and needs to be fixed.
I imagine LTT did that for meme purposes more than anything else. Threadrippers are not built for games. They're built for production workloads which don't translate to gaming performance.
That said, the point still stands. This game needs the most powerful gaming hardware (e.g. Ryzen X3D series and RTX 4090) on "recommended" settings and 1080p to get averages above 60fps, which is wild. There's a rather dedicated fellow on reddit who does detailed performance tests after each patch.
So very fucking glad I haven't bought this game.
I bought it for my girlfriend’s birthday and had to go through and refund it because of just how poorly the game ran even with everything set to minimum.
I got ok performance out of it on a 1660.
Are you on a potato?.
My system is 8 years old and it plays this game just fine. Granted I am not running 4K. I am still on 60 Hertz monitors. I also haven't gotten very far into the game so any population over 30k I have not experienced.
They did it because the developers said the game will use however many cores you can give it. And i mean, yeah it maxed out all cores. Likely doing nothing but struggling to keep them synchronized but it was using em
What are some characteristics of modern, multi-threaded games that don't match up to production workloads as far as the CPU is concerned? What do you consider a production workload? How does it differ from CS2's simulation system?
lol got a link to the video? That sounds hilarious and worth a watch.
https://youtu.be/R83W2XR3IC8?si=nTUMXFiFGFRcdtQa
Jump to the 3 minute mark.
Here is an alternative Piped link(s):
https://piped.video/R83W2XR3IC8?si=nTUMXFiFGFRcdtQa
Piped is a privacy-respecting open-source alternative frontend to YouTube.
I'm open-source; check me out at GitHub.
Much appreciated!
The game when it saw that CPU:
It seems like we have more power than we know what do do with.
That means we’re not cutting it close enough!
Edit: I don’t remember the exact quote but y’all get it.
Not sure why LTT or anyone else would have thought that would even help considering simulation games like that rely heavily on single core performance.
I mean... Watch the video? It uses 64 fucking cores when available. It's a heavily multithreaded game.
CS2 uses multiple cores for… something, but it’s a Unity game and there’s only so much you can do to avoid dependence on a main thread. Your single core perforemance is still going to be a limiting factor.
CS2 uses a design paradigm called Entity Component System, which allows for aggressive multi core utilization by splitting up game logic into self contained "systems" that operate on a subset of "Components" per "Entity". This allows for data dependencies to be statically analyzed and a scheduler to maximize CPU Utilization thanks to the better separated workflows.
It uses DOTS from Unity to accomplish this. There is a small bottleneck in communicating this work back to the game's renderer, but it is doing a lot of valuable work with all those cores.The communication with the renderer and their rendering implementation sucks right now and thats where the performance tanks.
I am very aware of how at some level there are less multicore workloads involved but a CPU core can do a metric shitload of work, it's the RAM and GPU transfers that kill performance. We dont need to blame Unity here, they are fucking this up 100% themselves.
Theres a video that explains all this but I cant find it and thats pretny annoying so whatever.
Wasn't most of the frame latency caused by shaders in graphics? There was a deep dive video but i forgot the title and YouTuber