60
submitted 6 months ago by misk@sopuli.xyz to c/hardware@lemmy.ml
top 44 comments
sorted by: hot top controversial new old
[-] pezmaker@programming.dev 29 points 6 months ago

I really hate it when questions are put this way. Forget about need. Needs are food, water, shelter, health, etc.

Why can't we just ask, "would anyone really benefit from 1khz gaming display?" Then we can discuss the merits of the subject and whether there's actually a reason to move to it or if it's just more marketing and dick waving points.

It feels like asking if anyone "needs" it is unnecessarily loaded with snark and judgment.

[-] jenny_ball@lemmy.world 2 points 6 months ago
[-] BmeBenji@lemm.ee 22 points 6 months ago
[-] pineapplelover@lemm.ee 15 points 6 months ago

I have a 240hz display and I think it's enough for me

[-] ColeSloth@discuss.tchncs.de 2 points 6 months ago

Type of games I play, I'm fine with my 120 and fps locked at 60.

[-] TwanHE@lemmy.world 15 points 6 months ago

Need, no. Want, YES.

Going from 144 to 280 was great, so if 1k Hz could become the next best thing maybe 480/540 will become affordable.

[-] UnfortunateShort@lemmy.world 13 points 6 months ago

Remember that blind test on LTT where no one, including an esporty player, could reliably differentiate between 120 und 240 Hz?

I know it's not exactly the most thorough experiment ever, but it's all I needed to know.

[-] sudo42@lemmy.world 3 points 6 months ago

Remember "Monster Cables"? Everyone had to have car-battery cables to connect their speakers? Remember when somebody asked "audiophiles" to compare "Monster Cables" and alternatives? Their preferred speaker wire was just a straightened-out metal coat-hanger?

[-] kogasa@programming.dev 2 points 6 months ago* (last edited 6 months ago)

It's extremely easy to tell the difference. I can't tell you what's wrong with their experiment as I don't know exactly what they did, but they clearly fucked it. If you're looking at a static image, you can't differentiate 240Hz from 30Hz. You need a test that actually demonstrates the difference.

[-] UnfortunateShort@lemmy.world 4 points 6 months ago

I have yet to draw the comparison in person (only have a 165Hz), but I mean, every time you double the FPS, the benefit of doubling them again halves. Going from 120 to 240 to 480 Hz is going from 8 to 4 to 2 ms in terms of frametime.

A 4 ms difference in delay might be somewhat noticeable, if you have a very well running game and amazing reflexes. Anything beyond 240 is marketing bs / e-sport 'I need every ms I can get'.

[-] kogasa@programming.dev 1 points 6 months ago

The benefits don't halve. It's the difference between noticing stroboscopic effects and not noticing them. Between not being able to comfortably track fast moving objects and being able to. 1000Hz is a point at which several limitations of LCD technology become invisible.

[-] Chef_Boyardee@lemm.ee 1 points 6 months ago

And I also remember in that same video the esports player performing better at higher refresh rates.

[-] BeigeAgenda@lemmy.ca 11 points 6 months ago

1 kHz is mid tier, real gamers use 1 MHz refresh rate.

[-] elvith@discuss.tchncs.de 7 points 6 months ago

My processor has a boost clock speed of more than 4Ghz - My display just has to keep up with this, otherwise I wouldn't be a real gamer!!!!111oneeleven

[-] jordanlund@lemmy.world 10 points 6 months ago

Considering HDMI 2.1 and Display Port 1.4 can't run that fast...hmmm... no.

[-] ShortN0te@lemmy.ml 7 points 6 months ago

Does anyone really need a 1,000 Hz gaming display?

yes

Does everyone?

no

[-] muhyb@programming.dev 7 points 6 months ago

Wait, we already have 480 Hz displays?

[-] TwanHE@lemmy.world 5 points 6 months ago

540 Hz already.

[-] umbrella@lemmy.ml 4 points 6 months ago

does it matter if we cant afford it

[-] slacktoid@lemmy.ml 6 points 6 months ago

Maybe VR could use this with some combo of hdr and variable refresh rate.

[-] Kerb@discuss.tchncs.de 5 points 6 months ago

i don't
i have a 144hz display, and can barely tell the diffrence between that and 60hz

[-] MrGerrit@feddit.nl 11 points 6 months ago

Do make sure you have set it to 144hz in advance display settings if your running windows.

[-] Kerb@discuss.tchncs.de 5 points 6 months ago

yeah, i changed that.

i only discovered that setting after a few months,
but now it has been on for years.

i can tell the diffrence if i actively look out for ghosting on my mouse coursor.

but unless i actively look out for it,
i dont notice it

[-] stevestevesteve@lemmy.world 5 points 6 months ago

One of the big benefits of 144 and 120 over 60hz display is actually how well they render lower frame rate content. Watching a 24fps (so cinematic!) movie on a 144hz ``display results in a new frame every 6 refreshes (or 5 for 120hz). With a 60hz display, you get an new frame every 2.5 refreshes. Generally this results in judder where every other frame is displayed for longer than the others

[-] fushuan@lemm.ee 1 points 6 months ago

Wouldn't by that logic also be impossible to show 60fps content at 60 fps in a 144 monitor?

[-] stevestevesteve@lemmy.world 1 points 6 months ago

Not impossible at all, but there would be similar judder without some compensation. It's pretty normal for 144hz monitors to support being driven at 60fps, but it's pretty abnormal for a 60hz monitor to advertise 48 or 24fps via edid. Most modern 60+hz TVs are perfectly capable of doing so, though.

Either way, that's one reason I'm very happy with the 240hz wave that seems to be going on. You can display 24 and 60 fps content simultaneously with no judder, as well as even higher frame rate content.

That combined with the popularity of VRR and free/g sync makes me even more optimistic for people to see just about everything the way it was meant to be seen

[-] RedC@sh.itjust.works 5 points 6 months ago

Does anyone need quad core? Does anyone need 1Tb of storage? Does anyone need more than 4GB of ram? Does anyone need...? Is a question we've asked time and time again in the modern history of computers. And the answer is always the same; not right now, but yes

[-] herrvogel@lemmy.world 4 points 6 months ago

Clearly not the same thing. There's no mechanism built into your very physiology that makes you biologically unable to make any meaningful use of anything above a certain amount of computer memory.

[-] You999@sh.itjust.works 4 points 6 months ago

I posted this in another thread but there are some applications where this display technology is actually needed. For example with VR/AR having a 1000Hz display would mean each frame is only displayed for 1ms. Being that quick would mean the headset would be able to better display the micro movements your head and body makes which inturn reduces the disconnect and motion sickness people get with VR/AR.

[-] mojofrododojo@lemmy.world 4 points 6 months ago

yes, but, it reduces the computation cycle for that frame to 1ms.

as a dev, that's daunting.

[-] SkyezOpen@lemmy.world 2 points 6 months ago

90hz is generally enough for most people to not get motion sick. Some headsets do 120 which is like 8ms frame time. Humans can barely detect a flash of light that lasts for that long.

[-] chonglibloodsport@lemmy.world 2 points 6 months ago

You absolutely can tell the difference between 90Hz and 1kHz. Just draw a squiggly line! See this video for a rather dramatic demonstration:

Microsoft Research: High Performance Touch

[-] SoleInvictus@lemmy.blahaj.zone 3 points 6 months ago* (last edited 6 months ago)

This is a demonstration of latency, not frame rate. Did you intend to link something different?

[-] chonglibloodsport@lemmy.world 3 points 6 months ago

A 1000Hz display necessarily has a latency of 1ms between frames. For 100Hz, that’s 10ms.

But this is only the lower bound. You have to include all other sources of latency, such as software, input hardware, drivers, graphics card, etc.

[-] SoleInvictus@lemmy.blahaj.zone 2 points 6 months ago* (last edited 6 months ago)

Ahhh, now I see the connection! It's the update interval. I had to chew on it for a minute but the math checks out.

[-] biber@feddit.de 1 points 6 months ago

The last sentence is simply incorrect. Humans can detect single photons in specific environments. https://www.nature.com/articles/ncomms12172

In real environments it depends very much on the brightness of the flash of light.

[-] Snoopey@lemmy.world 4 points 6 months ago

Getting closer to emulating CRT scan lines

[-] mojofrododojo@lemmy.world 4 points 6 months ago

I'm not sure, but know the tech would be great for vr. Every jump - from 60 dk1 to vive's 90 and index's 120 have been great.

this post was submitted on 21 May 2024
60 points (94.1% liked)

Hardware

5034 readers
1 users here now

This is a community dedicated to the hardware aspect of technology, from PC parts, to gadgets, to servers, to industrial control equipment, to semiconductors.

Rules:

founded 4 years ago
MODERATORS