We definitely are seeing things faster than 24 Hz, or we wouldn’t be able to tell a difference in refresh rates above that.
Edit: I don’t think we have a digital, on-off refresh rate of our vision, so fps doesn’t exactly apply. Our brain does turn the ongoing stream of sensory data from our eyes into our vision “video”, but compared to digital screen refresh rates, we can definitely tell a difference between 24 and say 60 fps.
People looking at a strobing light, start to see it as just “on” (not blinking anymore) at almost exactly 60Hz.
In double blind tests, pro gamers can’t reliably tell 90fps from 120.
There is however, an unconscious improvement to reaction time, all the way up to 240fps. Maybe faster.
However, when the modulated light source contains a spatial high frequency edge, all viewers saw flicker artifacts over 200 Hz and several viewers reported visibility of flicker artifacts at over 800 Hz. For the median viewer, flicker artifacts disappear only over 500 Hz, many times the commonly reported flicker fusion rate.
The real benefit of super high refresh rates is the decrease in latency for input. At lower rates the lag between input and the next frame is extremely apparent, above about ~144hz it’s much less noticable.
The other side effect of running at high fps is that when heavy processing occurs and there are frame time lags they’re much less noticable because the minimum fps is still very high. I usually tell people not to pay attention to the maximum fps rather look at the average and min.
I think having higher frame rates isn’t necessarily about whether our eyes can perceive the frame or not. As another commenter pointed out there’s latency benefits, but also, the frame rate affects how things smear and ghost as you move them around quickly. I don’t just mean in gaming. Actually, it’s more obvious when you’re just reading an article or writing something in Word. If you scroll quickly, the words blur and jitter more at low frame rates, and this is absolutely something you can notice. You might not be able to tell the frametime, but you can notice that a word is here one moment and next thing you know, it teleported 1 cm off
I think i read that fighter pilots need to be able to identify a plane in one frame at 300 fps, and that the theoretical limit of the eye is 1000+ fps.
Though, whether the brain can manage to process the data at 1000+ fps is questionable.
Both of these claims are kinda misguided. The brain is able to detect very short flashes of light (say, 1 thousandth of a second), and other major changes in light perception. Especially an increase in light will be registered near instantly. However, since it doesn’t have a set frame rate, more minor changes in the light perception (say, 100 fps) are not going to be registered. And the brain does try to actively correct discontinuities, that’s why even 12 fps animation feels like movement, although a bit choppy.
You’re right that is a continuous process, so there’s no frame rate as such. 24 fps is just the lowest framerate that doesn’t look “framey” in videos, but can go as low as 12 and still reliably perceive it as movement, which is why most stop motion films are done at 12fps.
The amount of motion blur we see on fast moving objects is similar to a 24fps camera with a normal shutter angle setting, but we don’t perceive any blur when we turn our heads or move our eyes like a camera at that fps does.
There’s also our reaction time, which can vary a lot depending on a plethora of factors, but averages around 250ms, which is similar to 4fps 😅… Or maybe since it’s a continuous process it’s more like ♾️fps but with a 250ms delay 🤔😅
So what you meant to say is that you don’t see a difference above 60 Hz. But other people definitely can tell the difference. Don’t generalize on everyone based on your own experiences.
It’s noticeable, it’s just not massive. My phone screen runs at 120hz but I don’t notice a difference unless I’m scrolling rapidly. Gaming culture (driven by corporations) really overemphasizes its importance. Gamers as a group seem to be easily duped by impressive sounding numbers, just like the rest of the population.
Also keep in mind there likely isn’t a lot of selective pressure on biological vision refresh rate, so I wouldn’t be surprised if there’s a fair amount of variability in the ability to discern a difference amongst individuals.
It’s a hot button topic for some people. I’m for the biological variation explanation - some people seem to really notice a difference while others don’t.
I think what the people who get really upset notice is that they dropped a few extra $100s on what’s often largely a marketing gimmick.
We definitely are seeing things faster than 24 Hz, or we wouldn’t be able to tell a difference in refresh rates above that.
Edit: I don’t think we have a digital, on-off refresh rate of our vision, so fps doesn’t exactly apply. Our brain does turn the ongoing stream of sensory data from our eyes into our vision “video”, but compared to digital screen refresh rates, we can definitely tell a difference between 24 and say 60 fps.
Yeah it’s not like frames from a projector. It’s a stream. But the brain skips parts that haven’t changed.
People looking at a strobing light, start to see it as just “on” (not blinking anymore) at almost exactly 60Hz.
In double blind tests, pro gamers can’t reliably tell 90fps from 120.
There is however, an unconscious improvement to reaction time, all the way up to 240fps. Maybe faster.
It seems to be more complicated than that
The real benefit of super high refresh rates is the decrease in latency for input. At lower rates the lag between input and the next frame is extremely apparent, above about ~144hz it’s much less noticable.
The other side effect of running at high fps is that when heavy processing occurs and there are frame time lags they’re much less noticable because the minimum fps is still very high. I usually tell people not to pay attention to the maximum fps rather look at the average and min.
I think having higher frame rates isn’t necessarily about whether our eyes can perceive the frame or not. As another commenter pointed out there’s latency benefits, but also, the frame rate affects how things smear and ghost as you move them around quickly. I don’t just mean in gaming. Actually, it’s more obvious when you’re just reading an article or writing something in Word. If you scroll quickly, the words blur and jitter more at low frame rates, and this is absolutely something you can notice. You might not be able to tell the frametime, but you can notice that a word is here one moment and next thing you know, it teleported 1 cm off
I think i read that fighter pilots need to be able to identify a plane in one frame at 300 fps, and that the theoretical limit of the eye is 1000+ fps.
Though, whether the brain can manage to process the data at 1000+ fps is questionable.
I’m using part of this comment to inform my monitor purchases for the rest of my life.
New 1,200 Hz displays? Well, it did say 1,000-plus…
Both of these claims are kinda misguided. The brain is able to detect very short flashes of light (say, 1 thousandth of a second), and other major changes in light perception. Especially an increase in light will be registered near instantly. However, since it doesn’t have a set frame rate, more minor changes in the light perception (say, 100 fps) are not going to be registered. And the brain does try to actively correct discontinuities, that’s why even 12 fps animation feels like movement, although a bit choppy.
You’re right that is a continuous process, so there’s no frame rate as such. 24 fps is just the lowest framerate that doesn’t look “framey” in videos, but can go as low as 12 and still reliably perceive it as movement, which is why most stop motion films are done at 12fps.
The amount of motion blur we see on fast moving objects is similar to a 24fps camera with a normal shutter angle setting, but we don’t perceive any blur when we turn our heads or move our eyes like a camera at that fps does.
There’s also our reaction time, which can vary a lot depending on a plethora of factors, but averages around 250ms, which is similar to 4fps 😅… Or maybe since it’s a continuous process it’s more like ♾️fps but with a 250ms delay 🤔😅
I would believe it if someone told me that an individual rod or cone in the eye was 24fps but they’re most likely not synched up
Generally humans don’t perceive a difference above 60 Hz.
Completely untrue and not even up for debate. You’d know this if you had ever used a high-refresh rate display.
Let me guess, you’ve also bought a Gamer Chair to go with your overpriced 144 Hz display.
So, have you actually used a 144 Hz display yourself?
Yes, I have observed a complete lack of improvement.
So what you meant to say is that you don’t see a difference above 60 Hz. But other people definitely can tell the difference. Don’t generalize on everyone based on your own experiences.
Well, that and most humans, like I said.
It’s noticeable, it’s just not massive. My phone screen runs at 120hz but I don’t notice a difference unless I’m scrolling rapidly. Gaming culture (driven by corporations) really overemphasizes its importance. Gamers as a group seem to be easily duped by impressive sounding numbers, just like the rest of the population.
Also keep in mind there likely isn’t a lot of selective pressure on biological vision refresh rate, so I wouldn’t be surprised if there’s a fair amount of variability in the ability to discern a difference amongst individuals.
Finally, a reasonable reply to my comment.
It’s a hot button topic for some people. I’m for the biological variation explanation - some people seem to really notice a difference while others don’t.
I think what the people who get really upset notice is that they dropped a few extra $100s on what’s often largely a marketing gimmick.