I don't like this line of thinking because especially now where new games seem to always suffer from performance issues, it lowers the bar that these developers feel like they'd need to set as far as the experience they're offering for their games.
I think the minimum standard should be at least 60fps, in cases with steam deck and other low-end hardware of course concessions must be made, so either lower graphics settings or deal with lower framerates.
But there's no reason a new game should be suffering poor framerates on modern desktop hardware (looking at you Dragons Dogma 2).
Lower frame rates can be perfectly fine, I find I'm far more bothered by inconsistent frametimes.
The main reason 40fps feels fine on the deck is that the display can come down to that same Hz and operate in lockstep.
I'll take consistent 60 over hitchy 165 most of the time, though VRR means you can occupy kind of a middle ground. But even there frametime inconsistencies can make for a shit experience.
My point is that game developers should aim to deliver games that render at similar framerates throughout.
So many of these recent games do hit decent framerates, but then there's that one in-game location, enemy type, player ability, or particle effect, that just makes the framerate completely shit itself.
It's like these studios are designing each element with a given GPU budget, pushing things right up to the limit, and then do a surprised pikachu face when things run like shit once they try to put more than of these elements together to make an actual game.
My point is that game developers should aim to deliver games that render at similar framerates throughout.
Scenes in most games usually have a high variety of complexity, so the way you'd achieve that is through getting a baseline quite a bit higher than your target FPS, and then limit FPS to your target FPS. This way the game won't utilize near 100 % of the GPU most of the time, but peaks in scene complexity won't cause FPS to drop below the set cap.
This is how it works or at least use to work for a lot of games on console. On PC, you almost always have to make the choice yourself (which is a good thing if you ask me).
For many games with a lot of changing scenery I have to target around 45 FPS with graphics settings to even have a chance of achieving somewhat consistent 30 FPS/33.33ms on the Deck.
On the one hand the Deck is heavily underpowered compared to even lower-end PCs. On the other hand tests show that the Z1 Extreme/7840U isn't much faster at these lower wattages (10-15 watts TDP), so there hasn't been a lot of progress yet.
But it's also that many games don't scale so well anymore. I feel like half the settings in many modern games don't affect performance to any noticeable degree, and even fewer settings affect CPU usage. And if there's low settings, the game often looks unrecognizable because these lower setting models, textures and lighting/shadows are simply generated by the engine SDK and rarely given second thoughts.
Eh, not caring too much about frame rates can be healthy in terms of how long you go between upgrades. Which could have a knock on effect of forcing devs to more often consider weaker builds as people don't upgrade as often
Just depends on how you approach it. If you have relatively new parts, then yeah, you should expect at least 60 FPS. If you have an older system, then the only thing that matters is that it's still enjoyable
Now that they don't have to optimize for last gen console hardware anymore, that's going to be even more rare for any triple-A game. Even a well optimized PS5 game is going to seriously struggle to run on the Deck as even if you reduce the graphical setting, the PS5 essentially has an 8 core version of the 4 core CPU in the Deck.
Combine that with the 15W shared TDP limit and the game would basically have to be able to run using only roughly 25% the CPU load.
And in the wider PC space, VRAM is a big issue as well.
Everybody keeps acting like every single big game that comes out is "unoptimised" because it uses more than 8GB of VRAM when playing at 1440p or above at ultra settings, and people with 8GB or lower GPUs are struggling.
Are all these games unoptimised? Or is it simply that they're done targeting consoles with 8GB RAM, are now exclusively focusing on ones with 16GB, and because of Nvidia being notoriously stingy with VRAM, lots of people on PC are suddenly finding their cards falling short?
Thank fuck Valve had the sense to put 16GB of unified memory in the Deck. Could you imagine if they only went with 8GB, or even 12GB?
the steam deck just increased my tolerance lol. i want my games running at minimum 90 fps on desktop, and for steam deck thats 40 or 60 fps depending on the game.
It's because of the screen. Low frame rates are much less impactful on a small screen. When the screen takes up most of your field of view, you notice the frame rate more.
It also depends on the type of game. Quick turns with lower frames an fps shooter will seem very different from low frames in a side scrolling platformer. Third person games often feel smoother with the same frames than with the first person view because the way the game turns is different.
I think a second factor is a lowered expectation of immersion that comes with playing elsewhere. The author specifically says they can play in the pub waiting for friends or on the train or whatever. You're not as invested in those situations, so a game is more of a distraction than an experience.
I've been playing some PS1 classics on my PC, and I can safely say, playing FF7 at 240 fps does not meaningfully improve the experience over what it was back in nineteen ninety eight.
Same here. I went through applying a 60 fps patch for dark souls 1 (cause I do prefer it) and once I hit a listed bug of getting flung off a ladder I unapplied that shit immediately. It's not worth it.
I'm getting kinda tired of the slideshow snobs, telling everyone how 30 FPS is enough. The games are supposed to be fun, and not cause nausea. I'm willing to compromise on the former, but not the latter.
Some people easily get motion sickness and it can be aggravated by many factors, including low and/or irregular framerates.
I'd be interested to know if people complaining about motion sickness at low fps have that issue with all games, or only FPS/TPS. And if they have the same issue with "first person" segments in movies (which are pretty damn rare in the first place, and basically always at a very consistent but low framerate)
To be fair, after getting a OLED TV, I can't stand 24 FPS content at all. With LCD, the blur between frames is just enough to mask the issue, but on OLED movement gets extremely stuttery, and if you get distracted focusing on it, you can even see the steps in each individual frame. It's nauseating.
I had to do the unthinkable and enable the less intrusive motion smoothing option on my TV, otherwise I'd straight up get a headache. This does not happen at any higher framerates. And I'm not talking about gaming at all, I mean TV and movie content.
theres a lot of factors that impact tolerance. for example motion blur. I personally hate a lot of post processing effects, so having motion blur off makes low fps gameplay pretty jarring.
Currently im going through pokemon violet, not on my switch, but on emulation, and the 30fps is really rough, and the 60fps mod has tradeoffs. If im complaining about ot on emulation on a reletively high end pc, I couldnt even imagine how bad the performance was on native hardware, given it was one of its biggest complaints.
Probably has something to do with the screen size. 30 fps on a small screen is way less exhausting on the eyes than sitting Infront of a 27" or bigger screen.
I the portability of the steamdeck. Years ago I had a desktop gaming PC that over time I used less and less because the lass thing I wanted to do after sitting at a desk all day was to sit at a desk some more.
I bring my deck with me when I travel for work, go on vacation, when my wife and I spend the afternoon at my parents. I'll even bring it with me if I have to bring my car to the shop and I have to wait. I play games in my recliner, and in bed. I've never been a frame rate snob so the steamdeck suits me just fine.
This depends on the game and the viewing distance + screen size for me. 30 fps in a tbs game like civilization is perfectly fine for me, but too slow for an rts like total war. 60 fps in total war works for me on the big screen living room tv, but I find it too slow on a desktop computer screen. I expect shooters with jumping and fast turns to benefit even more from faster fps than my rts games, but it's been years since I played one.
Never really had the privilege of being a frame rate snob. My teenager years was pretty much playing games that my computer didn't have the minimum requirements to run.
Now I am a steam deck owner and I am very happy. My only complaints are when games have really bugged UIs or when controller have these nonsensical key binds you can't change like L3 + left D pad (Path of Exile).
I've intentionally kept myself desensitized to higher frame rates over the years. I'll occasionally go up to 60fps on some titles that really need it, but usually if I play most everything at 30fps I never am bothered by it.
I've sort of done the same thing. Most console games are optimized around their 'quality' game mode. Many games will have the quality mode be a solid locked 30 fps while performance will be a low res, blurry mess running at an unlocked frame rate between 40-55 fps. I'll happily play at 30 fps to avoid those issues.
Do not, my friends, become addicted to 60 fps. It will take hold of you, and you will resent its absence!
Don't get addicted to 60? That's the standard. We've moved on to "don't get addicted to 144" because that is both easily achievable with moderately affordable hardware and widely supported now.
I find framerate more noticeable when I'm closer to the screen. Which is why I don't mind the framerates of my phone/Steam Deck, but on a desktop monitor 60 Hz is not enough
I've been solid in my belief that framestes are for suckers. In my experience anything running at 60 fps is fine, probably 30 even. I have to stress myself to really take note of fps in any meaningful way.
I still love my high frame rates, but playing with a controller is much more forgiving in that regard and i mostly play games with no camera controls on the Deck so it is perfectly fine.
As someone who used to play mostly on older consoles with framerates locked at 60 or even 30 FPS I never really understood framerate snobs, the only real time I see it making sense is for VR since you need high frame-rate for that, but for normal games (no I will not call them 'flat games') it's overkill, at least in my opinion.
This feels like cope. Controller input with limited movement and look velocity definitely hides a lot of performance issues, but also it also massively limits control complexity. Basically all modern AAA games are harmed to some degree by needing to support controllers imo.
From a mouse lover who's never heard of gyro aiming because eww joystick gross, that actually sounds really cool. I don't like that the screen is coupled with the movement, but it gets you multiple sensitivity ranges including tactile precision on the same input device.
It's not just look precision, it's the limited number of buttons. A great example is how amazing and complex combat is in the Witcher 2 because it was a PC first game, and how much it was simplified for the third game to target console.
Gyro aiming is definitely a big improvement though.