Stalker 2 PC performance testing and settings analysis — another demanding game that uses frame generation and upscaling as a crutch

Written on 11/20/2024
Jarred Walton

Stalker 2 arrives with steep PC system requirements and an assumption that everyone should be using some form of frame generation, or at least that seems to be the takeaway from the benchmarks we've conducted. The game looks good, but Unreal Engine 5 proves to be very demanding yet again.

(Image credit: GSC Game World)

Stalker 2 returns to the zone after a long absence

Stalker 2: Heart of Chornobyl finally marks a return to the zone after a 15 year absence. It leverages Epic's Unreal Engine 5 for some impressive visuals, and it also punishes lesser PCs. We've been running benchmarks and this is a game that needs both GPU and CPU horsepower — and don't neglect the VRAM if you're planning to run with maxed out settings. Our testing will look at many of the latest Nvidia, AMD, and Intel GPUs that are among the best graphics cards, and we'll provide a few CPU benchmarks as well. Let us know in the comments what other hardware you'd like to see tested, and we'll try to make it happen.

A lot has changed since Call of Prypiat, Clear Sky, and the original Shadow of Chornobyl came out in the late 00s (note that the spelling was officially updated after the Russian invasion of Ukraine). A state-of-the-art PC back in 2009, when Prypiat launched, consisted of a Core i7-965 or 975 Extreme Edition, or an overclocked i7-920 that was just as fast. The top graphics cards of the era were Nvidia's GeForce GTX 285 AMD's newly launched Radeon HD 5870. Windows 7 was also brand-spanking-new, having launched in October 2009, which means DirectX 11 had also just gone live. Call of Prypiat was one of the first (or perhaps the first) games to use DirectX 11, though it was also compatible with DX9 and DX10 GPUs.

Now, Windows 7 has officially been retired, along with the Windows 8 sequel, and Windows 10 is on its last legs. GPUs that were top of the line models 15 years ago often can't even run modern games. Stalker 2 as an example (and yes, we're going to skip the caps and periods rather than writing "S.T.A.L.K.E.R. 2" every time, since it's not even an acronym) requires at least a GTX 1060 6GB or RX 580 8GB graphics card — those arrived in 2016 and 2017, respectively. It also requires a Core i7-7700K or Ryzen 5 1600X or better, 16GB of RAM, and 160GB of disk space. That's for a target of 1080p with the low preset and 30 fps.

The recommended hardware — and based on our testing, you don't want to come up short — lists a Core i7-11700 or Ryzen 7 5800X, GeForce RTX 3070 Ti / 4070 or Radeon RX 6800 XT, and 32GB of RAM. That's for the high preset and 60 fps at 1440p. Also, an SSD is "required" — though I suspect the game can still load (slowly) from a hard drive. And those framerate targets? That's with TSR, DLSS, FSR, or XeSS upscaling and related technologies, including framegen.

Steep requirements might be an understatement, and once again we can only shake our heads sadly at the default enablement of upscaling and frame generation. We prefer to have those as ways to boost performance to higher levels and smooth out framerates, not as baseline requirements. To that end, we'll be testing at native resolution, as well as with upscaling only, and finally with upscaling and frame generation. So let's see how Stalker 2 runs on current PC hardware.

Stalker 2 Test Setup

GPU Test PC

Stalker 2 comes with the full complement of upscaling technologies from Nvidia, AMD, Intel, and Unreal Engine 5: DLSS (3.7.0), FSR 3.1, XeSS 1.3.1, and TSR (Temporal Super Resolution). Or you can turn off all of those and run with TAA (Temporal Anti-Aliasing), still with the option to enable (a rather poor looking) scaling if you want.

Our main test PC consists of our soon-to-be-retired from active duty Core i9-13900K, 32GB of DDR5-5600 CL28 memory, an MSI Z790 Ace motherboard, and a 4TB Crucial T700 SSD. We'll have additional results for the 4090 and 7900 XTX running on some other CPUs, to show just how CPU limited the game can be.

GPU drivers for our initial testing are Nvidia 566.14, Intel 6299, and AMD 24.10.1. This is an Nvidia promoted game, and AMD hasn't released any new drivers since October 18. Which is a bit concerning, given the number of new games coming out right now, like Flight Simulator 2024. AMD will probably drop new drivers right after we post this, but we didn't see any noteworthy issues in our testing.

Our aim here isn't to test every current generation GPU. We'll skip the Nvidia RTX 4080 and RTX 4070 Ti, for example, as well as the Arc A770 8GB and A380. We'll see about adding some previous generation cards as well in the next couple of days, just as a point of reference to help line up where such cards rank compared to newer offerings.

Our typical testing consists of using medium settings at 1080p, and then 1080p, 1440p, and 4K with ultra — or in this case, "epic" — settings. We'll do all of the baseline testing with TAA and 100% scaling, and then run some additional tests at 1440p and 4K with upscaling, and all three resolutions with upscaling and framegen. Upscaling alone won't help the fastest cards, as we'll hit CPU limits (at least on the 13900K), so keep that in mind.

Our benchmark sequence consists of a manual run through the starting village of Zalissya. The map, as in previous Stalker games, is effectively completely open from the start and you can wander seamlessly throughout the exclusion zone... assuming you can survive. The first town makes for a decent benchmark sequence as there are no mutants or enemies, but there are plenty of NPCs wandering about as well as buildings and trees. We run each GPU at each setting at least twice, using the higher score, and three times for the first run on each preset (i.e. medium and epic), discarding the first result.

We'll have screenshots and a discussion of image fidelity at the various settings further down the page, after the benchmarks. Stalker 2 does not utilize ray tracing hardware, opting instead for software-based rendering using Unreal's Lumen and Nanite technologies. It's plenty heavy even without using hardware RT, though, as we'll soon see.

[Note: We're still testing additional GPUs and will periodically update these charts. Let us know in the comments if there are any GPUs in particular that you'd like to see tested.]

Stalker 2 1080p Medium GPU performance

1080p medium represents our baseline performance metric. Yes, you can go much higher, and you can also go lower — on resolution and settings. You can enable upscaling and framegen as well. But we feel 1080p and medium settings represent a target that most "reasonable" gaming PCs should be able to manage. Here are the results.

sLsD5zhtnR5uiLLYV8TfJh.png

(Image credit: Tom's Hardware)

One thing that's immediately obvious is that we're hitting CPU bottlenecks here. AMD's top GPUs run slightly faster than Nvidia's top GPUs, but everything lands right around the 100 fps mark. Considering the relatively low target, that's a bit surprising, but Stalker games have often pushed the hardware of their time to the limit.

Some of this is also due to the use of Unreal Engine 5. It's possible to hit much higher framerates, with the right settings and game optimizations, but from the games we've tested it seems that UE5 is usually quite heavy on both the CPU and GPU fronts.

Stalker 2 1080p Epic GPU performance

A6kCtGDq8t5ZrRYL4p3uQh.png

(Image credit: Tom's Hardware)

The CPU bottleneck becomes far more obvious at 1080p and epic settings. Now, the three fastest GPUs are all tied at 85~86 fps, and the next two GPUs aren't too far off that pace. The 1% lows are also pretty consistent at less than 60 fps.

This is, sadly, the state of many games these days, particularly those using Unreal Engine 5. Running at native resolution like we're doing is no longer the goal. Instead, developers are using upscaling and frame generation technologies as crutches that allow them to post higher performance numbers.

Sure, Stalker 2 looks good graphically, but not amazing. UE5 also has a lot of pop-in where you'll see the geometry on objects massively shift — especially on trees, in this particular game. It's distracting and not something I like to see on top-end hardware with maxed out settings. Dragon Age: The Veilguard has similar quality graphics in my book, and manages to do away with a lot of the level of detail scaling pop-in that's so common with UE5 games. (And if it's not clear, no, Veilguard isn't using UE5.)

We also dropped the 8GB GPUs from testing. 1080p high settings seems to be viable, but 1080p epic exceeds their VRAM capacity and performance absolutely tanks. The RTX 4060 Ti 8GB card initially ran at around 48 fps, but after a minute or so it had dropped to around 20~25 fps with frequent dips into the single digits. The RTX 4060 never showed higher performance and ran in the single digits as well. AMD's RX 7600 likewise started at a higher level (~20 fps) but then dropped to <10 fps after a minute or so running around the game world.

Stalker 2 1440p Epic GPU performance

xgK8k3pyPTyFc5ea5e3eWh.png

(Image credit: Tom's Hardware)

1440p epic settings finally starts to move away from the CPU bottleneck, but the 4090 at least is still clearly being held back. Power use for many of the faster Nvidia GPUs is also quite low, a key indicator of CPU limitations. The 4090 used 294W, 4080 Super drew 242W, and the 4070 Ti Super needed 241W. In contrast, AMD's top GPUs were still close to their rated power limits, with 360W on the 7900 XTX, 317W for the 7900 XT, and 256W on the 7900 GRE.

In the overall standings, the 4090 and 4080 Super now take the top two spots, though the 7900 XTX matches the latter with only a 0.8 fps difference (a 1.1% difference, which is within the margin of error). The 7900 XT matches the 4070 Ti Super as well, which means it has the value advantage given its current price — you can check our GPU price index for current values, but right now the 4070 Ti Super costs around $780 while the 7900 XT only costs $640.

Similar discrepancies in price to performance exist throughout much of the current GPU landscape. The 4080 Super costs $950 or more while the 7900 XTX starts at $840. The 4070 and 7900 GRE both have a $550 MSRP, but here Nvidia undercuts AMD with a $500 card versus a $530 card — albeit with slightly lower performance. Nvidia's 4060 Ti 16GB doesn't look too good either, with a $450 price tag and performance that falls below the $390 7700 XT, never mind the $470 RX 7800 XT that's 30% faster at these settings.

Stalker 2 4K Epic GPU performance

kPsdXuehhmhy6rqLTEVach.png

(Image credit: Tom's Hardware)

Native 4K at maxed out settings as usual proves to be too much for most of the GPUs. Even the mighty RTX 4090 only manages 57 fps — definitely playable, but not quite perfectly smooth. And while there are less demanding areas in Stalker 2, we're sure there are also regions that will be even more taxing and would push framerates even lower.

The 4090 leads the second string 4080 Super by 20%, which still isn't as large of a gap as we often see — and the 4090 still only pulls 346W, so something is keeping it from reaching higher levels of performance. The RTX 4080 Super in turn beats the RX 7900 XTX by a couple of FPS in average performance, though minimums are more stable on the AMD cards. That's actually something we saw throughout our testing: AMD's GPUs, probably thanks to having more VRAM at most levels, end up with better minimum FPS.

VRAM also proves to be an issue on GPUs with less than 16GB now. The RTX 4070 falls below both the 7800 XT and 7900 GRE, but AMD's RX 7700 XT likewise takes a tumble down to the low 20s.

Stalker 2 with upscaling

Given the obvious CPU bottlenecks at lower resolutions, it shouldn't be too surprising that those same limitations show up when we turn on upscaling. We skipped 1080p testing for that reason, focusing on the 1440p and 4K results. They're still pretty underwhelming at times.

fTNnbyLSxBHefehMghkKih.png

(Image credit: Tom's Hardware)

At 1440p epic settings with quality mode upscaling (67% render resolution), the top six GPUs all land in the 80–85 fps range, with minimums of 48–52 fps. So, the 13900K isn't able to keep the GPUs fully saturated with work. We'll be testing a Ryzen 7 9800X3D shortly to see how that changes things, and maybe a few other CPUs as well, so stay tuned.

Lower tier GPUs do benefit from upscaling, and the further they are from 80 fps, the bigger the potential boost. the 7600 XT for example sees a 37% increase over native 1440p, and the 4060 Ti 16GB improves by a similar 35%. Everything from the RX 7800 XT and up manages to break 60 fps as well, and all of the tested cards are definitely playable (of course, we already dropped GPUs like the 8GB cards from our testing at 1080p due to performance issues).

But we do need to mention that DLSS, FSR3, and XeSS are not the same. That's part of why we stick to TAA 100% resolution as the baseline. Software algorithms to improve performance can be beneficial, but fundamentally the 7700 XT using FSR3 upscaling does not look the same as the 4060 Ti 16GB using DLSS upscaling. We would say both are viable, particularly if you're not too finicky, but DLSS continues to look better. (FSR3 upscaling looked noticeably worse at 1080p as well, if you're wondering.)

3VKo3tSE6WNmuu6BzTx5ph.png

(Image credit: Tom's Hardware)

4K epic with upscaling reduces the CPU bottleneck, though many of the positions still look closer than normal. The 4090 for example only beats the 4080 Super by 12%, which is basically in line with our larger GPU benchmarks hierarchy results for 1440p ultra in pure rasterization performance. But AMD's 7900 XTX shows a 13% lead over the 7900 XT, compared to just 7.4% in our hierarchy.

In games where we're clearly not CPU limited, though? The RTX 4090 can be up to 30–35 percent faster than the 4080 Super. That often requires the use of hardware ray tracing features, or at least a game that doesn't run into clear CPU limitations.

Stalker 2 with upscaling and frame generation

Frame generation is often a highly controversial feature, and rightly so. Some people refer to the generated frames as "fake frames," and they're not really wrong. Because there's no additional sampling of user input, and with the added overhead of framegen, it often feels more like two steps backward for two steps forward.

Stalker 2 supports FSR 3.1 and DLSS 3 framegen, which means if you really want to, you can use DLSS upscaling with FSR framegen. That can result in higher performance on Nvidia's cards — and of course you'd only have the option for FSR framegen if you have an RTX 30- or 20-series GPU, or anything else for that matter. But with pure FSR upscaling and framegen, we checked the 4070 Super and found that DLSS 3 upscaling and framegen was 38% faster than FSR 3 upscaling and framegen.

That's a critical point that many overlook. Yes, FSR 3.1 is technically open source, but that doesn't mean AMD didn't do a lot of work optimizing it to work better on its own hardware. Nvidia doesn't want FSR to be better than DLSS, so it's not actively working to improve performance with the technology. And so, we have apples and oranges that we get to compare and debate. Here are the results with both quality upscaling and framegen enabled.

wdA4FTC4fJNTFG26CTGtuh.png

(Image credit: Tom's Hardware)

Starting with 1080p, because of the heavy CPU bottleneck, framegen gives some significant improvements to perceived FPS. However, as we've said in the past, this is basically "frame smoothing" rather than "frame generation" — because no new user input gets factored into the generated frames. It's not inherently bad, an in Stalker 2 the overall performance uplift can make it worthwhile, but the user input sampling rate will be half of the shown FPS values.

And again, as we've seen in some other games, FSR3 gives AMD a bigger boost in overall performance than DLSS gives to Nvidia. It's not a "fair" comparison in that the GPUs are not running the same workloads, and using FSR3 on Nvidia doesn't fix things as it's only getting optimizations for AMD GPUs as far as we can tell. 1080p with upscaling and framegen absolutely looks worse on the AMD cards, even if it runs faster.

tDCdFfnNFWtZqhs8aCso2i.png

(Image credit: Tom's Hardware)

The higher output resolution of 1440p does make some of the rendering issues that were clearly visible at 1080p go away for FSR3. You have to start pixel peeping to really see differences, so in some sense this is a win. But this is also why we don't like games or graphics card companies using upscaling and framegen as the default comparison. That's ripe for abuse by using only the optimal results to make one company's products look better.

The important thing, insofar as Stalker 2 goes, is that every GPU in this chart is still imminently playable. Yes, user input still happens at half the rate shown in the chart, but it does look smoother and that tends to trick your brain into the game feeling smoother as well. When framegen only yields a 20–30 percent improvement in FPS, that's not very good, but here we're getting much better scaling.

The RTX 4060 Ti 16GB went from 53.3 to 76.3 fps, a 43% improvement. The higher tier 4090 on the other hand gets a 75% boost. So perhaps memory bandwidth or some other factor is holding back the lower tier cards.

AMD, as expected, gets an even larger boost from FSR3 framegen. The RX 7600 XT goes from 43.3 to 77.7 fps, a 79% increase. The 7900 XTX at the top of the chart sees an even larger 90% overall performance boost. That's very near to the theoretical 100% boost that framegen can offer (which shouldn't ever be realized unless there's other wonkiness involved).

RHaaVDDw7M242RR5xgWY8i.png

(Image credit: Tom's Hardware)

4K with upscaling and framegen also yields mostly playable results on all the GPUs we tested — we say "mostly" because the 4060 Ti 16GB and 7600 XT at just 44–45 fps definitely start to feel sluggish. That's because the input sampling rate is only around 22 fps. But the 4070 and above all break 60 fps of generated performance, which works fine overall.

Note that 12GB cards can still struggle at these settings, especially over longer play sessions. So the 4070 and 7700 XT (and 4070 Super and 4070 Ti) become more likely to see short bursts of highly unstable framerates — meaning, you might get individual frames falling into the single digit FPS range.

yzNo9YWjCQ2SuWgwbHyXZm.jpg

(Image credit: GSC Game World)

Stalker 2 settings and image quality

We haven't had time to do a full image quality analysis, so check back and we'll update this section.

Stalker 2 closing thoughts

There's a lot to digest, but as far as performance and system requirements go, our advice is to make sure you exceed the minimum recommendations. Also, plan on using upscaling and framegen for most GPUs, especially at higher resolutions.

We still need to test some CPUs, plus additional GPUs, and we'll be updating the article with that information in the next day or two (and testing Flight Simulator 2024 at the same time, assuming it continues to work). But if you're wondering if Stalker 2 properly carries forward the Stalker legacy, it does — in both good and bad ways.

The graphics have improved compared to the older games, as well they should after 15+ years. But a lot of the gameplay so far feels quite similar. You're constantly running into enemies that have PDAs that reveal stash locations, anomalies are all over the place, and you have to try and walk the fine line between the various factions. If you like the post-apocalyptic (-esque) setting of the Chornobyl area, returning to it again after all these years will almost certainly hit all the right pleasure centers of your brain.

It's still frustrating to see Unreal Engine 5 apparently requiring so much help to reach higher levels of performance. Framegen in particular should not be the baseline recommendation. It has some clear drawbacks, and tends to be more useful as a marketing feature. We'd rather have slightly less advanced graphics that run well over interpolated frames to smooth out some of the choppiness. Other games and game engines seem to manage this much better than UE5, but Unreal Engine gets used a lot because it makes life easier on the game developers — often at the cost of performance.

8uFgSGcCzKdFTTQdqonCPi.jpg

Jarred Walton is a senior editor at Tom's Hardware focusing on everything GPU. He has been working as a tech journalist since 2004, writing for AnandTech, Maximum PC, and PC Gamer. From the first S3 Virge '3D decelerators' to today's GPUs, Jarred keeps up with all the latest graphics trends and is the one to ask about game performance.