Fable Legends Early Preview: DirectX 12 Benchmark Analysis
by Ryan Smith, Ian Cutress & Daniel Williams on September 24, 2015 9:00 AM ESTDiscussing Percentiles and Minimum Frame Rates
Continuing from the previous page, we performed a similar analysis on AMD's Fury X graphics card. Same rules apply - all three resolution/setting combinations using all three system configurations. Results are given as frame rate profiles showing percentiles as well as choosing the 90th, 95th and 99th percentile values to get an indication of minimum frame rates.
Moving on to the Fury X at 4K and we see all three processor lineups performing similarly, giving us an indication that we are more GPU limited here. There is a slight underline on the Core i7 though, giving slightly lower frame rates in easier scenes but a better frame rate when the going gets tough beyond the 95th percentile.
For 1080p, the results take a twist. It almost seems as if we have some form of reverse scaling, whereby more cores is doing more damage to the results. If we have a look at the breakdown provided by the in-game benchmark (given in milliseconds, so lower is better):
Three areas stand out as benefitting from fewer cores: Transparency and Effects, GBuffer Rendering and Dynamic Lighting. All three are related to illumination and how the illumination interacts with its surroundings. One reason springs to mind on this – with large core counts, too many threads are issuing work to the graphics card causing thread contention in the cache or giving the thread scheduler a hard time depending on what comes in as high priority.
Nevertheless, the situation changes when we move down again to 720p:
Here the Core i3 takes a nose dive as we become CPU limited to pushing out the frames.
141 Comments
View All Comments
Alexvrb - Friday, September 25, 2015 - link
Ship them both to the East Coast and set up a Review Office / Beach Resort, complete with community events!zimanodenea - Thursday, September 24, 2015 - link
My Asus m5a97 has an option to do this.mdriftmeyer - Thursday, September 24, 2015 - link
Time to develop in a test harness of equal merits and scope across the globe for the reviewers. To do less is unprofessional. The whole point of a test harness is not to ductape simulations but to cover all bases.Spunjji - Friday, September 25, 2015 - link
Well said. This isn't some tinpot organisation, is it? ;)Drumsticks - Thursday, September 24, 2015 - link
That's a shame. I'd really like to see that comparison. With the improvements Zen should, in theory, bring, it could really give AMD its best chance in years to get some wind under its sails.beck2050 - Thursday, September 24, 2015 - link
A little too early to worry about. Hopefully both companies will improve when 12 becomes standard issue.DrKlahn - Thursday, September 24, 2015 - link
Epic has always worked closely with Nvidia and used their hardware, so the only thing that surprises me is that the gap doesn't favor Nvidia more. It's very early to make any predictions, but there are some interesting conversations on other forums about how both architectures behave in different situations. Nvidia's architecture does appear to have issues in some asynchronous workloads. What little evidence we have says this may be an issue in some games.My own opinion is that with Nvidia's market dominance we will see most developers try to avoid situations where problems occur. As an AMD owner my main hope is that we see DX12 squeeze out proprietary codes and level the playing field more. I'm also happy that the latest Unreal engine appears to run well on both vendors hardware.
jiao lu - Thursday, September 24, 2015 - link
not only close working relationship . the Unreal 3/4 use Nvidia Physics sdk outright. Epic engine is terribly optimized for console right now. basically it is PC engine, churn out pc demo now and then . Now far fewer AAA studio use unreal 4 like they do with unreal 3 in the ps 3/xbox 360 era. I am very much suspicious unreal 5 is not mult-threaded rendering enough , use dx 12 like do dx 11 before.Midwayman - Thursday, September 24, 2015 - link
Well, the xbox one is using AMD hardware and dx12. That's probably a bigger reason to keep it neutral than more nvidia share on the PC.Spunjji - Friday, September 25, 2015 - link
The PS4 is also using the same AMD GCN 1.0 architectures for CPU and GPU