Light – DiRT 3
DiRT 3 (DirectX 11)
A continuation of the Colin McRae series, but without his name, DiRT 3 is one of the top racing games in the world and offers stunning imagery along with support for features of DirectX 11.
Actual settings for our testing today was 1920×1080 at the LOW preset with No MSAA enabled.
DiRT 3 is another older title that should provide interesting results for us. Again, there is very little difference between our FRAPS based results and the capture-based performance metrics meaning that there aren’t any runt or dropped frames to be found. That’s good news. The AMD Richland system is again the best performer, pulling in an average frame rate of 50 FPS, followed by the Trinity system at 33 FPS, the MBA with HD 5000 at 31 FPS and then the Core i7-4702MQ + HD 4600 at 30 FPS. But again there are definitely smoothness and frame time differences to account for.
The Frame Times graph clearly shows that the Richland desktop APU and the MSI GE40 with the HD 4600 graphics result in the smoothest gameplay in DiRT3 thanks to faster CPU performance to go along with it. Both the Trinity and MBA / HD 5000 performance shows higher variability with differences between frame times over 4 ms. The A10-4600M (Trinity) is by far the worst in this manner though with frame time variance crossing 8 ms at about the 93rd percentile.
Again, this indicates that even though the HD 4600 results in a lower frame rate than the HD 5000, the experience of the former is smoother and more consistent.
Good reviews Ryan and as you
Good reviews Ryan and as you say in the (Performance per Watt and Final Thoughts) section : (Clearly we need to get in a Richland based notebook to see where it stands for the mobile market).
I couldn’t agree more will be interesting to see them results.
Good work.
Nice article Ryan, quick
Nice article Ryan, quick question though; On the Methodology page should the following sentence have the word “NOT” in it where I inserted it (capitalized for ease in finding it)?
“It might not stand out initially, but testing on the integrated panels was NOT possible with this testing; in order to get a capture of the graphical output from the system we need to intercept the signal from the GPU to the display and record it.”
Yup!
Yup!
With 2 out of the 3 laptops,
With 2 out of the 3 laptops, that I own, never get Intel HD graphics Driver updates from the OEM’s [who customizied the Intel HD graphics drivers, then never update the graphics drivers!]. Intel lets the OEMs customize their Intel HD graphics drivers, Once these Intel HD Graphics drivers are customizied by the OEM, Intel can not update them, it becomes the OEM’s responcability to update the OEM modded Intel HD graphics drivers! 2 of the 3 laptops that I own will never get updated graphics drivers. The big question here is What good is Intel graphics without updates? Intel’s record here, thorugh letting laptpop OEM customizie the Intel HD graphics, and OEMs never updating their customizied Intel HD graphics drivers, is Piss Poor! Intel needs to make the laptop OEMs use Intel generic HD graphics drivers, which can be updated by Intel, or Intel needs to require laptop OEM’s to keep the OEM customizied Intel HD graphics drivers updated on a regular basis! Without proper graphics drivers, and graphics driver update support, Intel graphics can not be trusted for gaming, no matter how good Intel’s graphics hardware is!
“And to be completely fair to
“And to be completely fair to Intel, I think that it was able to hold its own in terms of performance per watt of TDP.” –HAHAHA as if ANYONE needs to be fair to INTEL!!!
So ryan, what’s better for a
So ryan, what’s better for a budget game box?
A10-6800k vs a low end dedicated GPU such as GTX750 for 19×10 gaming?
Well the 750 will be faster,
Well the 750 will be faster, but it's a bigger configuration and will use more power / create more heat.
For gaming I would do hybrid
For gaming I would do hybrid crossfire by adding a low to mid range card.
I think if you are going to
I think if you are going to put up a graph on performance/watt then you should put up a performance/price graph as well. These two graphs are the main points of any hardware comparison and go hand in hand with each other. You can’t have one without the other. All your hardware reviews should have performance/watt and performance/price graphs included in them. Those are the best scales to use when comparing hardware. I usually have to make my own graphs because most don’t do that and when they do its only with a couple pieces of hardware. Another suggestion would be to include older hardware in benchmark comparisons because most who are buying new components are upgrading from something older. There is no point really not to include older hardware because the end-user can’t tell how much faster the new card is over theirs. I think the best comparisons for benchmarks should include one from each team of a high-end, mainstream and low-cost part from the current components back at least 2 generations. There are so many benchmarks of hardware of the same generation and there are plenty of places to find it but what is hard to find is how a new component is going to perform over someone’s old component. Just a thought.
@ryan:I did not follow the
@ryan:I did not follow the whole series of article,my question.
I believe one angle wasn’t covered that might affect result.pre-msi,post-msi and latest msi-x (message signal interrupt)
From what I read msi-x or newest possible is recommended.but I can’t find one mono that does not hybridize .(use pre-msi and msi-x or MSI.could you suggest a mono that do 100% msi-x.if your bored it would be a nice article to do.since it should definitely alter most result you had so far (if you didn’t think about interrupt!)
Not mono but motherboard
Not mono but motherboard