Crysis: Warhead

Kicking things off, we’ll start with Crysis: Warhead. Warhead is still the single most demanding game in our arsenal, with cards continuing to struggle to put out a playable frame rate with everything turned up.

AMD’s aspirations for the 5870 are that it will beat the GTX 295. Here they get close, but no cigar. Higher-end dual GPU solutions like the GTX 285 in SLI or the 5870 in Crossfire are still necessary to achieve playable framerates at 2560 in the Frost benchmark, and this isn’t even with the game at its highest settings. For those of you hoping to play Warhead completely maxed out, even the 5870 CF isn’t quite going to be able to deliver on that.

The Test Far Cry 2
Comments Locked

327 Comments

View All Comments

  • Xajel - Wednesday, September 23, 2009 - link

    Ryan,

    I've send a detailed solution for this problem to you ( Aero disabled when running and video with UVD accerelation ), but the basic for all readers here is just install HydraVision & Avivo Video Convertor packages, and this should fix the problem...
  • Ryan Smith - Wednesday, September 23, 2009 - link

    Just so we're clear, Basic mode is only being triggered when HDCP is being used to protect the content. It is not being triggered by just using the UVD with regular/unprotected content.
  • biigfoot - Wednesday, September 23, 2009 - link

    Well, I've been looking forward to reading your review of this new card for a little while now, especially since the realistic sounding specs leaked out a week or so ago. My first honest impression is that it looks like it'll be a little bit longer till the drivers mature and the game developers figure out some creative ways to bring the processor up to its full potential, but in the mean time, it looks like I'd have to agree with everyone's conclusion that even 150+GB/s isn't enough memory bandwidth for the beast, I'm sure it was a calculated compromise while the design was still on the drawing board. Unfortunately, increasing the bus isn't as easy as stapling on a couple more memory controllers, they probably would've had to resort back to a wider ring bus and they've already been down that road. (R600 anyone) Already knowing how much more die area and power would've been required to execute such a design probably made the decision rather easy to stick with the tested 4 channel GDDR5 setup that worked so well for the RV770 and RV790. As for how much a 6 or 8 channel (384/512 bit) memory controller setup would've improved performance, we'll probably never know; as awesome as it would be, I don't foresee BoFox's idea of ATI pulling a fast one on nVidia and the rest of us by releasing a 512-bit derivative in short order, but crazier things have happened.
  • biigfoot - Wednesday, September 23, 2009 - link

    Oh, I also noticed, they must of known that the new memory controller topology wasn't going to cut it all the time judging from all the cache augmentations performed. But like i said, given time, I'm sure they'll optimize the drivers to take advantage of all the new functionality, I'm betting that 90% of the low level functions are still handled identically as they were in the last generations architecture and it will take a while till all the hardware optimizations like the cache upgrades are fully realized.
  • piroroadkill - Wednesday, September 23, 2009 - link

    The single most disappointing thing about this card is the noise and heat.

    Sapphire have been coming out with some great coolers recently, in their VAPOR-X line. Why doesn't ATI stop using the same dustbuster cooler in a new shiny cover, and create a much, much quieter cooler, so the rest of us don't have to wait for a million OEM variations until there's one with a good cooler (or fuck about and swap the cooler ourselves, but unless you have one that covers the GPU AND the RAM chips, forget it. Those pathetic little sticky pad RAM sinks suck total donkey balls.)
  • Kaleid - Wednesday, September 23, 2009 - link

    Already showing up...

    http://www.techpowerup.com/104447/Sapphire_HD_5870...">http://www.techpowerup.com/104447/Sapphire_HD_5870...

    It will be possible to cool the card quietly...
  • Dante80 - Wednesday, September 23, 2009 - link

    the answer you are looking for is simple. Another, more elaborate cooler would raise prices more, and vendors don't like that. Remember what happened to the more expensive stock cooler for the 4770? ...;)
  • Cookie Monster - Wednesday, September 23, 2009 - link

    When running dual/multi monitors with past generation cards, the cards would always run at full 3d clocks or else face instabilities, screen corruptions etc. So even if the cards are at idle, it would never ramp down to the 2d clocks to save power, rending impressive low idle power consumption numbers useless (especially on the GTX200 series cards).

    Now with RV870 has this problem been fixed?
  • Ryan Smith - Wednesday, September 23, 2009 - link

    That's a good question, and something we didn't test. Unfortunately we're at IDF right now, so it's not something we can test at this moment, either.
  • Cookie Monster - Wednesday, September 23, 2009 - link

    It would be awesome if you guys do get some free time to test it out. Would be really appreciated! :)

Log in

Don't have an account? Sign up now