Mirror’s Edge: Do we have a winner?

And now we get to the final test. Something truly different: Mirror’s Edge.

This is an EA game. Ben had to leave before we got to this part of the test, he does have a wife and kid after all, so I went at this one alone.

I’d never played Mirror’s Edge. I’d seen the videos, it looked interesting. You play as a girl, Faith, a runner. You run across rooftops, through buildings, it’s all very parkour-like. You’re often being pursued by “blues”, police offers, as you run through the game. I won’t give away any plot details here but this game, I liked.

The GPU accelerated PhysX impacted things like how glass shatters and the presence of destructible cloth. We posted a video of what the game looks like with NVIDIA GPU accelerated PhysX enabled late last year:

"Here is the side by side video showing better what DICE has added to Mirror's Edge for the PC with PhysX. Please note that the makers of the video (not us) slowed down the game during some effects to better show them off. The slow downs are not performance related issues. Also, the video is best viewed in full screen mode (the button in the bottom right corner)."

 

In Derek’s blog about the game he said the following:

“We still want to really get our hands on the game to see if it feels worth it, but from this video, we can at least say that there is more positive visual impact in Mirror's Edge than any major title that has used PhysX to date. NVIDIA is really trying to get developers to build something compelling out of PhysX, and Mirror's Edge has potential. We are anxious to see if the follow through is there.”

Well, we have had our hands on the game and I’ve played it quite a bit. I started with PhysX enabled. I was looking for the SSD-effect. I wanted to play with it on then take it away and see if I missed it. I played through the first couple of chapters with PhysX enabled, fell in lust with the game and then turned off PhysX.

I missed it.

I actually missed it. What did it for me was the way the glass shattered. When I was being pursued by blues and they were firing at me as I ran through a hallway full of windows, the hardware accelerated PhysX version was more believable. I felt more like I was in a movie than in a video game. Don’t get me wrong, it wasn’t hyper realistic, but the effect was noticeable.

I replayed a couple of chapters and then played some new ones with PhysX disabled now before turning it back on and repeating the test.

The impact of GPU accelerated PhysX was noticeable. EA had done it right.

The Verdict?

So am I sold? Would I gladly choose a slower NVIDIA part because of PhysX support? Of course not.

The reason why I enjoyed GPU accelerated PhysX in Mirror’s Edge was because it’s a good game to begin with. The implementation is subtle, but it augments an already visually interesting title. It makes the gameplay experience slightly more engrossing.

It’s a nice bonus if I already own a NVIDIA GPU, it’s not a reason for buying one.

The fact of the matter is that Mirror’s Edge should be the bare minimum requirement for GPU accelerated PhysX in games. The game has to be good to begin with and the effects should be the cherry on top. Crappy titles and gimmicky physics aren’t going to convince anyone. Aggressive marketing on top of that is merely going to push people like us to call GPU accelerated PhysX out for what it is. I can’t even call the overall implementations I’ve seen in games half baked, the oven isn’t even preheated yet. Mirror’s Edge so far is an outlier. You can pick a string of cheese off of a casserole and like it, but without some serious time in the oven it’s not going to be a good meal.

Then there’s the OpenCL argument. NVIDIA won’t port PhysX to OpenCL, at least not anytime soon. But Havok is being ported to OpenCL, that means by the end of this year all games that use OpenCL Havok can use GPU accelerated physics on any OpenCL compliant video card (NVIDIA, ATI and Intel when Larrabee comes out).

While I do believe that NVIDIA and EA were on to something with the implementation of PhysX in Mirror’s Edge, I do not believe NVIDIA is strong enough to drive the entire market on its own. Cross platform APIs like OpenCL will be the future of GPU accelerated physics, they have to be, simply because NVIDIA isn’t the only game in town. The majority of PhysX titles aren’t accelerated on NVIDIA GPUs, I would suspect that it won’t take too long for OpenCL accelerated Havok titles to equal that number once it’s ready.

Until we get a standard for GPU accelerated physics that all GPU vendors can use or until NVIDIA can somehow convince every major game developer to include compelling features that will only be accelerated on NVIDIA hardware, hardware PhysX will be nothing more than fancy lettering on a cake.

You wanted us to look at PhysX in a review of an ATI GPU, and there you have it.

The Unreal Tournament 3 PhysX Mod Pack: Finally, a Major Title CUDA - Oh there’s More
Comments Locked

294 Comments

View All Comments

  • jeffrey - Thursday, April 2, 2009 - link

    1) ATI driver - Was 8.12 really used? Why? 9.3 was released last month.

    2) Conclusion - The edge should have gone to the 4890 for being ahead of the 275 in most games at resolutions targeting the price point.
  • Gary Key - Thursday, April 2, 2009 - link

    1. The 9.4 beta was used for the HD 4890 and the chart has been updated to reflect it. The 9.3 drivers are not any faster than the 8.12 HotFix for the other AMD cards in every test I have run but Crysis Warhead with a Core 2 Quad. A few improvements have been made for CF compatibility and video playback though.

    2. The conclusion has been updated to clarify our thoughts between the two cards.
  • can - Thursday, April 2, 2009 - link

    An ATI gpu with and nvidia gpu doing physx? I'm curious to see results of this kind of arrangement. Not a dedicated PCI Physx card, but on a faster bus, with a more powerful processor, as a video card. I'm wondering about pitfalls and performance and the literal looks of the application.
  • SiliconDoc - Monday, April 6, 2009 - link

    Sorry bub, you're stuck with ati, and as far as curiosity for physx - uhh... don't worry, you're not missing much, anand only got addicted to it for a bit.
    If you want the driver hack for it, there's a thread at techpowerup.
    Some genius figured something out on it- not sure which os.
  • Jamahl - Thursday, April 2, 2009 - link

    The conclusion in this review is awful beyond anything I have read before.

    How can the reviewer say the 275 is winning this one when the benchmarks clearly show dominance for the 4890 at most resolutions?

    Otherwise it was a good article but the conclusion leaves a sour taste in the mouth.
  • SiliconDoc - Monday, April 6, 2009 - link

    He could say it because he said it for ati for 6 months when ati won the top resolution. So his brain is in a "fart mode" that lied for ati for so long, he said it this time for nvidia - either that or he realized if he didn't he would look like an exposed raging red rooster fanboy.
    Good thing the reds started screaming NOW, after loving it for 6 months when their card was on top using the false method - because anand came in and saved the day - and changed the conclusion - for ati.
    LOL
    When nvidia doesn't give them a card for review again, it will be "them towing the line of honesty" that causes, no doubt, right ?
    BWAHAHAHAAAAAAA
    ( you all just tell yourselves that along with our dear leaders )
  • erikejw - Thursday, April 2, 2009 - link

    "On the NVIDIA side, we received a reference version of the GTX 275."

    You wish.
    "since there is no 275 ASIC, NV is telling OEMs that they can make it from either a 280 or 260 board. One costs much more, and one performs better, so guess what everyone is going to use?

    That isn't necessarily bad, but how NV is seeding reviewers is. They are only going to be giving out a very special run of ONLY 280 based parts.

    Quite special 280 based parts at that. Reviewers beware, what you are getting is not what you can buy."

    http://www.theinquirer.net/inquirer/news/599/10515...">http://www.theinquirer.net/inquirer/new...ia-hoodw...
  • bill3 - Thursday, April 2, 2009 - link

    A lot of people seem to be crying about lack of temp, power consumption, oc, and fan noise numbers..while I agree in a stand alone review these are glaring omissions, the fact is theres a dozen reviews around the web where you can get that info in triplicate. I would much rather have the insight on CUDA and PHysx!

    I mean, people act like the internet isnt free and we all arent a google search/mouse click away from that type of info! Geez.

    That said, I suppose reviews must be treated as "stand alone", however artificial a construct it may be. However if theres anything thats easily forgivable to be left out it's simple data numbers that can be found at a thousand other places. Which is exactly what temp, oc, etc are. I already know those numbers from a ton of other reviews. These people whining in the comments act like Anand is the only hardware review site there is. I would think if anybody was truly interested in laying out 250 to purchase one of these theyd be looking at more than one review!
  • SkullOne - Thursday, April 2, 2009 - link

    I second the question about why did you use Catalyst 8.12 Hotfix? Other sites are using what appears to be an Beta Catalyst 9.4 driver so is your listing of Catalyst 8.12 a misprint?

    Also why do you care if AMD sent you an overclocked version? The HD4890 is directly targeted at the overclocking enthusiasts which is a realm that AMD has ignored up until now while NV embraced it.

    The HD4890 has already been taken to 1+GHz on it's GPU and up to 4.8GHz on it's memory on other sites. That by far makes it the better buy.
  • SkullOne - Thursday, April 2, 2009 - link

    Forgot to mention that by having this extremely overclockable card AMD has opened up another entire SKU for themselves by selling "OC" cards with the 4890.

Log in

Don't have an account? Sign up now