PhysX in Warmonger: Fail

Cryostasis is a title due out this year, unfortunately there is no playable demo. Just a tech demo. Next.

Metal Knight Zero, MKZ for short, was another game on NVIDIA’s list. Once more, no playable demo, just a tech demo. We need real games here people, real titles, if you’re trying to convince someone to buy NVIDIA on the merits of PhysX.

Warmonger, ah yes, now we have a playable game. Warmonger is a first person shooter that uses GPU accelerated PhysX to enable destructible environments. Allow me to quote NVIDIA:

The first thing about Warmonger is that it runs horribly slow on ATI hardware, even with GPU accelerated PhysX disabled. I’m guessing ATI’s developer relations team hasn’t done much to optimize the shaders for Radeon HD hardware. Go figure.

The verdict here (aside from: I don’t want to play Warmonger), was that the GPU accelerated PhysX effects were not very, well, impressive. You could destroy walls, but the game itself wasn’t exactly fun so it didn’t matter. The realistic cloth that you could shoot holes through? Yeah, not terribly realistic looking.


Look at the hyper realistic cloth! Yeah, it looks like a highly advanced game from 6 years ago.

Warmonger itself wasn’t a triple A first person shooter, and the GPU accelerated PhysX effects on top of it weren’t going to make the game any better. Sorry guys, none of us liked this one. PC Gamer gave it a 55/100. Looks like we weren’t alone. Next.

PhysX in Sacred 2: There, but not tremendously valuable The Unreal Tournament 3 PhysX Mod Pack: Finally, a Major Title
Comments Locked

294 Comments

View All Comments

  • sbuckler - Thursday, April 2, 2009 - link

    Big difference between Havok Physics and HavokFX physics. With physx you can just turn on hardware acceleration and it works, with havok this is not possible - unlike physx it was never developed to be run on the gpu. Hence havok have had to develop a new physics engine to do that.

    No game uses the HavokFX engine - it's not even available to developers yet let alone in shipped games. The ati demo was all we have seen of it for several years. It's not even clear HavokFX is even a fully accelerated hardware physics engine - i.e. the version showed in the past (before intel took over havok) was basically the havok engine with some hw acceleration for effects. i.e. hardware accel could only be used to make it prettier explosions and rippling cloth - it could not be used to do anything game changing.

    Hence havok have a way to go before they can even claim to support what physX already does, let alone shipping it to developers and then seeing them use it in games. Like I said the moment that comes close to happening nvidia will just release an OpenCL version of physX and that will be that.
  • z3R0C00L - Thursday, April 2, 2009 - link

    It's integrated in the same way. Many game developers are already familiar with coding for Havok effects.

    Not to mention that OpenCL has chosen HavokFX (which is simply using either a CPU or a GPU to render Physics effect as seen here: http://www.youtube.com/watch?v=MCaGb40Bz58">http://www.youtube.com/watch?v=MCaGb40Bz58.

    Again... Physx is dead. OpenCL is HavokFX, it's what the consortium has chosen and it runs on any CPU or GPU including Intel's upcoming Larrabee.

    Like I said before (you seem to not understand logic). Physx is dead.. it's proprietary and not as flexible as Havok. Many studios are also familiar with Havok's tools.

    C'est Fini as they say in french.
  • erple2 - Friday, April 3, 2009 - link

    I think you're mistaken - OpenCL is analogous to CUDA, not to PhysX. HavokFX is analogous to PhysX. OpenCL is the GPGPU compiler that runs on any GPU (and theoretically, it should run on any CPU too, I think). It's what Apple is now trying to push (curious, given that their laptop base is all nVidia now).

    However, if NVidia ports PhysX to OpenCL, that's a win for everyone. Sort of. Except for NVIdia that paid a lot of money for the PhysX IP. I think that the conclusions given are accurate - NVidia is banking on "everyone" (ie Game Developers) coding for PhysX (and by extension, CUDA) rather than HavokFX (and by extension, OpenCL). However, if Developers are smart, they'll go with the actually open format (OpenCL, not CUDA). That means that any physics processing they do will work on ANY GPU, (NVidia and ATI). I personally think that NVidia banked badly this time.

    While I do believe that doing physics calculations on unused GPU cycles is a great thing (and the Razor's Edge demo shows some of the interesting things that can be done), I think that NVidia's pushing of PhysX (and therefore CUDA) is like what 3dfx did with pushing GLide. Everyone supported Direct3D and OpenGL, but only 3dfx supported Glide. While Glide was more efficient (it was catering to a single hardware vendor that made Glide, afterall), the fact that Game Developers could instead program for OpenGL (or Direct3D) and get all 3D accelerators supported meant that the days of Glide were ultimately numbered.

    I wonder if NVidia is trying to pull the industry to adopting its CUDA as a "standard". I think it's ultimately going to fail, however, given that the industry recognizes now that OpenCL is available.

    Is OpenCL as mature as CUDA is? Or are they still kind of finalizing it? Maybe that's the issue - OpenCL isn't complete yet, so NVidia is trying to snatch up support in the Developer community early?
  • sbuckler - Friday, April 3, 2009 - link

    CUDA is in many ways a simplified version of OpenCL - in that CUDA knows what hardware it will run on so has set functions to access it, OpenCL is obviously much more generic as it has to run on any hardware so it's not quite as easy. That part of the reason why CUDA is initially at least more popular then OpenCL - it's easier to work with. That said they are very similar so to port from one to the other won't be hard - hence develop for CUDA now then just port to OpenCL when the market demands it.

    All in my opinion Ati want is their hardware to run with whatever physics standard is out there. Right now they are at a growing competitive disadvantage as hardware physics slowly takes off. Hence they demo HavokFX in the hope that either (a) it takes off or (b) nvidia are forced to port PhysX to openCL. I don't think they care which one wins - both products belong to a competitor.

    Nvidia who have put a lot of money into PhysX want to maximise their investment so they will keep PhysX closed as long as possible to get people to buy their cards, but in the end I am sure they are fully aware they will have to open it up to everyone - it's just a matter of when. From our standpoint the sooner the better.
  • erple2 - Friday, April 3, 2009 - link

    Sure, but my point was simply that HavokFX and PhysX are physics API's, whereas OpenCL and CUDA are "general" purpose computing languages designed to run on a GPU.

    Is CUDA easier to work with? I don't really know, as I've never programmed for either. Is OpenGL harder to program for than Glide was? Again, I don't know, I'm not a developer.

    ATI's "CUDA" was "Stream" (I think). I recall ATI abandoning that for (or folding that into) OpenCL. That's a sound strategic decision, I think.

    If PhysX is ported to OpenCL, then that's a major win for ATI, and a lesser one for NVidia - the PhysX SDK is already free for any developer that wants it (support costs money, of course). NVidia's position in that market is that PhysX currently only works on NVidia cards. Once it works elsewhere (via OpenCL or Stream), NVidia loses that "edge". However, that's a good thing...
  • SiliconDoc - Monday, April 6, 2009 - link

    I guess you're forgetting that recently NVidia supported a rogue software coder that was porting PhysX to ATI drivers. Those drivers hit the web and the downloads went wild high - and ATI stepped in and slammed the door shut with a lawsuit and threats.
    Oh well, ATI didn't want you to enjoy PhysX effects. You got screwed, even as NVidia tried to help you.
    So now all you and Derek and anand are left to do is sourpuss and whine PhysX sucks and means nothing.
    Then anand tries Mirror's Edge ( because he HAS TO - cause dingo is gone - unavailable ) and falls in love with PhysX and the game. LOL
    His conclusion ? He's an ati fabbyboi so cannot recommend it.
  • tamalero - Monday, April 20, 2009 - link

    the ammount of fud you spit is staggering
  • z3R0C00L - Thursday, April 2, 2009 - link

    On one hand you have OpenCL, Havok, ATi, AMD and Intel on the other you have nVIDIA.

    Seriously.
  • z3R0C00L - Thursday, April 2, 2009 - link

    I'm an nVIDIA fan.. I'll admit. I like that you added CUDA and Physx.. but are we reading the same results?

    The Radeon HD 4890 is the clear winner here. I don't understand how it could be any different.

  • CrystalBay - Thursday, April 2, 2009 - link

    I agree if nV wants to sell more cards they need to include the video software at no charge...

Log in

Don't have an account? Sign up now