Gaming Performance
As usual, gaming performance was tested with a variety of current games. We ran benchmarks at a 1024x768 resolution. Any more than this resulted in completely unplayable game conditions, although we did find several older non-FPS games such as the Civilization series, Railroad Tycoon 3, Rollercoaster Tycoon series, and others played acceptability at 1280x1024 with the quality settings at normal. We concentrated on Direct X games as OpenGL support under Vista is bad at this time.
Battlefield 2
This benchmark is performed using DICE's built-in demo playback functionality with additional capture capabilities designed in house. During the benchmark, the camera switches between players and vehicles in order to capture the most action possible. There is a significant amount of smoke, explosions, and vehicle usage as this a very GPU intensive Battlefield 2 benchmark. We run Battlefield 2 using medium quality graphics settings available in the video settings. The game itself is best experienced with average in-game frame rates of 35 and up.
Half-Life 2: Lost Coast
We use the built-in timedemo feature to benchmark the game. Our timedemo consists of starting at the bottom of the hill near the lake and ending in the old church. The Source engine timedemo feature is similar to the nettimedemo of Id's Doom 3 engine, in that it plays back more than just the graphics. The visual quality settings were set to high or medium where possible with HDR off. While the Source engine is notorious for giving great frame rates for almost any hardware setup, we find the game isn't as enjoyable if it isn't running at 35fps or above.
Company of Heroes
Company of Heroes was recently released and is proving to be a very addictive RTS game around the office. This game is very GPU intensive and also requires a hefty CPU at times. The game contains a built-in performance test that utilizes the game engine to generate several different action scenes along with a coffee argument as a sideline distraction before the war starts. We found the performance test gives a good indication of how well your system will perform throughout the game on average. Some of the in-game action sequences are more demanding than the performance test, but we generally found the game to be enjoyable with an average performance test frame rate above 30fps.
Sims 2
Sims 2 was released over two years ago and is constantly being updated with best selling expansion packs. In testing with the various expansion packs we did not notice any measurable differences in performance so our benchmark will be reflective of game play using the base game title. This particular game requires a decent CPU and very good GPU when utilizing the antialiasing, shadow, or high quality texture options.
We utilize FRAPS to capture the results from our replay file. Our benchmark consists of a three character scenario that takes approximately twenty minutes to set up and cycle through a series of daily events. The camera movements are varied as is the movement so we feel like this benchmark represents the typical game. We set most video options to high but disable shadows. We generally found the game to be enjoyable with an average frame rate above 30fps.
Gaming Summary
What else can you really say after reviewing the results? They are generally pathetic from a gaming perspective but the 690G obviously has some potential left in it at the lower resolutions. The 6150 performs okay considering the age of its core and we will see the new 6150SE and older 6100 chipset performing a few percent better overall but not enough to catch the 690G.
Considering the G965 was launched last fall and promised to bring about a difference in on-board video performance, we are still not seeing the results live up to the hype. The G965 was incapable of running Battlefield 2 at all and Half-Life 2 was an interesting slide show that allowed you to see every detail of the benchmark session in an excruciatingly slow way that would make a visit to the dentist for a root canal seem pleasant. However, G965 was able to generate decent results (for this grouping) in Company of Heroes and Sims 2, though the minimum frame rates in Sims 2 would bring the game to a slow crawl at times. Image quality generally favored the 690G in the majority of games we tried but the NVIDIA 6150 was close in most titles. The G965 had acceptable image quality but we could tell certain details were not as sharp or even evident when compared to the competing AM2 solutions.
As usual, gaming performance was tested with a variety of current games. We ran benchmarks at a 1024x768 resolution. Any more than this resulted in completely unplayable game conditions, although we did find several older non-FPS games such as the Civilization series, Railroad Tycoon 3, Rollercoaster Tycoon series, and others played acceptability at 1280x1024 with the quality settings at normal. We concentrated on Direct X games as OpenGL support under Vista is bad at this time.
Battlefield 2
This benchmark is performed using DICE's built-in demo playback functionality with additional capture capabilities designed in house. During the benchmark, the camera switches between players and vehicles in order to capture the most action possible. There is a significant amount of smoke, explosions, and vehicle usage as this a very GPU intensive Battlefield 2 benchmark. We run Battlefield 2 using medium quality graphics settings available in the video settings. The game itself is best experienced with average in-game frame rates of 35 and up.
Half-Life 2: Lost Coast
We use the built-in timedemo feature to benchmark the game. Our timedemo consists of starting at the bottom of the hill near the lake and ending in the old church. The Source engine timedemo feature is similar to the nettimedemo of Id's Doom 3 engine, in that it plays back more than just the graphics. The visual quality settings were set to high or medium where possible with HDR off. While the Source engine is notorious for giving great frame rates for almost any hardware setup, we find the game isn't as enjoyable if it isn't running at 35fps or above.
Company of Heroes
Company of Heroes was recently released and is proving to be a very addictive RTS game around the office. This game is very GPU intensive and also requires a hefty CPU at times. The game contains a built-in performance test that utilizes the game engine to generate several different action scenes along with a coffee argument as a sideline distraction before the war starts. We found the performance test gives a good indication of how well your system will perform throughout the game on average. Some of the in-game action sequences are more demanding than the performance test, but we generally found the game to be enjoyable with an average performance test frame rate above 30fps.
Sims 2
Sims 2 was released over two years ago and is constantly being updated with best selling expansion packs. In testing with the various expansion packs we did not notice any measurable differences in performance so our benchmark will be reflective of game play using the base game title. This particular game requires a decent CPU and very good GPU when utilizing the antialiasing, shadow, or high quality texture options.
We utilize FRAPS to capture the results from our replay file. Our benchmark consists of a three character scenario that takes approximately twenty minutes to set up and cycle through a series of daily events. The camera movements are varied as is the movement so we feel like this benchmark represents the typical game. We set most video options to high but disable shadows. We generally found the game to be enjoyable with an average frame rate above 30fps.
Gaming Summary
What else can you really say after reviewing the results? They are generally pathetic from a gaming perspective but the 690G obviously has some potential left in it at the lower resolutions. The 6150 performs okay considering the age of its core and we will see the new 6150SE and older 6100 chipset performing a few percent better overall but not enough to catch the 690G.
Considering the G965 was launched last fall and promised to bring about a difference in on-board video performance, we are still not seeing the results live up to the hype. The G965 was incapable of running Battlefield 2 at all and Half-Life 2 was an interesting slide show that allowed you to see every detail of the benchmark session in an excruciatingly slow way that would make a visit to the dentist for a root canal seem pleasant. However, G965 was able to generate decent results (for this grouping) in Company of Heroes and Sims 2, though the minimum frame rates in Sims 2 would bring the game to a slow crawl at times. Image quality generally favored the 690G in the majority of games we tried but the NVIDIA 6150 was close in most titles. The G965 had acceptable image quality but we could tell certain details were not as sharp or even evident when compared to the competing AM2 solutions.
70 Comments
View All Comments
goinginstyle - Tuesday, March 6, 2007 - link
I think most of the people missed the comments or observations in the article. The article was geared to proving or disproving the capabilities of the 690g and in a way the competing platforms. It was obvious to me the office crowd was not being addressed in this article and it was the home audience that the tests were geared towards. I think the separation between the two was correct.The first computer I bought from Gateway was an IGP unit that claimed it would run everything and anything. It did not and pissed me off. After doing some homework I realized where I went wrong and would never again buy an IGP box unless the video and memory is upgraded, even if it is not for gaming. I have several friends who bought computers for their kids when World of WarCraft came out and bitched non-stop at work because their new Dell or HP would not run the game. At least the author had the balls to state what many of us think. The article was fair and thorough in my opinion although I was hoping to see some 1080P screen shots. Hint Hint
Final Hamlet - Tuesday, March 6, 2007 - link
Too bad one can't edit one's comments...My point (besides correcting a mistake) is, that I think that this test is gravely imbalanced... you are testing - as you have said yourself - an office chipset - then why do you do it with an overpowered CPU?
Office PC's in small businesses go after price and where is the difference in using a mail program between a Core 2 Duo for 1000$ and the smallest and cheapest AMD offering for less than 100$?
Gary Key - Tuesday, March 6, 2007 - link
We were not testing an office chipset. We are testing chipsets marketed as an all in solution to the home, home/office, multimedia, HTPC, and casual gaming crowd. The office chipsets are the Q965/963 and 690V solutions. The G965 and 690G are not targeted to the office workers and were not tested as such. Our goal was to test these boards in the environment and with applications they are marketed to run.
JarredWalton - Tuesday, March 6, 2007 - link
We mentioned this above, but basically we were looking to keep platform costs equal. Sure, X2 3800+ is half as expensive and about 30% slower than the 5200+. But since the Intel side was going to get an E6300 (that's what we had available), the use of a low-end AMD X2 would have skewed results the other direction. We could have used an X2 4800+ to keep costs closer, but that's an odd CPU choice as well as we would recommend spending the extra $15 to get the 5200+.The intent was not to do a strict CPU-to-CPU comparison as we've done that plenty (as recently as the http://www.anandtech.com/cpuchipsets/showdoc.aspx?...">X2 6000+ launch). We wanted to look at platform and keep them relatively equal in the cost department. All you have to do is look at the power numbers to see that the 5200+ with 690G compares quite well (and quiet well) to the E6300 with G965.
The major selling point of this chipset is basically that it supports HDMI output. That's nice, and for HTPC users it could be a good choice. Outside of that specific market, though, there's not a whole lot to put this IGP chipset above other offerings. That was what we were hoping to convey with the article. It's not bad, but neither is it the greatest thing since sliced bread.
If you care at all about GPU performance, all of the modern IGP solutions are too slow. If you don't care, then they're all fast enough to do whatever most people need. For typical business applications, the vast majority of companies are still running Pentium 4, simply because it is more than sufficient. New PCs are now coming with Core 2 Duo, but I know at least a few major corporations that have hundreds of thousands of P4 and P3 systems in use, and I'm sure there are plenty more. Needless to say, those corporations probably won't be touching Vista for at least three or four years - one of them only switched to XP as recently as two years back.
JarredWalton - Tuesday, March 6, 2007 - link
Perhaps it's because the companies releasing these products make so much noise about how much better their new IGP is compared to the older offerings from their competitors? If AMD had released this and said, "This is just a minor update to our previous IGP to improve features and video quality; it is not dramatically faster and is not intended for games" then we would cut them some slack. When all of the companies involved are going on about how much faster percentage-wise they are than the competition (never mind that it's 5 FPS vs. 4 FPS), we're inclined to point out how ludicrous this is. When Intel hypes the DX9 capability of their G965 and yet still can't run most DX9 applications, maybe someone ought to call them on the carpet?Obviously, these low performance IGPs have a place in the business world, but Vista is now placing more of a demand on the GPU than ever before, and bare minimum functionality might now be adequate for a lot of people. As for power, isn't it interesting that the HIGHEST PERFORMANCE IGP ends up using the least amount of power? Never mind the fact that Core 2 Duo already has a power advantage over the X2 5200+!
So, while you might like to pull out the names and call us inane 15 year olds, there was certainly thought put into what we said. Just because something works okay doesn't mean it's great, and we are going to point out the flaws in a product regardless of marketing hype. Given how much effort Intel puts into their CPUs, a little bit more out of their IGP and drivers is not too much to ask for.
TA152H - Wednesday, March 7, 2007 - link
Jared,Maybe they didn't intend their products to be tested in the way you did. As someone pointed out, playing at 800 x 600 isn't that bad, and doesn't ruin the experience unless you have an obsession. Incredibly crude games were incredibly fun, so the resolution isn't going to make or break a game, it's the ideas behind it that will.
You can't be serious about what you want AMD to say. You know they can't, they are in competition and stuff like that would be extremely detrimental to them. Percentages are important, because they may not running the same games as you are, at the same settings. You would prefer they use absolutes as if they would give more information? Did AMD actually tell anyone these were excellent for all types of game? I never saw that.
With regards to CPUs and GPUs, you are trying to obfuscate the point. Everyone uses a CPU, some more than others. But, they do sell lower power ones, and even single core ones. Not everyone uses 3D functionality. If you don't get it, I DON'T want it on certain machines of mine. I don't run stuff like that on them, and I don't want the higher power use or heat dissipation problems from it. What you call effort isn't at all, it's a tradeoff. Don't confuse it with you get something for nothing if Intel puts more into it. You pay for it, and that's the problem. People who use it should, people that don't, shouldn't, so the kiddies can play their shoot 'em ups.
Just so you know, I'm both. I have mostly work machines, but two play machines. I like playing some games that require a good 3D card, but just don't like the mentality the the whole world should subsidize a bunch of gameplayers when they don't need it. That's what add-in cards are for. I would be equally against it if no one made 3D cards because most people didn't need them. I like choices, and I don't want to pay for excessive 3D functionality on something that will never use it, to help gameplayers out. Both existing is great, and IGPs will creep up as they always have, when it becomes inexpensive (both in power and initial cost) to add capabality, so the tradeoff is minor.
StriderGT - Tuesday, March 6, 2007 - link
Does this chipset support 5.1 LPCM over HDMI or not??? Or more plainly can someone send 5.1 (games, HD movies, etc) digitally to receiver with the 690G? According to your previous article on the 690G 5.1 48khz was supported over the HDMI port. Now its back to 2 channel and AC3 bitstream. Which is it?Gary Key - Wednesday, March 7, 2007 - link
It is two channel plus AC3 over HDMI. That is the final spec on production level boards and drivers. We will have a full audio review up in a week or so that also utilizes the on-board codec.StriderGT - Thursday, March 8, 2007 - link
Why is this happening? Why on earth can't they produce a PC HDMI Audio solution that outputs up to 7.1 LPCM (96khz/24bit) for ALL sources!?! They already do that for 2 channel sources!!!! Do you have any info from the hardware vendors regarding the reason/s they will not produce such a straightforward and simple solution?!?PS There are lots of people demanding a TRUE PC HDMI Audio solution not this SPDIF hacks...
Renoir - Tuesday, March 6, 2007 - link
I'm also interested to know more specifics about the audio side of this chipset. The support of HDMI v1.3 suggests that with an appropriate driver and supporting playback software Dolby TrueHD and DTS-HD bitstreams should be able to be sent via HDMI to a v1.3 receiver with the necessary decoders. Is this a possibility?