Angle-Independent Anisotropic Filtering At Last

For a number of years now the quality of anisotropic filtering has been slowly improving. Early implementations from AMD and NVIDIA were highly angle-dependent, resulting in a limited improvement to image quality from such filtering. The angle-dependent nature lead to shimmering and other artifacting that was not ideal.

As of the previous generation of cards, the quality of anisotropic filtering had become pretty good. NVIDIA’s best filtering mode was pretty close to angle-independent, and AMD’s only slightly worse. Neither was perfect, but neither was bad either.


The Radeon HD 4890


The GeForce GTX 285

However so long as no one had an angle-independent implementation, there was room to improve. And AMD has gone there. The anisotropic filtering algorithm used by the 5000 series is now truly and completely angle-independent. There are no more filtering tricks being used.


The Radeon HD 5870: Perfection

As you can see, the MIP maps in our venerable D3D AF Tester are perfectly circular, the hallmark of an angle-independent implementation. With angle-independent filtering, this effectively marks the end of the filtering arms race. AMD has won, and should NVIDIA catch up in the future the two would merely be tied. There’s nowhere left to go for quality beyond angle-independent filtering at the moment.

AMD tells us that there is no performance hit with their new algorithm compared to their old one. This is a bit hard to test since we can’t enable the old algorithm on the 5870, but certainly whatever performance hit there is, is similarly minor. In all of the testing we’re doing today, you will see results done with 16x anisotropic filtering used.

What you won’t see however is a difference, particularly with our static screenshots. When discussing the matter, AMD noted that the difference in perceived quality between the old algorithm and the new one was practically the same. After looking at matters we find ourselves in agreement with AMD; we were not able to come up with any situations where there was a noticeable difference, beyond the obvious AF quality tests that are designed to identify such changes.

Regardless of the outcome, AMD deserves kudos for making angle-independent anisotropic filtering happen. It’s demonstrably perfect filtering with no speed hit versus the previous generation of filtering; making it in essence a “free” improvement in image quality, however slight the real-world results are. We’re always ready to get better image quality out of our video cards, after all.

More GDDR5 Technologies: Memory Error Detection & Temperature Compensation The Return of Supersample AA
Comments Locked

327 Comments

View All Comments

  • SiliconDoc - Thursday, September 24, 2009 - link

    Are you seriously going to claim that all ATI are not generally hotter than the nvidia cards ? I don't think you really want to do that, no matter how much you wail about fan speeds.
    The numbers have been here for a long time and they are all over the net.
    When you have a smaller die cranking out the same framerate/video, there is simply no getting around it.
    You talked about the 295, as it really is the only nvidia that compares to the ati card in this review in terms of load temp, PERIOD.
    In any other sense, the GT8800 would be laughed off the pages comparing it to the 5870.
    Furthermore, one merely needs to look at the WATTAGE of the cards, and that is more than a plenty accurate measuring stick for heat on load, divided by surface area of the core.
    No, I'm not the one not thinking, I'm not the one TROLLING, the TROLLING is in the ARTICLE, and the YEAR plus of covering up LIES we've had concerning this very issue.
    Nvidia cards run cooler, ati cards run hotter, PERIOD.
    You people want it in every direction, with every lying whine for your red god, so pick one or the other:
    1.The core sizes are equivalent, or 2. the giant expensive dies of nvidia run cooler compared to the "efficient" "new technology" "packing the data in" smaller, tiny, cheap, profit margin producing ATI cores.
    ------
    NOW, it doesn't matter what lies or spin you place upon the facts, the truth is absolutely apparent, and you WON'T be changing the physical laws of the universe with your whining spin for ati, and neither will the trolling in the article. I'm going to stick my head in the sand and SCREAM LOUDLY because I CAN'T HANDLE anyone with a lick of intelligence NOT AGREEING WITH ME! I LOVE TO LIE AND TYPE IN CAPS BECAUSE THAT'S HOW WE ROLL IN ILLINOIS!
  • SiliconDoc - Friday, September 25, 2009 - link

    Well that is amazing, now a mod or site master has edited my text.
    Wow.
  • erple2 - Friday, September 25, 2009 - link

    This just gets better and better...

    Ultimately, the true measure of how much waste heat a card generates will have to look at the power draw of the card, tempered with the output work that it's doing (aka FPS in whatever benchmark you're looking at). Since I haven't seen that kind of comparison, it's impossible to say anything at all about the relative heat output of any card. So your conclusions are simply biased towards what you think is important (and that should be abundantly clear).

    Given that one must look at the performance per watt. Since the only wattage figures we have are for OCCT or WoW playing, so that's all the conclusions one can make from this article. Since I didn't see the results from the OCCT test (in a nice, convenient FPS measure), we get the following:

    5870: 73 fps at 295 watts = 247 FPS per milliwatt
    275: 44.3 fps at 317 watts = 140 FPS per milliwatt
    285: 45.7 fps at 323 watts = 137 FPS per milliwatt
    295: 68.9 fps at 380 watts = 181 FPS per milliwatt

    That means that the 5870 wins by at least 36% over the other 3 cards. That means that for this observation, the 5870 is, in fact, the most efficient of these cards. It therefore generates less heat than the other 3 cards. Looking at the temperatures of the cards, that strictly measures the efficiency of the cooler, not the efficiency of the actual card itself.

    You can say that you think that I'm biased, but ultimately, that's the data I have to go on, and therefore that's the conclusions that can be made. Unfortunately, there's nothing in your post (or more or less all of your posts) that can be verified by any of the information gleaned from the article, and therefore, your conclusions are simply biased speculation.
  • SiliconDoc - Saturday, September 26, 2009 - link

    4780, 55nm, 256mm die, 150watts HOT
    G260, 55nm, 576mm die, 171watts COLD
    3870, 55nm, 192mm die, 106watts HOT

    That's all the further I should have to go.
    3870 has THE LOWEST LOAD POWER USEAGE ON THE CHARTS
    - but it is still 90C, at the very peak of heat,
    because it has THE TINIEST CORE !
    THE SMALLEST CORE IN THE WHOLE DANG BEJEEBER ARTICLE !
    It also has the lowest framerate - so there goes that erple theory.
    ---
    The anomlies you will notice if you look, are due to nm size, memory amount on board (less electricity used by the memory means the core used more), and one slot vs two slot coolers, as examples, but the basic laws of physics cannot be thrown out the window because you feel like doing it, nor can idiotic ideas like framerate come close to predicting core temp and it's heat density at load.
    Older cpu's may have horrible framerates and horribly high temps, for instance. The 4850 frames do not equal the 4870's, but their core temp/heat density envelope is very close to indentical ( SAME CORE SIZE > the 4850 having some die shaders disabled and ddr3, the 4870 with ddr5 full core active more watts for mem and shaders, but the same PHYSICAL ISSUES - small core, high wattage for area, high heat)
  • erple2 - Tuesday, September 29, 2009 - link

    I didn't say that the 3870 was the most efficient card. I was talking about the 5870. If you actually read what I had typed, I did mention that you have to look at how much work the card is doing while consuming that amount of power, not just temperatures and wattage.

    You sir, are a Nazi.

    Actually, once you start talking about heat density at load, you MUST look at the efficiency of the card at converting electricity into whatever it's supposed to be doing (other than heating your office). Sadly, the only real way that we have to abstractly measure the work the card is doing is "FPS". I'm not saying that FPS predict core temperature.
  • SiliconDoc - Wednesday, September 30, 2009 - link

    No, the efficiency of conversion you talk about has NOTHING to do with core temp AT ALL. The card could be massively efficient or inefficient at produced framerate, or just ERROR OUT with a sick loop in the core, and THAT HAS ABSOLUTELY NOTHING TO DO WITH THE CORE TEMP. IT RESTS ON WATTS CONSUMED EVEN IF FRAMERATE OUTPUT IS ZERO OR 300SECOND.
    (your mind seems to have imagined that if the red god is slinging massive frames "out the dvi port" a giant surge of electricity flows through it to the monitor, and therefore "does not heat the card")

    I suggest you examine that lunatic red notion.

    What YOU must look at is a red rooster rooter rimshot, in order that your self deception and massive mistake and face saving is in place, for you. At least JaredWalton had the sense to quietly skitter away.
    Well, being wrong forever and never realizing a thing is perhaps the worst road to take.

    PS - Being correct and making sure the truth is defended has nothing to do with some REDEYE cleche, and I certainly doubt the Gregalouge would embrace red rooster canada card bottom line crumbled for years ever more in a row, and diss big green corporate profits, as we both obviously know.

    " at converting electricity into whatever it's supposed to be doing (other than heating your office). "
    ONCE IT CONVERTS ELECTRICITY, AS IN "SHOWS IT USED MORE WATTS" it doesn't matter one ding dang smidgen what framerate is,

    it could loop sand in the core and give you NO screeen output,

    and it would still heat up while it "sat on it's lazy", tarding upon itself.

    The card does not POWER the monitor and have the monitor carry more and more of the heat burden if the GPU sends out some sizzly framerates and the "non-used up watts" don't go sailing out the cards connector to the monitor so that "heat generation winds up somewhere else".

    When the programmers optimize a DRIVER, and the same GPU core suddenly sends out 5 more fps everything else being the same, it may or may not increase or decrease POWER USEAGE. It can go ANY WAY. Up, down, or stay the same.
    If they code in more proper "buffer fills" so the core is hammered solid, instead of flakey filling, the framerate goes up - and so does the temp!
    If they optimize for instance, an algorythm that better predicts what does not need to be drawn as it rests behind another image on top of it, framerate goes up, while temp and wattage used GOES DOWN.
    ---
    Even with all of that, THERE IS ONLY ONE PLACE FOR THE HEAT TO ARISE... AND IT AIN'T OUT THE DANG CABLE TO THE MONITOR!
  • SiliconDoc - Friday, September 25, 2009 - link

    You can modify that, or be more accurate, by using core mass, (including thickness of the competing dies) - since the core mass is what consumes the electricity, and generates heat. A smaller mass (or die size, almost exclusively referred to in terms of surface area with the assumption that thickness is identical or near so) winds up getting hotter in terms of degrees of Celcius when consuming a similar amount of electricity.
    Doesn't matter if one frame, none, or a thousand reach your eyes on the monitor.
    That's reality, not hokum. That's why ATI cores run hotter, they are smaller and consume a similar amount of electricty, that winds up as heat in a smaller mass, that means hotter.
    Also, in actuality, the ATI heatsinks in a general sense, have to be able to dissipate more heat with less surface area as a transfer medium, to maintain the same core temps as the larger nvidia cores and HS areas, so indeed, should actually be "better stock" fans and HS.
    I suspect they are slightly better as a general rule, but fail to excel enough to bring core load temps to nvidia general levels.
  • erple2 - Friday, September 25, 2009 - link

    You understand that if there were no heatsink/cooling device on a GPU, it would heat up to crazy levels, far more than would be "healthy" for any silicon part, right? And you understand that measuring the efficiency of a part involves a pretty strong correlation between the input power draw of the card vs. the work that the card produces (which we can really only measure based on the output of the card, namely FPS), right?

    So I'm not sure that your argument means anything at all?

    Curiously, the output wattage listed is for the entire system, not just for the card. Which means that the actual differences between the ATI cards vs. the nvidia cards is even larger (as a percentage, at least). I don't know what the "baseline" power consumption of the system (sans video card) is for the system acting as the test bed is.

    Ultimately, the amount of electricity running through the GPU doesn't necessarily tell you how much heat the processors generate. It's dependent on how much of that power is "wasted" as heat energy (that's Thermodynamics for you). The only way to really measure the heat production of the GPU is to determine how much power is "wasted" as heat. Curiously, you can't measure that by measuring the temperature of the GPU. Well, you CAN, but you'd have to remove the Heatsink (and Fan). Which, for ANY GPU made in the last 15 years, would cook it. Since that's not a viable alternative, you simply can't make broad conclusions about which chip is "hotter" than another. And that is why your conclusions are inconclusive.

    BTW, the 5870 consumes "less" power than the 275, 285 and 295 GPUs (at least, when playing WoW).

    I understand that there may be higher wattage per square millimeter flowing through the 5870 than the GTX cards, but I don't see how that measurement alone is enough to state whether the 5870 actually gets hotter.
  • SiliconDoc - Saturday, September 26, 2009 - link

    Take a look at SIZE my friend.
    http://www.hardforum.com/showthread.php?t=1325165">http://www.hardforum.com/showthread.php?t=1325165

    There's just no getting around the fact that the more joules of heat in any time period (wattage used!= amount of joules over time!) that go into a smaller area, the hotter it gets, faster !

    Nothing changes this, no red rooster imagination will ever change it.
  • SiliconDoc - Saturday, September 26, 2009 - link

    NO, WRONG.
    " Ultimately, the true measure of how much waste heat a card generates will have to look at the power draw of the card, tempered with the output work that it's doing (aka FPS in whatever benchmark you're looking at)."
    NO, WRONG.
    ---
    Look at any of the cards power draw in idle or load. They heat up no matter how much "work" you claim they do, by looking at any framerate, because they don't draw the power unless they USE THE POWER. That's the law that includes what useage of electricity MEANS for the law of thermodynamics, or for E=MC2.
    DUHHHHH.
    ---
    If you're so bent on making idiotic calculations and applying them to the wrong ideas and conclusions, why don't you take core die size and divide by watts (the watts the companies issue or take it from the load charts), like you should ?
    I know why. We all know why.
    ---
    The same thing is beyond absolutely apparent in CPU's, their TDP, their die size, and their heat envelope, including their nm design size.
    DUHHH. It's like talking to a red fanboy who cannot face reality, once again.

Log in

Don't have an account? Sign up now