40nm Battle Heats Up

Page 5 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

chizow

Diamond Member
Jun 26, 2001
9,537
2
0
Originally posted by: qbfx
Ok, If there wasn't the possibility you don't see well, I'd think you're lying. Before typing anything about the tests, check if these are right or wrong because I've no intenstion to list the results again:
Yep, the GTX 295 suffers in bandwidth or VRAM limited resolutions, that was never in question. But again, what you results don't show is the GTX 295 winning 4 titles with 4xAA at 1280/1680 with the maximum level of AA allowed.

Look at the numbers above, they tell stories.
Yep, it shows you counted a split resolution as a win for ATI in one instance, and a tie for Nvidia in two instances.

Maybe it proves it when it's about AMD/ATi winning
Yes, the 4870X2 shines with 8xAA, that's no different than the 4870 vs 280, its just not winning nearly as much as Performance Rating would indicate.

You can look at every AA/AF mode for every res. as well, that can't summarize the results from 12 titles, can it?
Sure you can, but of course then you'd have to include the 4xAA results for 1280 and 1680, which you failed to list above, for some reason. :roll:

No, by handpicked I mean 5 titles that favor GF hardware and that were the only benches nVidia would let reviewers use during these tests.
Uh, no, they showed guidance telling reviewers their 180 drivers offered a boost in a variety of Top 10 titles, but they were free to review whatever games they wanted with Rel 180. Most sites chose to confirm/deny their claims and found there were significant increases in many of those titles. The GTX 295 preview limited reviewers to 5 titles, but that's really no different than ATI's restrictions on reviewers for their 4870X2 preview.

Anyway, according to these results it's clear that the HD4870X2 wipes the floor with the 20% more expensive GTX295 on the resolutions that actually matter (stop pointing out the GTX295 wins in the "majority" of the tests only because at 1280x1024 noAA/AF the GTX295 actually kicks arse (by 10% :laugh) and don't forget 1920x1200 that's the sweetspot right now.
It does win the majority though once you include no AA or lower AA settings. Yes the 4870X2 performs better at higher resolutions and AA settings where bandwidth and VRAM become an issue, but that doesn't stop this site and others from coming to the conclusion the GTX 295 is the faster overall part.

[/quote]

 

videogames101

Diamond Member
Aug 24, 2005
6,783
27
91
Originally posted by: chizow
Originally posted by: qbfx
Ok, If there wasn't the possibility you don't see well, I'd think you're lying. Before typing anything about the tests, check if these are right or wrong because I've no intenstion to list the results again:
Yep, the GTX 295 suffers in bandwidth or VRAM limited resolutions, that was never in question. But again, what you results don't show is the GTX 295 winning 4 titles with 4xAA at 1280/1680 with the maximum level of AA allowed.

Look at the numbers above, they tell stories.
Yep, it shows you counted a split resolution as a win for ATI in one instance, and a tie for Nvidia in two instances.

Maybe it proves it when it's about AMD/ATi winning
Yes, the 4870X2 shines with 8xAA, that's no different than the 4870 vs 280, its just not winning nearly as much as Performance Rating would indicate.

You can look at every AA/AF mode for every res. as well, that can't summarize the results from 12 titles, can it?
Sure you can, but of course then you'd have to include the 4xAA results for 1280 and 1680, which you failed to list above, for some reason. :roll:

No, by handpicked I mean 5 titles that favor GF hardware and that were the only benches nVidia would let reviewers use during these tests.
Uh, no, they showed guidance telling reviewers their 180 drivers offered a boost in a variety of Top 10 titles, but they were free to review whatever games they wanted with Rel 180. Most sites chose to confirm/deny their claims and found there were significant increases in many of those titles. The GTX 295 preview limited reviewers to 5 titles, but that's really no different than ATI's restrictions on reviewers for their 4870X2 preview.

Anyway, according to these results it's clear that the HD4870X2 wipes the floor with the 20% more expensive GTX295 on the resolutions that actually matter (stop pointing out the GTX295 wins in the "majority" of the tests only because at 1280x1024 noAA/AF the GTX295 actually kicks arse (by 10% :laugh) and don't forget 1920x1200 that's the sweetspot right now.
It does win the majority though once you include no AA or lower AA settings. Yes the 4870X2 performs better at higher resolutions and AA settings where bandwidth and VRAM become an issue, but that doesn't stop this site and others from coming to the conclusion the GTX 295 is the faster overall part.

[/quote]

Guys, it depends on how you like to play it. GTX295 is single-card, 4870x2 is dual card, that is the decision maker. Obviously, if dual cards scare you, get the GTX295, but if they don't, find a really cheap 4870x2.
 

AzN

Banned
Nov 26, 2001
4,112
2
0
Originally posted by: chizow

Already gone over all of this before, you claimed SP and TMUs had the greatest impact on performance with NV parts, and while the G92 clearly benefits from the ~50% increases to both over G80, it performs nowhere close to that much faster. Further, G80 saw a very linear increase to performance by simply increasing core clocks and nothing else. This would lead one to believe that the main difference in G80 to G92 in terms of reductions, ROPs, bandwidth and VRAM are whats holding it back. All areas that were addressed significantly with GT200. What areas weren't addressed as much compared to G92b? SP performance and TMUs. Yet GT200 always outperforms GT92b, and often significantly.

So if you raise core clocks it's only raising ROP clocks right? So you conclude ROP and bandwidth and vram is whats holding it back? :disgust: That's really retarded of you.

When you raise core clocks you are also raising texture clocks and WTF does VRAM have anything to do with it? Long as video card have enough vram to run the desired resolution and settings you don't need more than what's needed other than internet show farting. Games for the most part load and unload textures back from frame buffer which texture fillrate matters more than ROP. ROP is nothing more than running higher resolutions or using AA. It might add a frame or two but no way does it impact in a way TMU or SP would. If you remember back when BFG and I had disagreement over SP vs texture fillrate and bandwidth when G92 was about to release you know that I was the one who said you had to have right balance between bandwidth and fillrate to get efficient results. Then you came around with you don't need more than 256bit bus on a 16ROP then your stupid thread about ROP being the most impact in performance. That was about year and some days ago. Now things have changed a bit. We have quite bit of games where SP matter more so than fillrate.



Its possible G92 with around half the bandwidth is bandwidth limited, but the question is whether or not it benefits more from memory bandwidth or core frequency? Comparing the various G92 and G92b variants, all grounded with 256-bit bus and ~1000-1000MHz memory frequency and my experience with an 8800GT I'd still come to the conclusion core clocks (ROP, set-up, TMU etc) have a greater impact on performance than shader performance or texture fillrate.

You got to be kidding me? The more fillrate you have you need more bandwidth to use all that fillrate efficiently. G92 has whole lot of fillrate. More so than GTX 260 or 216 core. It sure makes HD 4700 faster over 4850 doesn't it and that card has whole lot less fillrate than G92.


First, counting number of transistors and die size for any given core function is a poor metric. For example, L2 cache attributes up to 80% of some Core 2 dice. Yet a chip with 1/2 or even 1/4 the L2 typically performs within 80%. Secondly, there's always going to be design decisions with regard to functional units and the goal is balance. Third, its possible Nvidia missed their target clocks or had other motives for certain design decisions (Tesla, Quadro, GPGPU etc).

You are making wild assumptions if that.

As for why Nvidia didn't address ROP with GT200? They most certainly did.....they doubled ROP compared to G92 which is also a 50% increase over G80. They obviously learned from their mistakes with G92 that simply increasing compute and texture units without increasing ROPs and bandwidth wasn't enough. They didn't kill off texture or SP units, but as has been shown already, the theoreticals for TMU and SP performance with GT200 are very close to G92b due to differences in clock speed.

And look at GT200 performance. It's not much better than g92. never mind the 2x the ROP with 2x the bandwidth just to get 10-40% performance over g92 with 16ROP and 256bit bus. When they Nvidia could have easily accomplished that by adding GDDR5 circuitry in their G92 chip with 1 gig of vram.

Now look at AMD 3870 to 4850. 4850 has same old 16ROP and 256bit bus but they've raised 50-60% performance over 3870 by raising TMU and SP count.



Yes, that's the point, that GTX 280 SLI is always faster even in situations that aren't bandwidth limited, discounting situations where all high-end solutions are CPU bottlenecked. There's more than just the AT and Bit-Tech benches, every other review site that tested GTX 295 came to the conclusion it performs closer to GTX 260 SLI compared to GTX 280 SLI. Certainly bandwidth and VRAM come into play at higher resolutions with AA, but even at lower resolutions with or without AA the GTX 280 SLI is significantly faster when CPU bottlenecking isn't an issue.

Because GTX 280 SLI core is clocked higher and has more bandwidth. With AA biggest factor was bandwidth. Without AA that bandwidth limitation shrink a bit and GTX 295 gain some momentum over GTX 260 SLI and catch upto GTX 280 SLI.


I actually did exactly that with WiC, CoH, Crysis and FC2 last night at 1680 with 4xAA. I'll put actual numbers up in a bit but from preliminary results, cutting bandwidth 26% at 602/1296/868 resulted in less than 5% difference in WiC, CoH and FC2. Crysis showed more difference, ~8-10%. I can guarantee you a 26% increase to core clock would result in more than 5-10% performance gain, unfortunately Precision and RT don't allow me to change the core/shader ratio that much individually.

There you have it. You said it makes absolute no difference at low resolution with AA but bandwidth sure had an impact in low resolution. :roll: Now downclock your core to 576 mhz and do the benchmark again. divide your percentage by Anandtech's GTX 295 vs GTX 280 SLI review there you should see those extra 4 ROP's impact in performance. Of course this is just a rough idea but you should get the point.



How am I contradicting myself? Yes the number of pixels drawn to the frame buffer are the same for any given frame, but they can only be drawn as fast as they're fed to the ROPs, and in cases with AA, post-processing or heavy shading that rate is going to be reduced.

First you say you have less pixels to draw per frame but then you go ahead and say more gpu intensive games or some game with AA the bottleneck isn't fillrate. Does this even make sense?


Its an issue at higher resolutions with AA, but not nearly as much at 1680 with 4xAA.

ROFL.. Seriously. You even benchmarked your GTX 280 that is uber bandwidth saturated and it dropped in performance @ 1680 4xAA... Card like GTX 280 or your old 8800gtx need core clocks because it had the bandwidth but not the fillrate to get balance which is confusing the heck of you. :laugh: It just doesn't have the juice to use all that bandwidth efficiently. Now imagine 9800gtx+ with just as much fillrate as GTX 280 and same amount of bandwidth. it won't be pretty but I think it would literally destroy the old GTX 260 and go head to head with GTX 260 216 core.


If you're comparing your GSO then you also have 25% less bandwidth than a full G92. This is similar to past comparisons with parts crippled with 64-bit or 128-bit memory buses. And yes Crysis has always been bandwidth and VRAM sensitive even at lower resolutions, so its not much of a surprise reducing bandwidth on a part that was already heavily bandwidth limited would have an adverse effect. Its also one of the few titles that is heavily shader intensive. Curious how you were able to unlink core and shader with that much difference though, I don't recall being able to change the ratio that much with my G80s or G92, and certainly not with my 280.

But then it also has 4 less ROP, 16 less TMU, and 32SP. In fact it's exactly 25% cut of the full G92 GPU. So 192bit is bandwidth hungry but 256bit isn't? That makes perfect sense. :roll:

I'll benchmark my 8800gts to show how retarded you sound. Don't give me that shit about Crysis being bandwidth hungry either.

 

AzN

Banned
Nov 26, 2001
4,112
2
0
Crysis 1.2 with system below. 1680x1050 4xAF dx9 high settings 28% less clocks with both memory and core.

Overclocked speeds 756/1836/1026
34.43 fps

CORE REDUCTION 590/1836/1026
32.13 fps -7.2% difference

BANDWIDTH REDUCTION 756/1836/800
29.72 fps -15.8% difference

It sure is bandwidth limited isn't it whether it be 256bit or 192bit. 256bit bus seeing bigger performance drop over 192bit bus. Then again 8800gts core is clocked higher while memory was clocked slightly lower over the 8800gs.
 

chizow

Diamond Member
Jun 26, 2001
9,537
2
0
Originally posted by: Azn
So if you raise core clocks it's only raising ROP clocks right? So you conclude ROP and bandwidth and vram is whats holding it back? :disgust: That's really retarded of you.

When you raise core clocks you are also raising texture clocks and WTF does VRAM have anything to do with it? Long as video card have enough vram to run the desired resolution and settings you don't need more than what's needed other than internet show farting. Games for the most part load and unload textures back from frame buffer which texture fillrate matters more than ROP. ROP is nothing more than running higher resolutions or using AA. It might add a frame or two but no way does it impact in a way TMU or SP would. If you remember back when BFG and I had disagreement over SP vs texture fillrate and bandwidth when G92 was about to release you know that I was the one who said you had to have right balance between bandwidth and fillrate to get efficient results. Then you came around with you don't need more than 256bit bus on a 16ROP then your stupid thread about ROP being the most impact in performance. That was about year and some days ago. Now things have changed a bit. We have quite bit of games where SP matter more so than fillrate.
Nope, raising core clock isn't just increasing ROP performance, it also increases things like staging and set-up however those units are attached to the ROPs (if you remove the ROPs/mem controllers, you lose these too). And we've seen multiple times now TMUs aren't the issue with G80 to G92 and G92 to GT200. Same for SP, where SP op/s increased nearly 50% from G80 to G92 and dropped 50% from G94 to G92 and yet, performance differences weren't anywhere close to 50% different. Now with G92+ to GTX 260, where again SP performance is very similar you see the GTX 260 still always outperforms it and in many cases, significantly.

Bottom-line is that you've repeatedly claimed TMU and SP performance are the most important factors with Nvidia performance when the last 3 generations of unified shader parts prove you wrong, again and again.

You got to be kidding me? The more fillrate you have you need more bandwidth to use all that fillrate efficiently. G92 has whole lot of fillrate. More so than GTX 260 or 216 core. It sure makes HD 4700 faster over 4850 doesn't it and that card has whole lot less fillrate than G92.
And bandwidth only matters if its being saturated, if any part of the GPU is slowing fillrate then bandwidth goes to waste, which is why we see sub 100 FPS instead of a few thousand frames per second.

And G92 has more fillrate than GTX 260? Rofl, might want to check on that, which ultimately shows once again texture fillrate and shader performance aren't as important as you think. Also, 4870 is clocked 20% faster than the 4850....I'm pretty sure an underclocked 4870 at 625 would perform very similarly to a 4850 as that additional bandwidth would be going to waste.

You are making wild assumptions if that.
Pull up any die shot of a Core 2, or any recent Intel CPU for that matter. Much of the die is cache, which shows it would be a huge mistake to judge performance or significance of any logical unit based on size alone.
E8600 Die Shot
So ya, Looks to be around 50% of the core for L2.

And look at GT200 performance. It's not much better than g92. never mind the 2x the ROP with 2x the bandwidth just to get 10-40% performance over g92 with 16ROP and 256bit bus. When they Nvidia could have easily accomplished that by adding GDDR5 circuitry in their G92 chip with 1 gig of vram.

Now look at AMD 3870 to 4850. 4850 has same old 16ROP and 256bit bus but they've raised 50-60% performance over 3870 by raising TMU and SP count.
Hahah! Well, I don't think anyone who has used both would agree with the bolded portion.

So again you're wrong. You claimed Nvidia didn't do anything with ROPs because they didn't think it was important when in reality, that was one of the major areas of improvement over G92. They spent a lot of transistors on the SPs as well, but due to lower clockspeeds, single precision performance wasn't drastically improved. Again, given the parallels I've already drawn, where SP and TMU with GTX 260 and G92+ are nearly identical, but ROP/fillrate is ~40% better, I think its clear ROPs have a greater influence on performance than SP and TMUs.

Because GTX 280 SLI core is clocked higher and has more bandwidth. With AA biggest factor was bandwidth. Without AA that bandwidth limitation shrink a bit and GTX 295 gain some momentum over GTX 260 SLI and catch upto GTX 280 SLI.
If bandwidth is the biggest factor, why isn't the GTX 295 always performing the same as GTX 260 then? They both have the same bandwidth, yet the GTX 295 is slightly faster than the 260 SLI (due to more TMUs and SPs), but still much slower than GTX 280 SLI. I highly doubt 4% clock difference and 27% less bandwidth is the reason for 10-15% differences in resolutions and settings that aren't bandwidth or cpu limited .

There you have it. You said it makes absolute no difference at low resolution with AA but bandwidth sure had an impact in low resolution. :roll: Now downclock your core to 576 mhz and do the benchmark again. divide your percentage by Anandtech's GTX 295 vs GTX 280 SLI review there you should see those extra 4 ROP's impact in performance. Of course this is just a rough idea but you should get the point.
No, I said bandwidth was less of an issue at lower resolutions, and while it had some impact, it clearly doesn't have as much impact as a 14% reduction in ROPs. Again, given GTX 295 and GTX 260 SLI have the same bandwidth, why aren't they performing the same if bandwidth is the main bottleneck? There's no need for me to re-perform the benches at 576MHz, as the GTX 295 is clocked at that speed already, with the same bandwidth as the GTX 260 SLI and outperforms it slightly. GTX 280 SLI performance is much higher than both in non-CPU bottlenecked situations.

1680 4xAA Memory Clocks at 864 (GTX 260/295 effective bandwidth) and 1107 (GTX 280 bandwidth)

3-8% difference from a 27% reduction in bandwidth. Not insignificant, but certainly not the biggest impact on performance either. And you're claiming that and a 4% difference in core clockspeed is the reason for the GTX 280's better performance? I'd say the difference in performance looks closer to the 14% difference in ROPs personally.....

First you say you have less pixels to draw per frame but then you go ahead and say more gpu intensive games or some game with AA the bottleneck isn't fillrate. Does this even make sense?
Yes, it does. Why do you think there's say, 6000 FPS in the loading screen in GTA4 and 20 FPS in Crysis on Enthusiast at the same resolution? The number of pixels drawn per frame is the same, the work load however is very different.

ROFL.. Seriously. You even benchmarked your GTX 280 that is uber bandwidth saturated and it dropped in performance @ 1680 4xAA... Card like GTX 280 or your old 8800gtx need core clocks because it had the bandwidth but not the fillrate to get balance which is confusing the heck of you. :laugh: It just doesn't have the juice to use all that bandwidth efficiently. Now imagine 9800gtx+ with just as much fillrate as GTX 280 and same amount of bandwidth. it won't be pretty but I think it would literally destroy the old GTX 260 and go head to head with GTX 260 216 core.
Rofl, it dropped 3-8% from a 27% reduction in bandwidth. You keep going back to the 9800GTX+, but once again, the GTX 260 is always faster than the 9800GTX+. You can't seem to get that through your head, I'm sure you can drop the bandwidth on the 260 and it'll still be faster, just as I dropped memory clocks on my 8800GTX to match 9800GTX bandwidth and again, found it had minimal impact on performance.

But then it also has 4 less ROP, 16 less TMU, and 32SP. In fact it's exactly 25% cut of the full G92 GPU. So 192bit is bandwidth hungry but 256bit isn't? That makes perfect sense. :roll:
I've covered this with you before, the problem is resolutions are a constant and as such will always require a certain amount of bandwidth to prevent significant bottlenecking. This is why 64-bit and 128-bit nowadays is an automatic red-flag, because as resolution increases, the minimum bandwidth required for higher frame rates is going to also increase.

I'll benchmark my 8800gts to show how retarded you sound. Don't give me that shit about Crysis being bandwidth hungry either
Rofl, I've never claimed Crysis wasn't bandwidth intensive, it clearly is as its constantly streaming large textures from memory. Also, make sure you test core/shader/memory individually, as we all know Crysis is one of the few titles that does actually benefit from shader clocks as much as core clocks.
 

AzN

Banned
Nov 26, 2001
4,112
2
0
Originally posted by: chizow
Nope, raising core clock isn't just increasing ROP performance, it also increases things like staging and set-up however those units are attached to the ROPs (if you remove the ROPs/mem controllers, you lose these too). And we've seen multiple times now TMUs aren't the issue with G80 to G92 and G92 to GT200. Same for SP, where SP op/s increased nearly 50% from G80 to G92 and dropped 50% from G94 to G92 and yet, performance differences weren't anywhere close to 50% different. Now with G92+ to GTX 260, where again SP performance is very similar you see the GTX 260 still always outperforms it and in many cases, significantly.

Bottom-line is that you've repeatedly claimed TMU and SP performance are the most important factors with Nvidia performance when the last 3 generations of unified shader parts prove you wrong, again and again.

The fact that you made a conclusion that ROP made the biggest impact in performance by raising core clocks tell other wise. Now it's staging and setup? :laugh: Gimme a break. Just change your posts however it fits why don't you? :roll:

The only thing you proved is that you can't stop replying even though you've been proven wrong again and again.


And bandwidth only matters if its being saturated, if any part of the GPU is slowing fillrate then bandwidth goes to waste, which is why we see sub 100 FPS instead of a few thousand frames per second.

Yeah bandwidth goes to waste like your 280gtx. :laugh:

And G92 has more fillrate than GTX 260? Rofl, might want to check on that, which ultimately shows once again texture fillrate and shader performance aren't as important as you think. Also, 4870 is clocked 20% faster than the 4850....I'm pretty sure an underclocked 4870 at 625 would perform very similarly to a 4850 as that additional bandwidth would be going to waste.

Here let me do the checking for you since you think it's so god damn funny. I'll even use the 216 core if it makes you feel any better.

GTX 260 216 core
41472 MTexels/sec


9800gtx+

47232 MTexels/sec

So which has more texture fillrate? :roll: Let's take a wild guess. Is it 9800gtx+?

Which ultimately shows G92 is bandwidth starved to stretch all that fillrate. Same reason why 9600gt does so well compared to 9800gt even though it has 40% more texture fillrate. Lot of that core is just being wasted waiting for the memory bandwidth to catch up. Same reason why my Crysis benches didn't take a nose dive when I downclocked my core to whopping 28% and only dropped 7.2% in frames.

4870 has much less texture fillrate compared to G92. Lot of that bandwidth wasted, waiting for more juice just like your GTX 280. You are comparing apples and oranges and trying to use the same logic to get an answer. That's your problem.

Pull up any die shot of a Core 2, or any recent Intel CPU for that matter. Much of the die is cache, which shows it would be a huge mistake to judge performance or significance of any logical unit based on size alone.
E8600 Die Shot
So ya, Looks to be around 50% of the core for L2.

What does that have to do with ROP being the most impact in performance? Nothing. Who cares. :roll:


Hahah! Well, I don't think anyone who has used both would agree with the bolded portion.

GTX 280 is roughly 30% faster than 9800gtx far as raw frame goes. You can turn your AA all you want but if it doesn't pull enough frame rates you won't be able to play. That's a fact.


So again you're wrong. You claimed Nvidia didn't do anything with ROPs because they didn't think it was important when in reality, that was one of the major areas of improvement over G92. They spent a lot of transistors on the SPs as well, but due to lower clockspeeds, single precision performance wasn't drastically improved. Again, given the parallels I've already drawn, where SP and TMU with GTX 260 and G92+ are nearly identical, but ROP/fillrate is ~40% better, I think its clear ROPs have a greater influence on performance than SP and TMUs.

You got the wrong impression. You are the one eating up Nvidia's marketing jargon. GT200 was failure far as performance per wafer. Ultimately eating through Nvidia's profits. Nvidia concentrated too much on ROP and memory bandwidth and not enough on Texture fillrate and SP over g92. In the other hand G92 was a success because it improved on G80 while shrinking the chip while outperforming G80 with less ROP and bandwidth ultimately made them whole lot more money. :music:




If bandwidth is the biggest factor, why isn't the GTX 295 always performing the same as GTX 260 then? They both have the same bandwidth, yet the GTX 295 is slightly faster than the 260 SLI (due to more TMUs and SPs), but still much slower than GTX 280 SLI. I highly doubt 4% clock difference and 27% less bandwidth is the reason for 10-15% differences in resolutions and settings that aren't bandwidth or cpu limited .

Why would it perform same as GTX 260 when it has more texture fillrate and SP? It's only when AA is used that's where the performance drop to GTX 260 SLI levels. Considering you downclocked your GTX 280 memory clocks and showing anywhere from 5% in low resolution. Now try downclocking the core while you are at it to match GTX 260 clocks to see if that 5% becomes 10%. GTX 280 is core hungry no doubt about it. Clocking it lower would detrimental effect on performance.


No, I said bandwidth was less of an issue at lower resolutions, and while it had some impact, it clearly doesn't have as much impact as a 14% reduction in ROPs. Again, given GTX 295 and GTX 260 SLI have the same bandwidth, why aren't they performing the same if bandwidth is the main bottleneck? There's no need for me to re-perform the benches at 576MHz, as the GTX 295 is clocked at that speed already, with the same bandwidth as the GTX 260 SLI and outperforms it slightly. GTX 280 SLI performance is much higher than both in non-CPU bottlenecked situations.

1680 4xAA Memory Clocks at 864 (GTX 260/295 effective bandwidth) and 1107 (GTX 280 bandwidth)

3-8% difference from a 27% reduction in bandwidth. Not insignificant, but certainly not the biggest impact on performance either. And you're claiming that and a 4% difference in core clockspeed is the reason for the GTX 280's better performance? I'd say the difference in performance looks closer to the 14% difference in ROPs personally.....

It sure was an issue when it's dropping performance at such low resolution. Look at those minimum frame rates. Now compare that to GTX295 vs GTX 280 SLI benchmarks. Are you seeing what I'm seeing? There's a pattern between your benchmarks and GTX 295 vs GTX 280 SLI when it comes to minimum frame rates. Now downclock your core clocks to 576mhz and see if you don't see 10% performance drop. Divide the percentage drop against Anandtech benches and that would be your ROP performance drop in the equation.

Yes, it does. Why do you think there's say, 6000 FPS in the loading screen in GTA4 and 20 FPS in Crysis on Enthusiast at the same resolution? The number of pixels drawn per frame is the same, the work load however is very different.

Make it up why don't you. :laugh: In a loading screen there's nothing going on on. It's mostly black and some logo and that's about it.


Rofl, it dropped 3-8% from a 27% reduction in bandwidth. You keep going back to the 9800GTX+, but once again, the GTX 260 is always faster than the 9800GTX+. You can't seem to get that through your head, I'm sure you can drop the bandwidth on the 260 and it'll still be faster, just as I dropped memory clocks on my 8800GTX to match 9800GTX bandwidth and again, found it had minimal impact on performance.

ROFL, you said bandwidth wouldn't matter at such low resolution but it did even on a bandwidth happy card like GTX 280. :laugh:

GTX 260 is barely faster in raw performance with almost 2x as much ROP and 2X as much bandwidth. That's the point.


I've covered this with you before, the problem is resolutions are a constant and as such will always require a certain amount of bandwidth to prevent significant bottlenecking. This is why 64-bit and 128-bit nowadays is an automatic red-flag, because as resolution increases, the minimum bandwidth required for higher frame rates is going to also increase.

Well you didn't cover it enough considering my full G92 GTS is showing same kind of results when I said it would.


Rofl, I've never claimed Crysis wasn't bandwidth intensive, it clearly is as its constantly streaming large textures from memory. Also, make sure you test core/shader/memory individually, as we all know Crysis is one of the few titles that does actually benefit from shader clocks as much as core clocks.

Reading comprehension much? I said "Don't give me that shit about Crysis being bandwidth hungry either"

Crysis is just a hungry game period. :laugh:
 

qbfx

Senior member
Dec 26, 2007
240
0
0
Originally posted by: chizow

Yep, the GTX 295 suffers in bandwidth or VRAM limited resolutions, that was never in question. But again, what you results don't show is the GTX 295 winning 4 titles with 4xAA at 1280/1680 with the maximum level of AA allowed.
I never claimed the GTX295 isn't bandwidth and VRAM insufficient, and that's clearly shown by the results. All I wanted to prove was that the HD4870X2 performs better where it needs to.
I think you'll agree with me that none would buy the card to game at 1280x1024 with low or no AA/AF settings unless they're insane.
The results I listed were the resolutions and AA/AF modes we were discussing and I've never had the intension to limit them to the ones that show the HD4870X2 in a good light. The link is there for everyone to consult the totality of the ratings.

Yep, it shows you counted a split resolution as a win for ATI in one instance, and a tie for Nvidia in two instances.
I never counted splits in favor of neither of the cards. I looked over all of the benchmarks and syntetisized the results myself. Again, go and look the numbers, the higher number wins, that's all.

Yes, the 4870X2 shines with 8xAA, that's no different than the 4870 vs 280, its just not winning nearly as much as Performance Rating would indicate.
The numbers I provided you with in my latest post aren't by no means biased and they indicate quite clearly what the performance ratings show. I don't think these ratings are boosted in any way, please don't make me do the math

Sure you can, but of course then you'd have to include the 4xAA results for 1280 and 1680, which you failed to list above, for some reason. :roll:
No, I'm going to repeat what I mean, when you want to summarize the results of say 12 tests in a single graph there is no better way to do so other than using %.
Again, I didn't fail to list these res/AA/AF modes because we weren't discussing those in our earlier posts. Everyone is free to look at all the results and judge by themselves. However I don't think 1280x1024 or 800x600 is a resolution to be considered when buying a dual-gpu card in the $400-$500 range

Uh, no, they showed guidance telling reviewers their 180 drivers offered a boost in a variety of Top 10 titles, but they were free to review whatever games they wanted with Rel 180. Most sites chose to confirm/deny their claims and found there were significant increases in many of those titles. The GTX 295 preview limited reviewers to 5 titles, but that's really no different than ATI's restrictions on reviewers for their 4870X2 preview.
Uh, no, all the reviewers were limited when testing the Big Bang II drivers, nV would let them review only a couple of games in the first tests. And no, by a couple of games I don't mean the 10 top titles. You call that guidance?

It does win the majority though once you include no AA or lower AA settings. Yes the 4870X2 performs better at higher resolutions and AA settings where bandwidth and VRAM become an issue, but that doesn't stop this site and others from coming to the conclusion the GTX 295 is the faster overall part.
I don't think an enthusiast willing to spend $500 on a card would consider the majority of tests, especially when that majority is comprised of 1280x1024 and 1680x1050 with no AA/AF. And I don't think they'd care about a conclusion based on that majority (again, including absurdly low resolutions a 9600GT could easily handle). I'm fairly confident if the benches included the 1920x1200 tests, they would be in favor of the HD4870X2, as I listed some numbers, that prove that, from another site, earlier.
The GTX295 is a card released half an year after its competitor and costs some $100 more so I don't think its bandwidth and memory limitations can be an excuse for it's limited performance at higher resolutions and AA/AF modes.
And I'm glad we concluded this discussion agreeing with each other.
 

chizow

Diamond Member
Jun 26, 2001
9,537
2
0
Originally posted by: Azn
The fact that you made a conclusion that ROP made the biggest impact in performance by raising core clocks tell other wise. Now it's staging and setup? :laugh: Gimme a break. Just change your posts however it fits why don't you? :roll:

The only thing you proved is that you can't stop replying even though you've been proven wrong again and again.
Yep, of course "ROPs" would mean all attached core units if those units are lost if "ROPs" are reduced. We can easily see this in G80 GTS vs G80 GTX and GT200 260 and GT 280. Even if you clock them identically with core, memory, and shader you'll still end up with a significant difference in performance, and that's due to "ROPs", which include all functional units lost when ROPs are reduced. Its really simple, if I say I'm going to cut off your hands, I shouldn't have to say I'm going to cut off 10 fingers, 10 finger nails as well, its implied.

As for being proven wrong, lol, you just change your argument each time evidence is presented proving you wrong and blatantly contradict yourself over and over again.

Yeah bandwidth goes to waste like your 280gtx. :laugh:
Yep, it does go to waste most of the time, which is exactly what I've claimed all along and its also why a 27% reduction results in 3-8% difference even with AA. Simply put, the GTX 295/260/280 differences aren't solely due to bandwidth, as you've claimed.

Here let me do the checking for you since you think it's so god damn funny. I'll even use the 216 core if it makes you feel any better.

GTX 260 216 core
41472 MTexels/sec


9800gtx+

47232 MTexels/sec

So which has more texture fillrate? :roll: Let's take a wild guess. Is it 9800gtx+?
You said fillrate, not texture fillrate, and the GT200 clearly dominates in that regard. I've already run the numbers for you so I'm well aware of what they are, and as I've already demonstrated, despite higher texture fillrate and similar SP performance, the 9800GTX+ still never beats the GTX 260.

Which ultimately shows G92 is bandwidth starved to stretch all that fillrate. Same reason why 9600gt does so well compared to 9800gt even though it has 40% more texture fillrate. Lot of that core is just being wasted waiting for the memory bandwidth to catch up. Same reason why my Crysis benches didn't take a nose dive when I downclocked my core to whopping 28% and only dropped 7.2% in frames.
Uh, no it shows there's no additional benefit from the additional texture fillrate as the bottleneck lies elsewhere. And yes G94 is another example I've provided to you as proof, where ROP, bandwidth and VRAM are the same as G92, with half the SP and TMU performance, yet performance is very similar. It really doesn't get easier than this really. If both parts were bandwidth limited, as you claimed, there would be no benefit from overclocking the core and shader, yet there clearly is as G92 parts have shown from the G92 GT to G92 GTS to G92 GTX to G92 GTX+ where memory frequency was typically limited to 1000-1100 clocks to due DDR3 limitations of the time. Yet there is a significant gain in performance with each from core/shader clock improvements.

4870 has much less texture fillrate compared to G92. Lot of that bandwidth wasted, waiting for more juice just like your GTX 280. You are comparing apples and oranges and trying to use the same logic to get an answer. That's your problem.
Huh? Why are you comparing the 4870 to Nvidia architectures? And if you are, why doesn't the 9800GTX+ beat the 4870? According to your texture fillrate and wasted bandwidth theories, the 9800GTX+ should run circles around the 4870, yet it doesn't. I'm sure you're going to bring up 4870 and its fewer ROPs compared to GT200 next, so I'll give you a hint: the 4870 had its ROPs reworked so that it can blend/write 2x as many pixels per clock as GT200 and 4x as many as G92.

What does that have to do with ROP being the most impact in performance? Nothing. Who cares. :roll:
Rofl, again, incompetence. You attempted to draw a parallel from core unit size to performance with GT200's die shot and ROP size. I clearly showed its a mistake to make any assumptions about proportional performance based on % of die size alone.

GTX 280 is roughly 30% faster than 9800gtx far as raw frame goes. You can turn your AA all you want but if it doesn't pull enough frame rates you won't be able to play. That's a fact.
GT200 is always faster with or without AA though, and pulling enough frame rates isn't going to be an issue with AA as FPS are always going to be less than or equal to what they were without AA. Saying GT200 isn't much better than G92 is ignorant, and that's a fact.

You got the wrong impression. You are the one eating up Nvidia's marketing jargon. GT200 was failure far as performance per wafer. Ultimately eating through Nvidia's profits. Nvidia concentrated too much on ROP and memory bandwidth and not enough on Texture fillrate and SP over g92. In the other hand G92 was a success because it improved on G80 while shrinking the chip while outperforming G80 with less ROP and bandwidth ultimately made them whole lot more money. :music:
I'm not eating up any of Nvidia's marketing jargon, I'm looking at theoreticals, die changes, and actual performance in coming to the conclusion that the biggest difference between G92 and GT200, ROPs, more closely match the actual real-world performance difference between G92 and GT200. Certainly more than the minimal difference in performance from TMU and SP increases, as you've claimed.

Why would it perform same as GTX 260 when it has more texture fillrate and SP? It's only when AA is used that's where the performance drop to GTX 260 SLI levels. Considering you downclocked your GTX 280 memory clocks and showing anywhere from 5% in low resolution. Now try downclocking the core while you are at it to match GTX 260 clocks to see if that 5% becomes 10%. GTX 280 is core hungry no doubt about it. Clocking it lower would detrimental effect on performance.
That's exactly the point, there's still a difference in performance with GTX 260 SLI and GTX 295 even though you've claimed bandwidth is the bottleneck with AA, yet differences in core architecture still yield results proportionate to the differences in the core architecture.

I know exactly what happens when I downclock the core on the GTX 280, performance scales almost linearly, which is the same as G80. It was significantly less than linear with G92 because changes in core yield smaller gains due to fewer ROPs. And yes I'm well aware core clock also controls TMU and all other parts, but again, G92 already received a 50% boost to TMU performance over G80 so any more benefit from clockspeed is insignificant.

It sure was an issue when it's dropping performance at such low resolution. Look at those minimum frame rates. Now compare that to GTX295 vs GTX 280 SLI benchmarks. Are you seeing what I'm seeing? There's a pattern between your benchmarks and GTX 295 vs GTX 280 SLI when it comes to minimum frame rates. Now downclock your core clocks to 576mhz and see if you don't see 10% performance drop. Divide the percentage drop against Anandtech benches and that would be your ROP performance drop in the equation.
I'm well aware of what my benchmarks are showing, its showing minimum frame rates drop in bandwidth intensive frame sequences, but this occasional drop results in only a 3-8% difference in average despite a 27% reduction in bandwidth, which again proves my earlier points that bandwidth is largely wasted and has less impact on performance than core/clock differences.

Downclocking my GTX 280 wouldn't matter because again, a 4% core clock difference on a GTX 280 is not going to result in a 10% difference in performance, as performance is at best, linear. There's about 20 GTX 285 Reviews you can look over to confirm this. Not to mention you want to compare a single GPU to multi-GPU, which wouldn't be accurate unless the multi-GPU showed 100% scaling, and that's assuming CPU bottlenecking isn't an issue.

Make it up why don't you. :laugh: In a loading screen there's nothing going on on. It's mostly black and some logo and that's about it.
Yeah, that's exactly the point, there's nothing going on and as a result the work load to render each frame is different, which is why we don't see frame rates based on theoretical maximums for fillrate and bandwidth.

ROFL, you said bandwidth wouldn't matter at such low resolution but it did even on a bandwidth happy card like GTX 280. :laugh:

GTX 260 is barely faster in raw performance with almost 2x as much ROP and 2X as much bandwidth. That's the point.
No, you claimed any results with AA were insignificant because of bandwidth limititations. I said bandwidth was less of an issue and showed the performance impact from bandwidth differences were much less than the difference in actual bandwidth. This is clearly reflected in GTX 295/260SLI/280SLI comparisons.

And if you're going back to GTX 260 vs. 9800GTX+, then yes, GTX 260 is always faster and in more GPU intensive resoutions and settings, its performance advantage closely matches its 40% increase to ROP performance and 40% increase to bandwidth.


Well you didn't cover it enough considering my full G92 GTS is showing same kind of results when I said it would.
What? It scales in accordance to differences in architecture? Why don't you look up some reviews of those G92s that were shipping with 128-bit memory buses? They have identical core specs and perform just fine at lower resolutions, but once you increase resolution they fall off drastically. Again, bandwidth only matters when you clearly don't have enough of it.

Reading comprehension much? I said "Don't give me that shit about Crysis being bandwidth hungry either"

Crysis is just a hungry game period. :laugh:
And I've never claimed otherwise, I just find it funny you keep clinging to it as your shining example as its really one of the few titles that show significant gains/decreases from shader/memory clocks.
 

chizow

Diamond Member
Jun 26, 2001
9,537
2
0
Originally posted by: qbfx
Old news

PCGH GTX 285 Review with 295 and 4870X2, updated drivers

Like I said earlier, single reviews aren't much to go by, looks like PCGH tests come to the conclusion the GTX 295 is the faster part. Also looks to be a pretty well balanced testing suite, mostly new and popular titles with a few older or technically intensive games:

Game...GTX 295...4870X2
COD5...4...0
Crysis...2...3
FO3...4...0
FC2...2...2
HL2...0...4
L4D...4...0
Grid...2.5...1.5
STALKER...1...3

That's 4 games for GTX 295, 3 for 4870X2 and a tie for FC2, which the GTX 295 won with 8xAA, by the way. Surely a close race, but that's now 2 reviews for the GTX 295 and one for the 4870X2, with the new drivers of course.
 

AzN

Banned
Nov 26, 2001
4,112
2
0
Originally posted by: chizow

Yep, of course "ROPs" would mean all attached core units if those units are lost if "ROPs" are reduced. We can easily see this in G80 GTS vs G80 GTX and GT200 260 and GT 280. Even if you clock them identically with core, memory, and shader you'll still end up with a significant difference in performance, and that's due to "ROPs", which include all functional units lost when ROPs are reduced. Its really simple, if I say I'm going to cut off your hand, I shouldn't have to say I'm going to cut off 10 fingers, 10 finger nails as well, its implied.

Quite a story about finger nails trying to make relevance to GPU. :laugh:


As for being proven wrong, lol, you just change your argument each time evidence is presented proving you wrong and blatantly contradict yourself over and over again.

Change my argument? How hypocritical of you. I haven't changed my argument since I joined the board but you seem to be in the constant flux. First you agreed with everybody that it was SP then ROP now ROP and bandwidth. :disgust: Same with this post. First ROP then staging and setup. :laugh:


Yep, it does go to waste most of the time, which is exactly what I've claimed all along and its also why a 27% reduction results in 3-8% difference even with AA. Simply put, the GTX 295/260/280 differences aren't solely due to bandwidth, as you've claimed.

So you agree that bandwidth is wasted on your GTX 280 but G92 isn't bandwidth starved even though it has much as texture fillrate as GTX 280? :laugh:


You said fillrate, not texture fillrate, and the GT200 clearly dominates in that regard. I've already run the numbers for you so I'm well aware of what they are, and as I've already demonstrated, despite higher texture fillrate and similar SP performance, the 9800GTX+ still never beats the GTX 260.

Now you have nothing to nit pick yet you knew exactly what I meant but hey if it made you feel any better. All power to you.


Uh, no it shows there's no additional benefit from the additional texture fillrate as the bottleneck lies elsewhere. And yes G94 is another example I've provided to you as proof, where ROP, bandwidth and VRAM are the same as G92, with half the SP and TMU performance, yet performance is very similar. It really doesn't get easier than this really. If both parts were bandwidth limited, as you claimed, there would be no benefit from overclocking the core and shader, yet there clearly is as G92 parts have shown from the G92 GT to G92 GTS to G92 GTX to G92 GTX+ where memory frequency was typically limited to 1000-1100 clocks to due DDR3 limitations of the time. Yet there is a significant gain in performance with each from core/shader clock improvements.

That's quite amusing considering 8800gt is 20-35% faster without AA compared to 9600gt and only within 10% with AA. Which ultimately lies in bandwidth limitations of the g92 core when AA is compared while G94 is more of a balanced product. I've already proven time again and again that G92 is bandwidth starved. You can argue whatever you like but it doesn't change any facts about the tests shown.


Huh? Why are you comparing the 4870 to Nvidia architectures? And if you are, why doesn't the 9800GTX+ beat the 4870? According to your texture fillrate and wasted bandwidth theories, the 9800GTX+ should run circles around the 4870, yet it doesn't. I'm sure you're going to bring up 4870 and its fewer ROPs compared to GT200 next, so I'll give you a hint: the 4870 had its ROPs reworked so that it can blend/write 2x as many pixels per clock as GT200 and 4x as many as G92.

If I were to compare RV770 800sp vs G92 128SP which I shouldn't be comparing because they work differently. Now if I were to talk about balance of fillrate and bandwidth they work just as same whether it be AMD or Nvidia.

You don't need to be hinting anything about AMD's architecture. I've already bunch of reviews to know that ROP was reworked. You only mentioned because 4870 does so well compared to GTX 260 with only 16ROP.


Rofl, again, incompetence. You attempted to draw a parallel from core unit size to performance with GT200's die shot and ROP size. I clearly showed its a mistake to make any assumptions about proportional performance based on % of die size alone.

LOL. but then you tried to compare CPU to GPU and say it's the same. :laugh:


GT200 is always faster with or without AA though, and pulling enough frame rates isn't going to be an issue with AA as FPS are always going to be less than or equal to what they were without AA. Saying GT200 isn't much better than G92 is ignorant, and that's a fact.

Crysis is sure a big issue. Still doesn't change the fact it's only 30% faster without AA. Cut your bandwidth to half and see how well your card performs. :brokenheart: It won't be pretty. I bet it will barely edge out 9800gtx+. :laugh:


I'm not eating up any of Nvidia's marketing jargon, I'm looking at theoreticals, die changes, and actual performance in coming to the conclusion that the biggest difference between G92 and GT200, ROPs, more closely match the actual real-world performance difference between G92 and GT200. Certainly more than the minimal difference in performance from TMU and SP increases, as you've claimed.

You sure are eating up Nvidia's marketing like how it's meant to be brainwashed. :laugh:
Look at Nvidia's financial results from last year and look at it now. Cut in half.


That's exactly the point, there's still a difference in performance with GTX 260 SLI and GTX 295 even though you've claimed bandwidth is the bottleneck with AA, yet differences in core architecture still yield results proportionate to the differences in the core architecture.

LOL... What point? with AA it perform similarly to GTX 260 SLI without AA it performs more closer to GTX 280 SLI. :lips:

I know exactly what happens when I downclock the core on the GTX 280, performance scales almost linearly, which is the same as G80. It was significantly less than linear with G92 because changes in core yield smaller gains due to fewer ROPs. And yes I'm well aware core clock also controls TMU and all other parts, but again, G92 already received a 50% boost to TMU performance over G80 so any more benefit from clockspeed is insignificant.

So add in that 3-8% from memory and your 4% down in clocks. Combination of both core and memory would have more of an effect than just clocking it alone. There you will see your 4 extra ROP performance difference. 1 or 2% if that. I'm being a generous guy here.

G92 is bandwidth starved man. How many times do I have to repeat this? It's a different beast compared to GTX 280 that is core hungry.




I'm well aware of what my benchmarks are showing, its showing minimum frame rates drop in bandwidth intensive frame sequences, but this occasional drop results in only a 3-8% difference in average despite a 27% reduction in bandwidth, which again proves my earlier points that bandwidth is largely wasted and has less impact on performance than core/clock differences.

No it doesn't prove a thing except that your GTX 280 is core starved like I've mentioned hundred times already and G92 is bandwidth starved. You are comparing your 8800gtx and GTX 280 to a G92 which isn't even in the same ball park.


Downclocking my GTX 280 wouldn't matter because again, a 4% core clock difference on a GTX 280 is not going to result in a 10% difference in performance, as performance is at best, linear. There's about 20 GTX 285 Reviews you can look over to confirm this. Not to mention you want to compare a single GPU to multi-GPU, which wouldn't be accurate unless the multi-GPU showed 100% scaling, and that's assuming CPU bottlenecking isn't an issue.

Downclock the core to 576 and memory to 872 and see how much of a performance drop you get. I'm guessing it will be very similar to GTX 295 vs GTX 280 SLI differences of 10-15%. Compare to anandtech's review there you will see what those extra 4 ROP on your GTX 280 does.



Yeah, that's exactly the point, there's nothing going on and as a result the work load to render each frame is different, which is why we don't see frame rates based on theoretical maximums for fillrate and bandwidth.

There's no textures or anything not even a 3d model. Just a logo with few pixels if that. Of course it's going to run at 1000fps.



No, you claimed any results with AA were insignificant because of bandwidth limititations. I said bandwidth was less of an issue and showed the performance impact from bandwidth differences were much less than the difference in actual bandwidth. This is clearly reflected in GTX 295/260SLI/280SLI comparisons.

LOL. You even tested your GTX 280 at low resolution showing a difference when you said it wouldn't matter. Funny. You say one thing but the tests show other wise.

And if you're going back to GTX 260 vs. 9800GTX+, then yes, GTX 260 is always faster and in more GPU intensive resoutions and settings, its performance advantage closely matches its 40% increase to ROP performance and 40% increase to bandwidth.

Why cause GTX 260 has more vram to run a certain resolution with AA? Has more bandwidth for AA? other wise it's not all that much faster. give g92 more bandwidth it would easily crush GTX 260. :light:

What? It scales in accordance to differences in architecture? Why don't you look up some reviews of those G92s that were shipping with 128-bit memory buses? They have identical core specs and perform just fine at lower resolutions, but once you increase resolution they fall off drastically. Again, bandwidth only matters when you clearly don't have enough of it.

You were saying how 192 bit wasn't enough but 256bit is enough even though the chip is exact 25% cut whether it be ROP, TMU, SP, Memory. 1440x900 is high resolution? Since when? :roll:

Mind linking those 128bit bus compared to 256bit bus benchmarks on the exact same chip showing exact same frame rates? Of course not there's no such things. make it up why don't you. :roll:


And I've never claimed otherwise, I just find it funny you keep clinging to it as your shining example as its really one of the few titles that show significant gains/decreases from shader/memory clocks.

ROFL. of course you did. You said crysis is bandwidth starved.

 

chizow

Diamond Member
Jun 26, 2001
9,537
2
0
Originally posted by: Azn
Quite a story about finger nails trying to make relevance to GPU. :laugh:
It obviously got the point across, so I don't have to state the obvious and say "Everything included when you cut ROP and memory clusters" and can just say ROPs instead.

Change my argument? How hypocritical of you. I haven't changed my argument since I joined the board but you seem to be in the constant flux. First you agreed with everybody that it was SP then ROP now ROP and bandwidth. :disgust: Same with this post. First ROP then staging and setup. :laugh:
When did I agree with anyone that SP was ever an issue for Nvidia parts? I was one of the first to argue SP performance on G80 was largely overrated, and that was confirmed with some tests done by Keys and others who disabled shader clusters. Of course there were some shader intensive games more adversely affected, like Crysis, but none of that was ever disputed.

As for changing your argument...lol, when comparing 512-bit bus to 256-bit and 2900XT and 3870 in a recent thread....just funny since some months ago you were claiming that additional bandwidth on those parts was significant, when it clearly was not.

So you agree that bandwidth is wasted on your GTX 280 but G92 isn't bandwidth starved even though it has much as texture fillrate as GTX 280? :laugh:
Yep I agree GTX 280's bandwidth is largely wasted as it shows the GTX 280 needs somewhere between what it has and what the G92 has in terms of bandwidth, but will still always outperform the G92 because bandwidth simply isn't the biggest factor when it comes to performance.

Now you have nothing to nit pick yet you knew exactly what I meant but hey if it made you feel any better. All power to you.
Its not nitpicking when you keep claiming 9800GTX+'s dominates in fillrate based on TMU theoreticals, but ignore the fact its slower than the GTX 260, which typically outperforms the 9800GTX+ proportionately to its 40% advantage in pixel fillrate and ROP performance.

That's quite amusing considering 8800gt is 20-35% faster without AA compared to 9600gt and only within 10% with AA. Which ultimately lies in bandwidth limitations of the g92 core when AA is compared while G94 is more of a balanced product. I've already proven time again and again that G92 is bandwidth starved. You can argue whatever you like but it doesn't change any facts about the tests shown.
No the facts show G94 performance compared to G92 more closely mirrors it ROP and bandwidth theoreticals than its TMU and SP theoreticals. It has almost identical ROP and bandwidth performance but half the TMU and SP performance. Simple question, does G94 perform closer to 100% of the G92 or 50% of the G92?

If I were to compare RV770 800sp vs G92 128SP which I shouldn't be comparing because they work differently. Now if I were to talk about balance of fillrate and bandwidth they work just as same whether it be AMD or Nvidia.

You don't need to be hinting anything about AMD's architecture. I've already bunch of reviews to know that ROP was reworked. You only mentioned because 4870 does so well compared to GTX 260 with only 16ROP.
Why are you dodging the question? You brought up 4870, it has fewer TMUs than 9800GTX+, yet it also always beats the 9800GTX+. Based on your assessment the 9800GTX+ should be running circles around it.

And obviously I have to mention the differences as you've again tried to diminish GT200's ROP performance with a comment about 4870 when I've already covered why AMD is able to keep up or exceed Nvidia with regards to ROP and AA performance. RV770 is able to blend/write 2x as many pixels per clock, but GT200 has 2x as many ROPs to compensate.

LOL. but then you tried to compare CPU to GPU and say it's the same. :laugh:
I never said they were the same, however, it does clearly show drawing parallels between core size and performance is clearly a flawed assessment. You'd have to be incompetent to not understand this.

Crysis is sure a big issue. Still doesn't change the fact it's only 30% faster without AA. Cut your bandwidth to half and see how well your card performs. :brokenheart: It won't be pretty. I bet it will barely edge out 9800gtx+. :laugh:
LOL Crysis again. I'm sure I could cut bandwidth again by another 25% for another 5-10% decrease in performance, but I'm sure it'll still be faster than a 9800GTX+.

You sure are eating up Nvidia's marketing like how it's meant to be brainwashed. :laugh:
Look at Nvidia's financial results from last year and look at it now. Cut in half.
And what does any of that have to do with me proving you wrong, over and over again, showing your claims based on theoreticals simply do not bear out in real world testing with actual hardware?

My analysis and comparisons have nothing to do with marketing or being brainwashed, my decisions are driven by what I can expect for the next generation of GPU so that I can make a sound decision on my next GPU upgrade.

LOL... What point? with AA it perform similarly to GTX 260 SLI without AA it performs more closer to GTX 280 SLI. :lips:
No, the differences are much greater with GTX 280 SLI with or without AA, which shows bandwidth isn't the issue, core differences are. Given the only core difference between GTX 295 and GTX 280 is ROPs, the natural conclusion is that ROPs have a greater impact on performance than TMUs and SPs as 295 performs more like GTX 260 SLI than GTX 280 SLI.

So add in that 3-8% from memory and your 4% down in clocks. Combination of both core and memory would have more of an effect than just clocking it alone. There you will see your 4 extra ROP performance difference. 1 or 2% if that. I'm being a generous guy here.
Funny the results I linked to earlier show much more than 7-10% difference between GTX 295 and GTX 280 SLI, more like 15-25%, which is again much more than the 5-12% difference gained from 260 SLI and GTX 295.

Game - 295...280SLI...260SLI

COD5 - 119...133...115
AOC - 52...62...42
CRY- 41...46...36
FC2- 77...88...75
Grid- 109...138...108
L4D- 124...121...118

G92 is bandwidth starved man. How many times do I have to repeat this? It's a different beast compared to GTX 280 that is core hungry.
Prove it. Increase bandwidth, not decrease, increase bandwidth and see how much performance you gain. If you're saying its bandwidth starved increasing bandwidth should drastically improve performance. Again, don't show me a 25% decrease in bandwidth and 10% drop, I want to see a 25% increase and at least a 10% increase in performance. Again, this goes back to static resolutions and settings requiring certain bandwidth requirements, and also why much of GTX 280's bandwidth goes to waste at lower resolutions or without AA.

No it doesn't prove a thing except that your GTX 280 is core starved like I've mentioned hundred times already and G92 is bandwidth starved. You are comparing your 8800gtx and GTX 280 to a G92 which isn't even in the same ball park.
Actually it proves exactly what I sought to prove, that the bandwidth difference you claimed were the cause of the huge differences between GTX 280SLI and GTX 295 isn't even close to the resulting difference in performance, even in non-bandwidth limited situations.

Downclock the core to 576 and memory to 872 and see how much of a performance drop you get. I'm guessing it will be very similar to GTX 295 vs GTX 280 SLI differences of 10-15%. Compare to anandtech's review there you will see what those extra 4 ROP on your GTX 280 does.
Again, look up, the % differences between GTX 295 and GTX 280 are much greater than the differences derived from a 3-8% difference from bandwidth and 4% difference in core clock.

There's no textures or anything not even a 3d model. Just a logo with few pixels if that. Of course it's going to run at 1000fps.
Which again proves my original point about a single frame at a certain resolution having the same number of pixels to draw, and that GPU work load and not pixel fillrate ultimately determine performance and FPS, not theoreticals. That's why fillrate theoreticals, pixel or texture, ultimately don't mean anything if there's other bottlenecks in the pipeline.

LOL. You even tested your GTX 280 at low resolution showing a difference when you said it wouldn't matter. Funny. You say one thing but the tests show other wise.
Where did I say it wouldn't matter? I didn't, I said it was less of an issue at lower resolutions, and it is, and certainly less of an impact than core differences.

Why cause GTX 260 has more vram to run a certain resolution with AA? Has more bandwidth for AA? other wise it's not all that much faster. give g92 more bandwidth it would easily crush GTX 260. :light:
So now its VRAM right? Again, even at lower resolutions without AA, the GTX 260 is always faster than the 9800GTX+, so you can't keep crying about bandwidth, and now VRAM. When you compare slower lower clocked G92s, like 9800GX2 or GTX, or GTS, the 260 outperforms them even more. How is that possible if those G92s are all bandwidth starved? Yet they all improve with increases to core and shader clock.

As for G92 crushing the GTX 260, prove it. Show me anything close to a linear increase in performance from increases to memory clock alone. You won't get it. Why? Because memory clock increases by themselves don't matter without increasing the core/shader proportionately, as unused bandwidth is wasted bandwidth. Improvements to G92b ran their course and core clock increases could no longer be improved to increase performance, hence GT200.

You were saying how 192 bit wasn't enough but 256bit is enough even though the chip is exact 25% cut whether it be ROP, TMU, SP, Memory. 1440x900 is high resolution? Since when? :roll:
Yep based on bandwidth requirements for static resolutions, increases in bandwidth are always going to be needed, but not anywhere close to increases in GPU performance. So a 192-bit part with identical specs to a 256-bit version might perform identically up to a certain resolution, at which point it tanks.

Mind linking those 128bit bus compared to 256bit bus benchmarks on the exact same chip? Of course not there's no such things. make it up why don't you. :roll:
128-bit 8800GT
Surprised you missed this one given your penchant for buying low-end/neutered GPUs. :laugh:

ROFL. of course you did. You said crysis is bandwidth starved.
Yep, I said it was one of the few titles that would benefit from core/shader/memory clock increases.
 

apoppin

Lifer
Mar 9, 2000
34,890
1
0
alienbabeltech.com
i am getting a little frustrated and am beginning to lose interest ..

i need someone to *summarize* their position and disagreements
- they appear to be undergoing some modification and i cannot keep up with the myriad of issues being ... discussed





 

BFG10K

Lifer
Aug 14, 2000
22,709
2,995
126
Originally posted by: chizow

Oh noes, you don't care what he claimed, problem is what you're quoting is a direct reply to his claim.
So what? If you think he posted something wrong that doesn?t change the fact that you posted something wrong, and I?m calling you out on it. Again stop changing the subject and address the issue of your comments.

He claimed this review was somehow objective, and not just "12 games on Wolfgang's hard drive", saying they were less biased than hand-picked titles in reviews that favor Nvidia, especially when Nvidia's "hand-picked" criteria is clearly more objective to begin with.
It is less biased than nVidia?s list. They tested a larger range of titles with a larger range of ages and did not have any restrictions with titles. Again those 2007 titles are still very demanding and are actively played, so they?re very relevant.

GPU performance evaluation shouldn?t solely focus on ?hot? games at the expense of everything else. Many slightly older titles still have a large following and those that play said titles are very interested in the scores.

My point is there's always going to be subjectivity when selecting a testing suite, but I'd always prefer a criteria like "Any Top 10 title from the last 3 months" over "The Good, the Bad, and the Ugly from the last 3 years, aka "Whatever is on Wolfgang's Hard Drive Today."
Who gives a shit what you prefer? Not everyone has the same preference, so stop stating your preference as fact.

Certainly, what would make you think IHVs aren't supporting them?
I don?t think that at all, I?m pointing out the fact they?re supported makes them legitimate benchmarks.

So what are you saying? That AT's game selections are subjective?
Nope, I?m saying if you think the scores are wrong from their previous reviews then you must also consider the possibility that they?re wrong here.

I?m also saying they didn?t test the same cross-section of games ComputerBase did.

GRID showed some improvement but considering only 1 resolution was tested and the drivers claim results may only be evident with certain hardware configs or settings, that's not much to go by at all.
Another total back-pedal on your part. You stated his Big Bang results were clearly an outlier and I pointed that out, especially given Nvidia did not list improvements in the titles AT tested..

You need to retract this lie. AT tested games on nVidia?s list; two of them in fact. Retract your lie Chizow and stop your rhetoric.

After retesting with the drivers Derek posted an update saying they did see larger improvements at higher resolutions or with AA, saying the drivers "handled high memory usage situations better".
Right, so the benchmark scores turned out to be wrong. Now what?s to say the same won?t happen with the figures you linked?

Its possible the 4870X2 could be using older results
Exactly. Or everything could be correct except for the scores, as was the case with Big Bang.

Heheh ya, what'd you ultimately claim was the reason again?
Nothing you?d understand given you don?t even know how the vendors stack these days, you just constantly mouth off about them.

My point is that the PR 160% or 120% aggregates clearly do not reflect actual performance even in their individual settings where things are at worst, even in high bandwidth situations.
Again stop deflecting the issue with averages. If you can?t understand what an average is then ignore it. It?s a fact that when a larger cross-section of games is used the 4870X2 can come out in a more positive light.

Rofl no, I haven't read your 4850 vs GTX 260+ comparison, I'm sure its fine and all, and will ultimately come down to ugly textures in Thief 2 and Red Faction.
It?s nothing to do with the review, it?s everything to do with my comments afterward about drivers while having used both relevant parts, comments you were still dismissing to the very end.

But of course that has nothing to do whatsoever with your idiotic claims made over a year ago where you claimed ATI drivers were better than Nvidia's based on your experience when you didn't have any relevant experience with an ATI part in over 3 years.
No, I claimed based on my past experiences with ATi drivers combined with online information I wouldn?t have any trouble believing ATi were superior. Then Derek came along and said that multiple reviewers backed my claims.

You however claimed my past experiences were invalid and so was online feedback, and then you refused to accept that Derek backed my claims. Then even after I had used a 4850 for several months of heavy gaming, you still refused to accept the fact that my new findings were mimicking that of my past statements.

Meanwhile you hadn?t touched an ATi part since the 9700 Pro, yet you felt more qualified to comment than me. :roll:

As for ATI driver problems, are you claiming the FC2 issues didn't exist and still don't exist even to this day, despite numerous hot fixes specifically addressing it? Are you claiming the CF/Vista problems didn't exist and still don't exist, despite numerous hot fixes specifically addressing it?
Absolutely not. I?m claiming you can?t make sweeping generalizations based on what is a clearly an outlier like you were. Many that have actually used monthly drivers prefer them to nVidia?s ?scatter? method. I?ve had fixes in legacy games in as little as one month after I reported them.

So it took 2-3 months to fix Far Cry 2. Well considering it?s generally much longer between officially supported drivers from nVidia - and official drivers still don?t guarantee a fix (see Unreal 2 stuttering, Nov 06 - Jun 08) - ATi still have an advantage. And again ATi?s best case scenario is that we have an official fix the next month and I?ve seen this happen multiple times.

Once again, referencing multiple sources, particularly those with concurrent experience with hardware from both camps is certainly compelling evidence
I have experience with hardware from both camps. You don?t, yet you somehow feel you?re more qualified to discuss the issue than I am. :roll:
 

MegaWorks

Diamond Member
Jan 26, 2004
3,819
1
0
Originally posted by: apoppin
i am getting a little frustrated and am beginning to lose interest ..

i need someone to *summarize* their position and disagreements
- they appear to be undergoing some modification and i cannot keep up with the myriad of issues being ... discussed


I'm still trying to get an direct answer to my question from Mr. Member of Nvidia Focus Group wannabe, so let me ask him again. chizow, Are you saying that nVidia drivers are more robust than ATI? Yes or No?
 

GoodRevrnd

Diamond Member
Dec 27, 2001
6,801
581
126
Good timing. Right when Dawn of War 2 and Empire: Total War come out and I'll be looking to upgrade.
 

AzN

Banned
Nov 26, 2001
4,112
2
0
Originally posted by: chizow
It obviously got the point across, so I don't have to state the obvious and say "Everything included when you cut ROP and memory clusters" and can just say ROPs instead.

To who? Maybe yourself that's about it. :laugh:

When did I agree with anyone that SP was ever an issue for Nvidia parts? I was one of the first to argue SP performance on G80 was largely overrated, and that was confirmed with some tests done by Keys and others who disabled shader clusters. Of course there were some shader intensive games more adversely affected, like Crysis, but none of that was ever disputed.

Yeah right. You were agreeing with everyone else in that thread about a year and some months ago when we had the long thread before it got locked. It was me who made Keys pissed off to disable or downclock his SP in the 9600gt thread. You can ask Keys yourself. Things have changed though. We have quite bit of games that are dictated by shader.

As for changing your argument...lol, when comparing 512-bit bus to 256-bit and 2900XT and 3870 in a recent thread....just funny since some months ago you were claiming that additional bandwidth on those parts was significant, when it clearly was not.

Umh.. That's not even remotely close to what I said. That discussion was part of your on going pixel performance matters most arguments you've been trying to spread on the internet. To prove that pixel performance doesn't matter as much I compared 2900xt and 3870 because more bandwidth a card has more pixel fillrate it has even though it's clocked the same yet 2900xt doesn't perform any better. That was the argument. Obviously you didn't get it but then again you don't get much of anything.


Yep I agree GTX 280's bandwidth is largely wasted as it shows the GTX 280 needs somewhere between what it has and what the G92 has in terms of bandwidth, but will still always outperform the G92 because bandwidth simply isn't the biggest factor when it comes to performance.

So you agree that GTX 280 bandwidth is wasted but G92 card that has nearly same amount of texture fillrate as GTX 280 and you are saying bandwidth isn't the biggest factor for G92? ROFL!!! Go ahead downclock your memory to 550mhz and see that performance shrink to 9800gtx+ levels. :laugh:



Its not nitpicking when you keep claiming 9800GTX+'s dominates in fillrate based on TMU theoreticals, but ignore the fact its slower than the GTX 260, which typically outperforms the 9800GTX+ proportionately to its 40% advantage in pixel fillrate and ROP performance.

But then how many times do I have to repeat myself to you that G92 is bandwidth starved. If it had the bandwidth it could easily outclass GTX 260. You can easily figure this out by downclocking your memory to 550mhz and compare your results with a 9800gtx+.


No the facts show G94 performance compared to G92 more closely mirrors it ROP and bandwidth theoreticals than its TMU and SP theoreticals. It has almost identical ROP and bandwidth performance but half the TMU and SP performance. Simple question, does G94 perform closer to 100% of the G92 or 50% of the G92?

Only when AA is applied. Again bandwidth limitations as shown even in your own benchmarks when you downclocked your memory. Once you remove most of the bandwidth limitations by not using AA it's becomes obvious. Then again you can't figure simple thing as bandwidth limitations and restrictions when it comes to AA.


Why are you dodging the question?

I'm really dodging anything you muster. :laugh:

You brought up 4870, it has fewer TMUs than 9800GTX+, yet it also always beats the 9800GTX+. Based on your assessment the 9800GTX+ should be running circles around it.

Because it has bandwidth which 9800gtx+ does not. It can also do 1200 GFLOPS and 9800gtx+ can not. Isn't it obvious where the raw performance difference comes from? If you even paid attention to what I was saying you wouldn't even be asking me stupid questions like this. But you thought it was a bright idea to ask me anyway.

And obviously I have to mention the differences as you've again tried to diminish GT200's ROP performance with a comment about 4870 when I've already covered why AMD is able to keep up or exceed Nvidia with regards to ROP and AA performance. RV770 is able to blend/write 2x as many pixels per clock, but GT200 has 2x as many ROPs to compensate.

How would 4870 supersede GTX 280 far as render back performance go when 4870 can only do 16pix/clk with no MSAA while GTX 260 can do 28pix/clk. Only time a 4870 should be 12% better than GTX 260 is with AA and absolutely no advantage over GTX 280.


I never said they were the same, however, it does clearly show drawing parallels between core size and performance is clearly a flawed assessment. You'd have to be incompetent to not understand this.

You were trying to compare CPU L2 cache performance in games to GPU core size. In a server environment L2 cache does whole lot more than CPU mhz which crumble your feeble arguments.


LOL Crysis again. I'm sure I could cut bandwidth again by another 25% for another 5-10% decrease in performance, but I'm sure it'll still be faster than a 9800GTX+.

Why the "F" not? Crysis is the most GPU intensive game there is to PC gaming. Cut you bandwidth to 50% to match 9800gtx+ which would be 550mhz. Considering GTX 280 is only 30% faster than 9800gtx+ in raw performance you should get a pretty good idea 9800gtx+ is bandwidth starved. Test AA performance and then comeback. Perhaps you don't want to downclock your memory and benchmark because it would literally make your arguments obsolete.


And what does any of that have to do with me proving you wrong, over and over again, showing your claims based on theoreticals simply do not bear out in real world testing with actual hardware?

You proved me wrong? Like when you said bandwidth wouldn't matter at lower resolutions with AA? :laugh: How about when I said full g92 would show same results like my 8800gs? :brokenheart:

My analysis and comparisons have nothing to do with marketing or being brainwashed, my decisions are driven by what I can expect for the next generation of GPU so that I can make a sound decision on my next GPU upgrade.

Maybe you need to go back to your analysis and do some more testing because you've proven wrong multiple times in this thread alone.

Making a sound upgrade? Like a $500 GTX 280 which you over paid that perform worse than 9800gx2 or a 4870x2? :laugh:


No, the differences are much greater with GTX 280 SLI with or without AA, which shows bandwidth isn't the issue, core differences are. Given the only core difference between GTX 295 and GTX 280 is ROPs, the natural conclusion is that ROPs have a greater impact on performance than TMUs and SPs as 295 performs more like GTX 260 SLI than GTX 280 SLI.

Without AA 10-12% better if that. That's really awesome considering it's also clocked 4% higher, 4% higher shader clocks and has 27% more bandwidth. :roll:

Did you know that you've been repeating that same line about 3x in every single one of your post without even proving anything? "ROP the greater impact in performance" How sad you have to tell yourself just to make yourself feel better.




Funny the results I linked to earlier show much more than 7-10% difference between GTX 295 and GTX 280 SLI, more like 15-25%, which is again much more than the 5-12% difference gained from 260 SLI and GTX 295.

Game - 295...280SLI...260SLI

COD5 - 119...133...115
AOC - 52...62...42
CRY- 41...46...36
FC2- 77...88...75
Grid- 109...138...108
L4D- 124...121...118

It's useless when you just keep on telling me the same crap over and over again without no proof. I said bandwidth would be factor in low resolutions with AA which it was. That's been proven at this point. You need to prove yourself wrong again by downclocking your GTX 280 to 576/1242/872 then come back with the results. :thumbsup:


Prove it. Increase bandwidth, not decrease, increase bandwidth and see how much performance you gain. If you're saying its bandwidth starved increasing bandwidth should drastically improve performance. Again, don't show me a 25% decrease in bandwidth and 10% drop, I want to see a 25% increase and at least a 10% increase in performance. Again, this goes back to static resolutions and settings requiring certain bandwidth requirements, and also why much of GTX 280's bandwidth goes to waste at lower resolutions or without AA.

I already proved with Crysis benches on my 8800gts but hey you want to be ignorant that's on you.



Actually it proves exactly what I sought to prove, that the bandwidth difference you claimed were the cause of the huge differences between GTX 280SLI and GTX 295 isn't even close to the resulting difference in performance, even in non-bandwidth limited situations.

ROFL.. In your own benchmarks bandwidth made the difference. Look at those minimum frames. Same pattern as GTX 295 vs GTX 280 SLI. I also said you need to downclock your core to GTX 260 levels which you haven't even done to prove yourself wrong again. I'm guessing you are too scared to prove yourself wrong at this point.


Again, look up, the % differences between GTX 295 and GTX 280 are much greater than the differences derived from a 3-8% difference from bandwidth and 4% difference in core clock.

Just do it and do come back. Make sure your shader is also clocked to GTX 295 levels which is 1242. If I had GTX 280 I would prove this to you but then again I'm just not gullible enough to pay $500 for a video card that perform 30% faster in raw frame rates than a $100 card.


Which again proves my original point about a single frame at a certain resolution having the same number of pixels to draw, and that GPU work load and not pixel fillrate ultimately determine performance and FPS, not theoreticals. That's why fillrate theoreticals, pixel or texture, ultimately don't mean anything if there's other bottlenecks in the pipeline.

What point? Because it has nothing to draw in the first place.

You've been going back and forth and changing it up to CPU now loading screens in a game that has no relevance. Then you go on to say "this proves my point". You've proven nothing. :laugh:


Where did I say it wouldn't matter? I didn't, I said it was less of an issue at lower resolutions, and it is, and certainly less of an impact than core differences.

It sure was an issue if it dropped 3-8% average depending on the game and made minimum frame rates drop by 21%.



So now its VRAM right?

Vram so you don't texture thrash. :roll: In a non vram or non bandwidth limited situation it's not much faster.


Again, even at lower resolutions without AA, the GTX 260 is always faster than the 9800GTX+, so you can't keep crying about bandwidth, and now VRAM.

Crying? :laugh: Is that you've been doing in front of your computer when you benchmarked your GTX 280 to find out I was right all along?

Then again my 8800gts showed much bigger performance drop without AA in Crysis when I lowered the bandwidth over Core clocks which proves that g92 is bandwidth starved.

When you compare slower lower clocked G92s, like 9800GX2 or GTX, or GTS, the 260 outperforms them even more. How is that possible if those G92s are all bandwidth starved? Yet they all improve with increases to core and shader clock.

260 outperforming a 9800gx2? :laugh: You are quite funny. GX2 outperforms GTX 280.

All G92 cards are bandwidth starved.



As for G92 crushing the GTX 260, prove it. Show me anything close to a linear increase in performance from increases to memory clock alone. You won't get it. Why? Because memory clock increases by themselves don't matter without increasing the core/shader proportionately, as unused bandwidth is wasted bandwidth. Improvements to G92b ran their course and core clock increases could no longer be improved to increase performance, hence GT200.

Considering GTX 260 only perform 10% better than 9900gtx in raw numbers and my 8800gts dropped 15.8% when I reduced my bandwidth by 28%. When you consider GTX 260 has 60% more bandwidth than 9800gtx+ this pretty much proves my theories.



Yep based on bandwidth requirements for static resolutions, increases in bandwidth are always going to be needed, but not anywhere close to increases in GPU performance. So a 192-bit part with identical specs to a 256-bit version might perform identically up to a certain resolution, at which point it tanks.

Exactly it's not static and no it's not going to perform identically. When you have a card that's already bandwidth limited it's going to be more bandwidth limited when you raise everything by 25%. More bandwidth limitations than the card with 75% power of the chip. In this case my 8800gs. That's why my frames dropped more with my 8800gts than it did with 8800gs when I lowered my memory clocks.



And where is the benchmarks to show that it performs identically compared to 256 bit bus at lower resolutions? :roll:

That card has 96SP. It's not even a 8800gt. It's more like 8800gs on a 128bit bus.

Surprised you missed this one given your penchant for buying low-end/neutered GPUs. :laugh:

I'm sorry I'm just not cool as you to be dropping $500 on a card just so you can have larger e-penis when GTX 280 is only 30% faster in raw performance over my 8800gts.



Yep, I said it was one of the few titles that would benefit from core/shader/memory clock increases.

You denied it then you acknowledge it only when you've been caught red handed.
 

legcramp

Golden Member
May 31, 2005
1,671
113
116
Guys, it depends on how you like to play it. GTX295 is single-card, 4870x2 is dual card, that is the decision maker. Obviously, if dual cards scare you, get the GTX295, but if they don't, find a really cheap 4870x2.

You fail
 

chizow

Diamond Member
Jun 26, 2001
9,537
2
0
Originally posted by: BFG10K
So what? If you think he posted something wrong that doesn?t change the fact that you posted something wrong, and I?m calling you out on it. Again stop changing the subject and address the issue of your comments.
How is what I stated wrong? I said a criteria like "Top 10 games from the last 2-3 months" is always going to be more objective and always going to be more relevant than what he claimed were objective benchmarks, especially if he's going to claim it was a list of popular titles. Surely I don't need to link definitions for criteria, objective, popular and relevant do I?

Are you saying the inclusion of random older titles like CoJ, RS:Vegas, and Jericho that never even satisfied any such popularity, relevance or best-selling criteria are more relevant than recent titles like COD5, L4D, FO3 or FC2? So again, how am I changing the subject when I directly replied to his claims about Nvidia's guidance being subjective while claiming the Computerbase selections were somehow more objective or popular?

It is less biased than nVidia?s list. They tested a larger range of titles with a larger range of ages and did not have any restrictions with titles. Again those 2007 titles are still very demanding and are actively played, so they?re very relevant.
How is a list of "Top 10 titles in the last 3 months" less biased than "Whatever is on Wolfgang's Hard Drive"? What a joke, this has nothing to do with the titles still being very demanding, this has to do with those titles having never been popular or relevant titles to begin with. Unless your criteria is perhaps "Titles that hit the $9.99 Bargain Bin within 3 months of release" or perhaps "Poorly reviewed titles with Metacritic scores less than 60", I'm really not sure how you can argue Nvidia's guidance is somehow "cherry-picked" or "marketing".

GPU performance evaluation shouldn?t solely focus on ?hot? games at the expense of everything else. Many slightly older titles still have a large following and those that play said titles are very interested in the scores.
Yet that criteria is always going to be more relevant to more people and ultimately more objective as the selection criteria is not determined directly by the reviewer.

Who gives a shit what you prefer? Not everyone has the same preference, so stop stating your preference as fact.
You should if you're going to claim "12 random titles on Wolfgang's Hard Drive" is somehow more objective or relevant than Top 10 titles from the last 3-4 months.

I don?t think that at all, I?m pointing out the fact they?re supported makes them legitimate benchmarks.
Sure, if you claim the criteria is going to be "supported titles from the last 3 years." Of course that wouldn't make that criteria more relevant or those titles more popular than Top 10 selling games from the last 3 months.

Nope, I?m saying if you think the scores are wrong from their previous reviews then you must also consider the possibility that they?re wrong here.

I?m also saying they didn?t test the same cross-section of games ComputerBase did.
I've already considered the possibility and cross-referenced older results and found they're not a carbon copy as they were in the past with months-old archived results.

And of course AT didn't test the same cross-section of games as ComputerBase did, as all reviews have some degree of subjectivity to them where reviewers select which titles are tested.

Another total back-pedal on your part. You stated his Big Bang results were clearly an outlier and I pointed that out, especially given Nvidia did not list improvements in the titles AT tested..

You need to retract this lie. AT tested games on nVidia?s list; two of them in fact. Retract your lie Chizow and stop your rhetoric.
And Grid showed improvement, so you need to retract your lie. And Crysis showed improvement once resolution/AA was increased. Also I wasn't directly referring to their "benchmark result", as I've stated numerous times not all games, even if listed would show improvement, I was referring to his conclusion, which was clearly the outlier. He stated the drivers did not make any noticeable impact which was more or less a foregone conclusion as he didn't bother to test enough of the games listed or sufficient resolutions and settings to come to that conclusion.

Right, so the benchmark scores turned out to be wrong. Now what?s to say the same won?t happen with the figures you linked?
No the benchmark wasn't wrong, Derek simply didn't test thoroughly enough to come to the conclusion he came to and later corrected his mistake by testing more titles and more resolutions/settings, which was my point about it being the outlier all along.

Exactly. Or everything could be correct except for the scores, as was the case with Big Bang.
I've already cross-referenced the benches, have you? And why do you keep referencing the Big Bang review? Is that a joke? Are you really going to hang your hat (and yourself) on this "benchmark"?
  • Rel 178 Driver Rel 180 Driver Percent Increase
    Crysis 29.7 29.6 -0.34
    Far Cry 2 38.2 37.1 -2.88
    Race Driver GRID 55.8 56.8 1.79
    Oblivion 43.8 40.5 -7.53
    Enemy Territory: Quake Wars 85.1 84.5 -0.71
5 games, 1 resolution. Are you really going to get behind those benchmarks as evidence to back your point? Just shows how disingenuous you are when it comes to making a point.

Nothing you?d understand given you don?t even know how the vendors stack these days, you just constantly mouth off about them.
Yep, I understand you claimed numerous times that ATI had better and more robust drivers based on your experience, and now you're claiming you bought another Nvidia part because Nvidia's drivers are better and more robust. Makes total sense. :laugh:

Again stop deflecting the issue with averages. If you can?t understand what an average is then ignore it. It?s a fact that when a larger cross-section of games is used the 4870X2 can come out in a more positive light.
No it shows averages can clearly be skewed by subjective selections that favor one vendor or another, which is why averages and aggregates should not be used as a cumulative indication of performance. As for it being a "fact", I'm not so sure of that given they couldn't even replicate their performance gains a day later:
Cat 9.x Comparison
GTX 295 Review with Cat 9.x
Odd, where'd those driver performance increases go?

It?s nothing to do with the review, it?s everything to do with my comments afterward about drivers while having used both relevant parts, comments you were still dismissing to the very end.
Afterward? No you made all those idiotic comments about ATI drivers being superior in your experience long before you touched a 4850, which was what? 4 years after the last ATI part you used? Its hilarious you're attempting to justify comments made years before you finally decided to refresh your frame of reference. After which, you decided on Nvidia, again. LMAO.

No, I claimed based on my past experiences with ATi drivers combined with online information I wouldn?t have any trouble believing ATi were superior. Then Derek came along and said that multiple reviewers backed my claims.
LMAO. Derek didn't even enter the discussion until months later. But nice try, you explicitly claimed your opinion was based on your experience despite the fact you hadn't used an ATI part in years. I continued to say you couldn't make any such claim based on your experience as you hadn't used an ATI part in years, at which point you continued to claim you had extensive experience with both vendors, when you didn't.

You however claimed my past experiences were invalid and so was online feedback, and then you refused to accept that Derek backed my claims. Then even after I had used a 4850 for several months of heavy gaming, you still refused to accept the fact that my new findings were mimicking that of my past statements.
And there you go again, trying to clump experiences you didn't have with online feedback, which aren't your experiences. You still can't seem to make the distinction, but this isn't surprising given your comments about Vista, Nvidia drivers and hot fixes, as a devout XP user.

I never refused to accept Derek's claims, as has already been linked for you. And I haven't said anything about your 4850 experiences other than I'm sure the conclusion was predictable in order to justify previously ignorant comments. I also found it incredibly ironic and not surprisingly hypocritical that you would still choose to purchase an Nvidia part that was by most accounts inferior to the 4870 1GB based on criteria you've set. And now, you're claiming your decision was based on Nvidia having superior driver features?!?!? LMAO. We certainly have come full circle with your hypocrisy.

Meanwhile you hadn?t touched an ATi part since the 9700 Pro, yet you felt more qualified to comment than me. :roll:
Again, the difference is, I didn't make an idiotic claim that the comments were based on my experience. And yes, quoting the likes of Anand, Derek, and now Jarred is certainly compelling testimony, as they're absolutely more qualified to comment than you given they actually have access and relevant experience with the hardware simultaneously at any given time, unlike you.

Absolutely not. I?m claiming you can?t make sweeping generalizations based on what is a clearly an outlier like you were. Many that have actually used monthly drivers prefer them to nVidia?s ?scatter? method. I?ve had fixes in legacy games in as little as one month after I reported them.
I'm not making sweeping generalizations and these results don't seem to be the outlier, they're pervasive. I've already linked to appropriate links that show this. As for your fixes...what's that supposed to mean other than they were already working on a fix? You think your bug reports are like the Catalyst Bat Phone or something? LOL. Its already been demonstrated numerous times and confirmed by ATI's own driver team that it would most likely take 2 months in order to get a fix in due to alternating driver trunks.

So it took 2-3 months to fix Far Cry 2. Well considering it?s generally much longer between officially supported drivers from nVidia - and official drivers still don?t guarantee a fix (see Unreal 2 stuttering, Nov 06 - Jun 08) - ATi still have an advantage. And again ATi?s best case scenario is that we have an official fix the next month and I?ve seen this happen multiple times.
Is FC2 studdering even fixed? Its possible Nvidia takes longer with fixes for legacy titles, but its also clearly obvious Nvidia has better support for new titles, a claim I made very early on. No need to start this again, I know you prefer support for old titles and aren't like the majority of users who buy new video cards for new games, and that's fine. Another reason I was surprised you'd go with Nvidia over ATI, again.

I have experience with hardware from both camps. You don?t, yet you somehow feel you?re more qualified to discuss the issue than I am. :roll:
That's nice, except my references, quotes and links to credible sources are always going to be more relevant than your non-concurrent experiences riddled with 3 year holes.
 

chizow

Diamond Member
Jun 26, 2001
9,537
2
0
Originally posted by: Azn
To who? Maybe yourself that's about it. :laugh:
Oh right, I should've said it was obvious to those who weren't incompetent.

Yeah right. You were agreeing with everyone else in that thread about a year and some months ago when we had the long thread before it got locked. It was me who made Keys pissed off to disable or downclock his SP in the 9600gt thread. You can ask Keys yourself. Things have changed though. We have quite bit of games that are dictated by shader.
Uh, no I didn't. I stated early on SPs were overrated in games that were current at the time, as it was obvious to anyone who actually used the hardware and you even took exception to it in that G94 HardOCP comparison thread where you couldn't make sense of simple apple-to-apple comparisons.

Shader clocks were new so obviously messing with the clocks and ratios was interesting. Anyone who was curious enough to test clearly saw shader clock increases did not have nearly the impact as raising the core clock. With a G80 GTS at the time with a stock 513MHz core clock it was obvous that increasing core clock 20% yielded a much larger increase than increases to shader or memory, same with G92, G94, GT200, etc/ etc.

Umh.. That's not even remotely close to what I said. That discussion was part of your on going pixel performance matters most arguments you've been trying to spread on the internet. To prove that pixel performance doesn't matter as much I compared 2900xt and 3870 because more bandwidth a card has more pixel fillrate it has even though it's clocked the same yet 2900xt doesn't perform any better. That was the argument. Obviously you didn't get it but then again you don't get much of anything.
Uh no, I said bandwidth was wasted if it was unused and pointed to these nearly identical parts with different bandwidth as definitive proof. You kept trying to point to insignificant differences between the parts when they performed identically, despite superior bandwith in favor of the 2900XT. Glad you finally understand the error you made then, perhaps you can take a step back and apply that knowledge here. :laugh:

So you agree that GTX 280 bandwidth is wasted but G92 card that has nearly same amount of texture fillrate as GTX 280 and you are saying bandwidth isn't the biggest factor for G92? ROFL!!! Go ahead downclock your memory to 550mhz and see that performance shrink to 9800gtx+ levels. :laugh:
Yep I've shown bandwidth has less impact on performance by underclocking by 27% and seeing only 3-8% decrease in performance. I can absolutely say for certain performance would drop by more than 3-8% by dropping core clocks. Likewise for a 9800GTX+, I can say gains from Core/Memory are going to be greater than increases to memory bandwidth. And that's something I don't need to prove as its been proven time and time again when comparing 8800GT to 8800GTS to 9800GTX to 9800GTX+ where memory bandwidth was fixed due to GDDR3 speed limitations, yet performance always increased between parts based on clock speed. Pretty impressive gains from a part that's bandwidth limited.

But then how many times do I have to repeat myself to you that G92 is bandwidth starved. If it had the bandwidth it could easily outclass GTX 260. You can easily figure this out by downclocking your memory to 550mhz and compare your results with a 9800gtx+.
Heh, sure it could, except the 9800GTX+ is never faster than a GTX 260, even at lower resolutions that aren't bandwidth intensive. I don't see how you can't get this through your head, despite theoreticals that exceed the GTX 260, the 9800GTX+ is never faster. You could drop resolution to 1280x1024 and as long as you weren't completely CPU limited, the GTX 260 would be faster. Is it still bandwidth limited at 1280?

Only when AA is applied. Again bandwidth limitations as shown even in your own benchmarks when you downclocked your memory. Once you remove most of the bandwidth limitations by not using AA it's becomes obvious. Then again you can't figure simple thing as bandwidth limitations and restrictions when it comes to AA.
BS, bandwidth is the same with G94 and G92, you're wrong here, you never could explain away G94 and you clearly still can't. G94 clearly shows TMU and SP have less impact on performance than ROPs given it has the same bandwidth as G92.

I'm really dodging anything you muster. :laugh:
Of course you are, that's what you do when you're clearly wrong and just trolling.

Because it has bandwidth which 9800gtx+ does not. It can also do 1200 GFLOPS and 9800gtx+ can not. Isn't it obvious where the raw performance difference comes from? If you even paid attention to what I was saying you wouldn't even be asking me stupid questions like this. But you thought it was a bright idea to ask me anyway.
What does bandwidth matter when it has less Texture Fillrate to begin with? According to your claims it shouldn't need the additional bandwidth, yet the 4870 runs circles around the 9800GTX+. Are you saying the 9800GTX+ is completely bandwidth limited to the point a further increase in core clocks (and as a result, higher texture fillrate) would yield no improvement? And why are you bringing up SP performance, are you finally going to acknowledge other factors impact performance before fillrate becomes an issue?

How would 4870 supersede GTX 280 far as render back performance go when 4870 can only do 16pix/clk with no MSAA while GTX 260 can do 28pix/clk. Only time a 4870 should be 12% better than GTX 260 is with AA and absolutely no advantage over GTX 280.
Hilarious that you claimed you knew what you were talking about. Here's a hint, RV770's
ROPs were reworked so that they can write/blend 2 pixels per clock, meaning they have an effective 32 ROPs that write 1 pixel per clock.

You were trying to compare CPU L2 cache performance in games to GPU core size. In a server environment L2 cache does whole lot more than CPU mhz which crumble your feeble arguments.
So were you referring to server performance with GT200's ROP size? You're just grasping at straws here now. You claimed the size of a logical unit should somehow be proportional to its impact on performance, when that's clearly a flawed assessment.

Why the "F" not? Crysis is the most GPU intensive game there is to PC gaming. Cut you bandwidth to 50% to match 9800gtx+ which would be 550mhz. Considering GTX 280 is only 30% faster than 9800gtx+ in raw performance you should get a pretty good idea 9800gtx+ is bandwidth starved. Test AA performance and then comeback. Perhaps you don't want to downclock your memory and benchmark because it would literally make your arguments obsolete.
I already know why, its because Crysis is one of the few titles that does actually scale with all facets of the GPU, core/memory/shader. But of course the exception doesn't make the rule, which is why you're reluctant to test any other games except for Crysis.

You proved me wrong? Like when you said bandwidth wouldn't matter at lower resolutions with AA? :laugh: How about when I said full g92 would show same results like my 8800gs? :brokenheart:
I said bandwidth was less relevant at lower resolutions with AA, when taken into account GT200 doesn't need all of its bandwidth and the performance difference with less bandwidth, I have proven you wrong. Bandwidth isn't the issue you claim it is with GTX 280, 295 and 260, plain and simple. As for G92 and GS, I never made any such claim, I asked which results you were showing and that it was obvious a bandwidth crippled card would show adverse effects if running higher resolutions.

Maybe you need to go back to your analysis and do some more testing because you've proven wrong multiple times in this thread alone.

Making a sound upgrade? Like a $500 GTX 280 which you over paid that perform worse than 9800gx2 or a 4870x2? :laugh:
Its hilarious really how Nvidia just happens to release parts so perfect in proving you wrong. G94 was the perfect part that really blows away all your arguments, the fact you continue to ignore these clear indicators just reinforce my belief you're too incompetent to understand such fundamental and elementary concepts. GTX 295 was another clear indicator, and you still don't get it.

As for my buying habits, don't worry about me, I got my GTX 280 for $314 ($65 after selling my 8800GTX ) on launch day, like many others on these forum. Its been worth every penny and is easily one of the best pieces of hardwarer I've owned. Awesome part without a doubt, anyone who has owned one would concur, and certainly avoid making ignorant comments like "the GTX 280 isn't much better than the 9800GTX+". Maybe if you held off on all the low-end parts you've bought in the same period of time you might've been able to pick one up yourself.

Without AA 10-12% better if that. That's really awesome considering it's also clocked 4% higher, 4% higher shader clocks and has 27% more bandwidth. :roll:

Did you know that you've been repeating that same line about 3x in every single one of your post without even proving anything? "ROP the greater impact in performance" How sad you have to tell yourself just to make yourself feel better.
And that 10% turns into 15-25% with AA. Its OK though really, believe what you like, it'll just be different from what knowledgeable observers conclude, including the reviewers I've already linked to.

It's useless when you just keep on telling me the same crap over and over again without no proof. I said bandwidth would be factor in low resolutions with AA which it was. That's been proven at this point. You need to prove yourself wrong again by downclocking your GTX 280 to 576/1242/872 then come back with the results. :thumbsup:
LMAO, no proof. That clearly shows the difference between GTX 280 SLI is far greater than GTX 295 and GTX 260 SLI between each other. Again, how would downclocking the GTX 280 4% show anything other than <=4% decrease in performance? I've already shown a 27% from bandwidth yields 3-8% difference, combined with a 4% linear decrease max from core/shader, it still can't account for the remaining differences between GTX 295 and GTX 280 SLI. How hard is that to understand?

I already proved with Crysis benches on my 8800gts but hey you want to be ignorant that's on you.
No you didn't, you had already increased the performance of the shader/core so that they did need more bandwidth and then reduced bandwidth further. This is very different than you claiming G92 at a certain performance level was bandwidth starved, at which point you would increase memory clockspeed to demonstrate a tangible gain from the increase. If you're going to claim the 9800GTX+ would beat the GTX 260 if it had the same bandwidth, you don't drop its bandwidth down to that of an 8800GS, heh.

ROFL.. In your own benchmarks bandwidth made the difference. Look at those minimum frames. Same pattern as GTX 295 vs GTX 280 SLI. I also said you need to downclock your core to GTX 260 levels which you haven't even done to prove yourself wrong again. I'm guessing you are too scared to prove yourself wrong at this point.
Again, 27% drop in bandwidth for a 3-8% difference in actual performance. Minimum FPS numbers were never in question, I'm sure they would be similar with the GTX 295 with 111GB/s, except we weren't comparing minimums, we were comparing averages. I said bandwidth at lower resolutions with AA was less significant an issue, and it clearly is. Throw in the 4% difference from core/shader and again you'll see the differences cannot be reconciled by bandwidth and clockspeed alone, which leaves ROPs.

Just do it and do come back. Make sure your shader is also clocked to GTX 295 levels which is 1242. If I had GTX 280 I would prove this to you but then again I'm just not gullible enough to pay $500 for a video card that perform 30% faster in raw frame rates than a $100 card.
Prove what? A 4% drop in core/shader will result in a linear decrease, at best? Anyone who has owned a G80 or higher part would see this is the case with most parts. Until then it looks like you're willing to make ignorant comments about $100 and $500 video cards when you clearly don't know what you're talking about. G92, G80 GTX, whatever other low-end part....been there done that, and I can say for 100% certainty I'm not in any hurry to go back to that level of performance.

What point? Because it has nothing to draw in the first place.

You've been going back and forth and changing it up to CPU now loading screens in a game that has no relevance. Then you go on to say "this proves my point". You've proven nothing. :laugh:
Rofl, nothing to draw? A blank screen still has pixels being drawn. Why don't you run some of your favorite 3DMark benches and note how many FPS you get in theoretical fillrate benches. Are frame rates so high because there's nothing to draw there as well?

It sure was an issue if it dropped 3-8% average depending on the game and made minimum frame rates drop by 21%.
So are you saying a 27% drop in core clock will result in less than 3-8% difference in performance? Like I said, bandwidth is clearly less of an issue, as I've stated numerous times.

Vram so you don't texture thrash. :roll: In a non vram or non bandwidth limited situation it's not much faster.
So why bring it up when it once again, is less of an issue at lower resolutions?

Crying? :laugh: Is that you've been doing in front of your computer when you benchmarked your GTX 280 to find out I was right all along?

Then again my 8800gts showed much bigger performance drop without AA in Crysis when I lowered the bandwidth over Core clocks which proves that g92 is bandwidth starved.
My results were as expected, much less impact in performance relative to the decrease in bandwidth, which also reinforces my claims ROPs are the biggest impact on performance as that's the only remaining difference with regards to GTX 295 and 280s core architecture.

Again, what do your results show other than a G96 is bandwidth limited at whatever resolution you're running? You're claiming a G92 is bandwidth limited at a certain performance level, then decreasing memory bandwidth, when you should be increasing it to show a tangible gain.

260 outperforming a 9800gx2? :laugh: You are quite funny. GX2 outperforms GTX 280.

All G92 cards are bandwidth starved.
Actually that should read 9800GTX+ compared to G92, GTS, 9800GX2 (compared to GTX+ SLI). You're claiming all G92 are bandwidth starved, yet the 9800GTX+ manages to outperform all slower G92 parts significantly despite the fact memory clocks are limited to available DDR3 speeds in the 1000-1100 range.

Considering GTX 260 only perform 10% better than 9900gtx in raw numbers and my 8800gts dropped 15.8% when I reduced my bandwidth by 28%. When you consider GTX 260 has 60% more bandwidth than 9800gtx+ this pretty much proves my theories.
Again, your GTS numbers already show far less than linear performance decrease from memory bandwidth alone, but you still ignore the fact that test is flawed as you've already overclocked the core/memory, increasing bandwidth requirements further. If you're going to claim something is limiting, you need to demonstrate gain from increasing that factor, not decreasing it. Increase memory 10% and see if you get anything close to 10% gain, I doubt you will.

Exactly it's not static and no it's not going to perform identically. When you have a card that's already bandwidth limited it's going to be more bandwidth limited when you raise everything by 25%. More bandwidth limitations than the card with 75% power of the chip. In this case my 8800gs. That's why my frames dropped more with my 8800gts than it did with 8800gs when I lowered my memory clocks.
Huh? No. If an 8800GS performs within 25% of an 8800GTS at 1280 due to core differences and memory differences, it just shows everything is scaling as expected and bandwidth isn't an issue. If an 8800GS performs 50% slower than an 8800GTS at 1920 or with AA, then it becomes apparent bandwidth isn't sufficient and is badly crippling the part.

And where is the benchmarks to show that it performs identically compared to 256 bit bus at lower resolutions? :roll:

That card has 96SP. It's not even a 8800gt. It's more like 8800gs on a 128bit bus.
Who cares, do your own research. The point is it won't even perform like an 8800GT or 8800GS at higher resolutions because the 128-bit bus is so crippling. You had an 8600GT, you should know exactly what I'm talking about.

I'm sorry I'm just not cool as you to be dropping $500 on a card just so you can have larger e-penis when GTX 280 is only 30% faster in raw performance over my 8800gts.
Uh ya, keep tellin yourself that while I'm enjoying all my games at resolutions and settings that my 8800GTX (faster than your 8800GTS) couldn't manage.

You denied it then you acknowledge it only when you've been caught red handed.
Uh, no I didn't, I said very early on that I knew Crysis was one of a few titles that benefitted from core/shader/memory bandwidth increases.
 

chizow

Diamond Member
Jun 26, 2001
9,537
2
0
Originally posted by: MegaWorks
I'm still trying to get an direct answer to my question from Mr. Member of Nvidia Focus Group wannabe, so let me ask him again. chizow, Are you saying that nVidia drivers are more robust than ATI? Yes or No?
What would it matter what I think? You'll just ignore it and maybe over react again. I'm very happy with Nvidia's drivers though, especially support for new and popular titles, which is the main reason I keep my hardware up-to-date in the first place.
 

AzN

Banned
Nov 26, 2001
4,112
2
0
Originally posted by: chizow
Oh right, I should've said it was obvious to those who weren't incompetent.

incompetent like you finger nail explanation and tried to make sense that ROP makes the most performance impact in games? :laugh:


Uh, no I didn't. I stated early on SPs were overrated in games that were current at the time, as it was obvious to anyone who actually used the hardware and you even took exception to it in that G94 HardOCP comparison thread where you couldn't make sense of simple apple-to-apple comparisons.

Uh. Yes you did. :roll: You also said bandwidth didn't matter either. :laugh:


Shader clocks were new so obviously messing with the clocks and ratios was interesting. Anyone who was curious enough to test clearly saw shader clock increases did not have nearly the impact as raising the core clock. With a G80 GTS at the time with a stock 513MHz core clock it was obvous that increasing core clock 20% yielded a much larger increase than increases to shader or memory, same with G92, G94, GT200, etc/ etc.

WRONG!!! G92 has minimal impact when you raise core frequencies. This has already been tested on 2 G92 cards which you have been ignoring with the evidence provided. 8800gs and 8800gts. G94 is a balanced product so you would need to raise everything. Not just the core.


Uh no, I said bandwidth was wasted if it was unused and pointed to these nearly identical parts with different bandwidth as definitive proof. You kept trying to point to insignificant differences between the parts when they performed identically, despite superior bandwith in favor of the 2900XT. Glad you finally understand the error you made then, perhaps you can take a step back and apply that knowledge here. :laugh:

ROFL... I even provided evidence back then which you ignored because it just made you look foolish with your ever retarded ROP makes the biggest performance difference. Here it is again.

http://techreport.com/r.x/rade...850/3dm-color-fill.gif

2900xt 3.8 Gpixels/second
3870 3.1 Gpixels/second

Why does it outperform in High color fill? Oh that's right ROP is tied down to the memory controller. In this case 2900xt had more bandwidth so it has more pixel performance than 3870. That 20% advantage in ROP performance equates to 1% performance difference in the real world. :roll:


Yep I've shown bandwidth has less impact on performance by underclocking by 27% and seeing only 3-8% decrease in performance. I can absolutely say for certain performance would drop by more than 3-8% by dropping core clocks. Likewise for a 9800GTX+, I can say gains from Core/Memory are going to be greater than increases to memory bandwidth. And that's something I don't need to prove as its been proven time and time again when comparing 8800GT to 8800GTS to 9800GTX to 9800GTX+ where memory bandwidth was fixed due to GDDR3 speed limitations, yet performance always increased between parts based on clock speed. Pretty impressive gains from a part that's bandwidth limited.

ROFL. You've proven to yourself that bandwidth mattered even at low resolution when you applied AA which you said it wouldn't. You even said it yourself... GTX 280 bandwidth was being wasted. When you downclocked your memory even @ 28% GTX 280 still has enough bandwidth to run efficiently. Now try downclocking it to 9800gtx+ memory bandwidth and see your card perform more like 9800gtx+ than GTX 280. :brokenheart:

Your arguments are hypocritical and illogical at best. If someone actually read your arguments to Nvidia or AMD engineer or in a court they would surely laugh in your face. :laugh:


Heh, sure it could, except the 9800GTX+ is never faster than a GTX 260, even at lower resolutions that aren't bandwidth intensive. I don't see how you can't get this through your head, despite theoreticals that exceed the GTX 260, the 9800GTX+ is never faster. You could drop resolution to 1280x1024 and as long as you weren't completely CPU limited, the GTX 260 would be faster. Is it still bandwidth limited at 1280?

It was bandwidth limited @ 1680x1050. What makes you think it's not bandwidth limited @ 1280x1024? :laugh: Do you want me to test Crysis again but @ 1280x1024 with my 8800gts? Don't try to tell me about bandwidth limitation when you don't even have the slightest idea. You made a wild assumption and you were wrong. Everyone can make a guess and be wrong but what you need to do is learn from those guesses and mistakes which you haven't even begun to do. You are being ignorant at best and it's making you look stupid.

Bandwidth plays a role whether it be low or high resolution. What you haven't figured out is that more fillrate you have more bandwidth you need to run efficiently. Just because you have certain amount of bandwidth doesn't mean you aren't bandwidth limited.

Your GTX 280 or GTX 260 isn't bandwidth starved. You and I both can agree to that but you keep implying g92 isn't bandwidth starved when it has texture fillrate closer to GTX 280 is ludicrous..



BS, bandwidth is the same with G94 and G92, you're wrong here, you never could explain away G94 and you clearly still can't. G94 clearly shows TMU and SP have less impact on performance than ROPs given it has the same bandwidth as G92.

That's all you can come up with (BS) when evidence is provided for you. :disgust:

I pretty much explained this in my previous post in this thread and 9600gt thread with Keys. Here I will explain it again since you seem to think I couldn't explain it.

G92 is bandwidth starved and I've pretty much said this from the very start while G94 is a balanced product. With AA 9600gt comes within 10% of 9800gt because of bandwidth limitations of the 9800gt but when you compare raw performance without AA 9800gt is roughly 20-35% faster than 9600gt when bandwidth restrictions were less critical. Now if both cards has all the bandwidth it needs 9800gt would be much faster with or without AA.


Of course you are, that's what you do when you're clearly wrong and just trolling.

ROFL. you are the one here spreading Nvidia marketing jargon. I've proven almost everything you claimed with my benchmarks and your benchmarks. :laugh: If that makes me a troll to you I guess I am.


What does bandwidth matter when it has less Texture Fillrate to begin with?

Because you need bandwidth to take full advantage of that texture fillrate not to mention AA needs more bandwidth.


According to your claims it shouldn't need the additional bandwidth, yet the 4870 runs circles around the 9800GTX+.

I wouldn't say circles now. It's roughly 20% faster.


Are you saying the 9800GTX+ is completely bandwidth limited to the point a further increase in core clocks (and as a result, higher texture fillrate) would yield no improvement? And why are you bringing up SP performance, are you finally going to acknowledge other factors impact performance before fillrate becomes an issue?

Exactly my point. 9800gtx+ is bandwidth limited to a point increasing core had very little impact on performance as shown on my Crysis benchmark. Did I say shader didn't impact performance? Of course I didn't I said texture and shader makes bigger impact than ROP. :roll:


Hilarious that you claimed you knew what you were talking about. Here's a hint, RV770's
ROPs were reworked so that they can write/blend 2 pixels per clock, meaning they have an effective 32 ROPs that write 1 pixel per clock.

ROFL. What hint? That you are making shit up? RV770 can only do 16 pixels per clock at 32bit color and GTX 280 can do 32 pixels per clock. Only time it can write/blend 2 pixels per clock is with MSAA or 64bit color consisting of HDR scenes. :laugh:

http://techreport.com/r.x/rade...ender-backend-ppcs.gif


So were you referring to server performance with GT200's ROP size? You're just grasping at straws here now. You claimed the size of a logical unit should somehow be proportional to its impact on performance, when that's clearly a flawed assessment.

Grasp straws is right. You were trying to compare L2Cache to GPU logical units which I corrected your flawed argument. :gift: Depending on what you were doing those units can have bigger performance impact. That's the point. But in a game ROP doesn't pull nearly as much frames as texture or SP would long as you weren't limited anyway.


I already know why, its because Crysis is one of the few titles that does actually scale with all facets of the GPU, core/memory/shader. But of course the exception doesn't make the rule, which is why you're reluctant to test any other games except for Crysis.

Why don't you test your GTX 280 while you are at it. Try to see if you get the same results as I did. Of course not because GTX 280 is core hungry while my GTS is bandwidth hungry. Now if your GTX 280 showed same results as I did than I would be wrong and you would be right but then again you are too chicken shit to test your GTX 280 because you already know what the outcome will be. I would test more games but majority of games out there don't have a built in game benchmark. I don't have WIC and demo doesn't work with the latest drivers. GTA4 shouldn't be a factor because that game is CPU dependent.



I said bandwidth was less relevant at lower resolutions with AA, when taken into account GT200 doesn't need all of its bandwidth and the performance difference with less bandwidth, I have proven you wrong. Bandwidth isn't the issue you claim it is with GTX 280, 295 and 260, plain and simple.

Look at your minimum frame rate in Crysis and WIC at 21% and 14%. I don't know how much more relevance you need. :roll: Now clock your GTX 280 memory clocks to 550mhz and do the benches again and compare it with 9800gtx+. I dare you to post your results.


As for G92 and GS, I never made any such claim, I asked which results you were showing and that it was obvious a bandwidth crippled card would show adverse effects if running higher resolutions.

ROFL. you never made any claims? WTF??? This post alone you were taking jabs about GTX 260 vs 9800gtx+ being not bandwidth starved for performance and you made no claims? W@W!! I'm speechless. Pathetic. High resolutions? I tested 1680x1050 with my 8800gts and 1440x900 with my 8800gs. :roll:


Its hilarious really how Nvidia just happens to release parts so perfect in proving you wrong. G94 was the perfect part that really blows away all your arguments, the fact you continue to ignore these clear indicators just reinforce my belief you're too incompetent to understand such fundamental and elementary concepts. GTX 295 was another clear indicator, and you still don't get it.

That's because you don't have the slightest clue when it comes to GPU. I think you are no better than any of the new forum members asking which card is faster when it comes to GPU except those newbies are willing to learn but you. :laugh: Nazi's used to believe something too. That didn't make them right.



As for my buying habits, don't worry about me, I got my GTX 280 for $314 ($65 after selling my 8800GTX ) on launch day, like many others on these forum. Its been worth every penny and is easily one of the best pieces of hardwarer I've owned. Awesome part without a doubt, anyone who has owned one would concur, and certainly avoid making ignorant comments like "the GTX 280 isn't much better than the 9800GTX+". Maybe if you held off on all the low-end parts you've bought in the same period of time you might've been able to pick one up yourself.

Employee discount I bet for spreading Nvidia marketing jargon. :laugh: There's no way I'm paying more than $200 for GTX 280. The card performs fine for older games but new games like Crysis it still chokes not to mention ridiculous power requirements for such a slow ass card.


And that 10% turns into 15-25% with AA.

I bet you downclocked your core and SP clocks by now only to find out I was right all along. That 10% turned into 15-25% with AA. ROFL... No wonder you don't want to show results. So silly.


Its OK though really, believe what you like, it'll just be different from what knowledgeable observers conclude, including the reviewers I've already linked to.

Knowledgeable observers like who? You? :laugh: You've also linked bit-tech benchmarks with GTX 260 SLI outperforming GTX 295 too when it's not theoretically possible. :laugh:


LMAO, no proof. That clearly shows the difference between GTX 280 SLI is far greater than GTX 295 and GTX 260 SLI between each other.

Exactly no proof. When you lowered your bandwidth your minimum frames were all over the place dropping as much as 21%.


Again, how would downclocking the GTX 280 4% show anything other than <=4% decrease in performance? I've already shown a 27% from bandwidth yields 3-8% difference, combined with a 4% linear decrease max from core/shader, it still can't account for the remaining differences between GTX 295 and GTX 280 SLI. How hard is that to understand?

Only if it worked that way. Poor Chizow. :frown: When you have combination for both core, SP, bandwidth you get much greater drops than just clocking each clocks separately. Of course you would have to be knowledgeable with GPU to actually know this but then again you know anything about GPU.



No you didn't, you had already increased the performance of the shader/core so that they did need more bandwidth and then reduced bandwidth further. This is very different than you claiming G92 at a certain performance level was bandwidth starved, at which point you would increase memory clockspeed to demonstrate a tangible gain from the increase. If you're going to claim the 9800GTX+ would beat the GTX 260 if it had the same bandwidth, you don't drop its bandwidth down to that of an 8800GS, heh.

Of course I did but it's okay. I can easily clock my 8800gts to stock clocks and only raise memory clocks and core clocks separately to show you same kind of results but then again it would make your arguments look silly. You wouldn't want that. You will just say it doesn't prove anything when the evidence is before you.



Again, 27% drop in bandwidth for a 3-8% difference in actual performance. Minimum FPS numbers were never in question, I'm sure they would be similar with the GTX 295 with 111GB/s, except we weren't comparing minimums, we were comparing averages. I said bandwidth at lower resolutions with AA was less significant an issue, and it clearly is. Throw in the 4% difference from core/shader and again you'll see the differences cannot be reconciled by bandwidth and clockspeed alone, which leaves ROPs.

Benchmarks talk chizow barks.



Prove what? A 4% drop in core/shader will result in a linear decrease, at best? Anyone who has owned a G80 or higher part would see this is the case with most parts. Until then it looks like you're willing to make ignorant comments about $100 and $500 video cards when you clearly don't know what you're talking about. G92, G80 GTX, whatever other low-end part....been there done that, and I can say for 100% certainty I'm not in any hurry to go back to that level of performance.

Ruff ruff. errr.... :laugh: Useless without benches.



Rofl, nothing to draw? A blank screen still has pixels being drawn. Why don't you run some of your favorite 3DMark benches and note how many FPS you get in theoretical fillrate benches. Are frame rates so high because there's nothing to draw there as well?

Make some sense. Why would I bench 3dmark fillrate test to measure fps? In 3dmark fillrate test the screen isn't black with a logo. It's testing peak texures with multiple layers of textures.



So are you saying a 27% drop in core clock will result in less than 3-8% difference in performance? Like I said, bandwidth is clearly less of an issue, as I've stated numerous times.

What does downclocking the core of your GTX 280 have anything to do with what was stated? You said memory clocks wouldn't matter but it showed huge drops in minimum frame rates and in average frame rates by 3-8%. As for core I've already mentioned this numerous times in this thread. Your GTX 280 is core hungry and my 8800gts is bandwidth hungry. :roll:



So why bring it up when it once again, is less of an issue at lower resolutions?

In lower resolutions you need less vram. I thought you were brighter than this but apparently you are not. I only brought it up because GTX 260 having more VRAM for higher resolutions and or with AA. When those vram abnormally isn't effected I implied it's not much faster.



My results were as expected, much less impact in performance relative to the decrease in bandwidth, which also reinforces my claims ROPs are the biggest impact on performance as that's the only remaining difference with regards to GTX 295 and 280s core architecture.

Again, what do your results show other than a G96 is bandwidth limited at whatever resolution you're running? You're claiming a G92 is bandwidth limited at a certain performance level, then decreasing memory bandwidth, when you should be increasing it to show a tangible gain.

So you are implying all cards are built like your GTX 280 so all cards behave the same to core frquency like your GTX 280. How retarded of you.



Actually that should read 9800GTX+ compared to G92, GTS, 9800GX2 (compared to GTX+ SLI). You're claiming all G92 are bandwidth starved, yet the 9800GTX+ manages to outperform all slower G92 parts significantly despite the fact memory clocks are limited to available DDR3 speeds in the 1000-1100 range.

Significantly? :laugh: I wouldn't say all that when it only performs 5-10% faster.

This is where the extra shader clocks and core clocks come in. Although it's bandwidth limited it still able to pull more fillrate with high core clocks but not as much as a card that has combination of bandwidth and core efficiency.



Again, your GTS numbers already show far less than linear performance decrease from memory bandwidth alone, but you still ignore the fact that test is flawed as you've already overclocked the core/memory, increasing bandwidth requirements further. If you're going to claim something is limiting, you need to demonstrate gain from increasing that factor, not decreasing it. Increase memory 10% and see if you get anything close to 10% gain, I doubt you will.

If I prove it will you acknowledge that g92 is bandwidth starved? You will once again say something as stupid as Crysis being bandwdith starved and we are back to square 1. It wouldn't matter anyway because I already have the results ready if you are willing to acknowledge g92 is bandwidth starved and you are wrong.


Huh? No. If an 8800GS performs within 25% of an 8800GTS at 1280 due to core differences and memory differences, it just shows everything is scaling as expected and bandwidth isn't an issue. If an 8800GS performs 50% slower than an 8800GTS at 1920 or with AA, then it becomes apparent bandwidth isn't sufficient and is badly crippling the part.

Did you even understand what I had originally posted? You are disagreeing with something you didn't even understand? :roll:

I mentioned that if 8800gs showed 10% dropped by downclocking memory, 8800gts should show bigger drops at the same resolution. In this case my 8800gts dropped 16% while 8800gs 10%. Why you ask? Because 8800gs was bandwidth starved in the first place. If you add 25% more of everything in a bandwidth limited card you would get amplified results.



Who cares, do your own research. The point is it won't even perform like an 8800GT or 8800GS at higher resolutions because the 128-bit bus is so crippling. You had an 8600GT, you should know exactly what I'm talking about.

ROFL. You said you would show me benchmarks of same chip performing same at lower resolution with lower bandwidth in fact there are no benchmarks available on the web which you have lied about. When you couldn't find any you told me go research? Pathetic feeble attempts. :laugh: I can easily do this with my 8800gts to prove you wrong once again. Cut my bandwidth to half and show you how stupid you sound. :roll:


Uh ya, keep tellin yourself that while I'm enjoying all my games at resolutions and settings that my 8800GTX (faster than your 8800GTS) couldn't manage.

You cool you drop lot of money on gaming hardware. I'm 33 years old. I game on the side. Nothing major but I do it as a hobby but that's me. I'm content with what I have because that's all I need.


Uh, no I didn't, I said very early on that I knew Crysis was one of a few titles that benefitted from core/shader/memory bandwidth increases.

ROFL... Now you are denying... Seriously Lame. :thumbsdown:
 

MegaWorks

Diamond Member
Jan 26, 2004
3,819
1
0
Originally posted by: chizow
Originally posted by: MegaWorks
I'm still trying to get an direct answer to my question from Mr. Member of Nvidia Focus Group wannabe, so let me ask him again. chizow, Are you saying that nVidia drivers are more robust than ATI? Yes or No?
What would it matter what I think? You'll just ignore it and maybe over react again. I'm very happy with Nvidia's drivers though, especially support for new and popular titles, which is the main reason I keep my hardware up-to-date in the first place.

No I won't, I've asked you directly how is that ignoring? But by reading some of your posts I guest the answer is yes. But I disagree with you on that! Why? I play a lot of games, and I've used the R300, R420, R520, RV670 and the RV770 and none of them gave me problems with any of my games. OK maybe the RV770 with FC2 but that's it, my situation with CrossFire was bios than software drivers.

The GTX 260 that I'm using is ok! I mean nothing revolutionary it works just like my 4850. I really don't see the difference the way you put it.
 
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |