AnTuTu and Intel

Page 4 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

jfpoole

Member
Jul 11, 2013
43
0
66
Thank you for the additional information on this.

Using uninitialized data would certainly explain some presence of denorms. But Jon Tyler's claim of 100% of the data set consisting of denorms seems very bizarre. If we're talking about single-precision floats the exponent field has to be 0 for a denorm and this doesn't include zero and negative zero. So given a uniform distribution there'd be a less than 1/256 (0.4%) chance of an uninitialized 32-bit floating being a denorm. Uninitialized data doesn't tend to really follow uniform distributions, it may well have a lot of stuff that looks like denorms due to containing 32-bit integers that aren't large enough (> 0x7FFFFF) to trip out of denorms. But they could also have a lot of zeroes, negative numbers, < 32-bit int datatypes, strings, pointers, code, etc etc that would tend to sway away from this. If you just so happen to have really bad luck in what sort of stuff you were doing beforehand it could have ended up with a dataset that's fully denormal, but I still find this pretty bizarre. It'd be good to get some measurements on this.

I do have to ask, is Geekbench doing anything to verify that the results of the benches are correct, like a checksum? IMO this is an important step in benchmarks, you want to at least make sure that the compiled code is doing what it's supposed to and not cutting corners that result in the wrong answers.

I'm not sure why Jon Tyler claimed 100% of the data set contained denormals. Given the nature of the bug between 30% and 40% of the operations were at risk of using denormals. I don't know how many of these operations actually used denormals, though; I'll have to write a test to check the actual data being referenced.

Geekbench 2 did not verify that the results of the workloads were correct (results were verified during development on the platform the developer was working on, but there were no automated checks in place that would run on all platforms).

Geekbench 3, on the other hand, will have extensive build-time tests that verify the workloads are operating as expected. Geekbench 3 will also have lightweight run-time tests as well, but we're putting most of our validation effort on tests that we run internally so that we can keep the Geekbench run-time as low as possible.
 

Concillian

Diamond Member
May 26, 2004
3,751
8
81
Didn't someone say that there were "lies, damn lies, and benchmarks"?

And then there are phone benchmarks where the thermal allowance starts high... then slowly degrades as the chassis heats up.

Phone benchmarks are basically completely useless not that phone SoCs have gotten bigger and bigger gaps between turbo speeds and realistic steady state speeds. Nothing more frustrating than playing a game on your phone and performance starts fine, then 10 minutes in things are turning into a slide-show as speeds start throttling.
 
Last edited:

mrmt

Diamond Member
Aug 18, 2012
3,974
0
76
Have they though? You see they now put a disclaimer everywhere that says that their compilers are optimized for their processors (http://software.intel.com/sites/default/files/m/0/1/3/opt-notice-en_080411.gif)

As of September 2010 at least they didn't drop the biased dispatcher, see the post on 2010-09-22 here:

Yes, that's what I said in the post. The FTC wanted, among other things, Intel to drop the biased dispatcher but could not bend them on the issue even in the settlement.
 

PPB

Golden Member
Jul 5, 2013
1,118
168
106

Exophase

Diamond Member
Apr 19, 2012
4,439
9
81
Yes, that's what I said in the post. The FTC wanted, among other things, Intel to drop the biased dispatcher but could not bend them on the issue even in the settlement.

Well I guess Intel didn't really have to do anything then

I really don't get the lengths they're going through in avoiding this. I don't know if they just don't like changing their approach on anything because it looks like they're admitting they're wrong or what. Their CPUs (Core series but I'm sure Silvermont, even Saltwell) are plenty competitive on their own merits and they don't need to resort to compiler games to be in a good position. This just damages their credibility.
 

Exophase

Diamond Member
Apr 19, 2012
4,439
9
81
I'm not sure why Jon Tyler claimed 100% of the data set contained denormals. Given the nature of the bug between 30% and 40% of the operations were at risk of using denormals. I don't know how many of these operations actually used denormals, though; I'll have to write a test to check the actual data being referenced.

Geekbench 2 did not verify that the results of the workloads were correct (results were verified during development on the platform the developer was working on, but there were no automated checks in place that would run on all platforms).

Geekbench 3, on the other hand, will have extensive build-time tests that verify the workloads are operating as expected. Geekbench 3 will also have lightweight run-time tests as well, but we're putting most of our validation effort on tests that we run internally so that we can keep the Geekbench run-time as low as possible.

Thanks for the update. I'm looking forward to Geekbench 3.
 

jfpoole

Member
Jul 11, 2013
43
0
66
Thanks for the update. I'm looking forward to Geekbench 3.

If you (or anyone else) would like to try out a beta build of Geekbench 3 send me an email (john at primatelabs dot com) and I'll set you up. We're about a week away from having something ready for external testing.
 

Mondozei

Golden Member
Jul 7, 2013
1,043
41
86
Great post, Exophase. A friend who works on an ARM design has been ranting for a while about Antutu and Geekbench and the quality of code currently coming out of JITs on ARM... I really hate the cross-ISA situation in terms of benchmarking. The worst part is that generally-credible reviewers don't caveat their articles enough, so people actually give credit to these results. It's worse than the 80s IPC comparisons across RISC/CISC because the macroscopic workload characteristics aren't even the same.

That was very interesting.

Now please explain to me like I was a 5 year old.
 

mrmt

Diamond Member
Aug 18, 2012
3,974
0
76
I really don't get the lengths they're going through in avoiding this. I don't know if they just don't like changing their approach on anything because it looks like they're admitting they're wrong or what. Their CPUs (Core series but I'm sure Silvermont, even Saltwell) are plenty competitive on their own merits and they don't need to resort to compiler games to be in a good position. This just damages their credibility.

From what I get they don't want the compiler as reference for anything, they want ICC to be the compiler on the market, but one that is only useful to Intel customers. And for most customers, this is a non issue as AMD market share is too small to bother. My company, for example, has over 400.000 PCs and servers scattered around multiple sites, not a single one is AMD.
 

StrangerGuy

Diamond Member
May 9, 2004
8,443
124
106

They used to cheat in Quake 3 in the P4 days too, Q3 ran optimized codepath on the P4 but non-optimized for Athlon XP. With a modified dll the Athlon XP actually beat the P4 without any loss of image quality. Xbit had an article about it but I'm not sure whether it is still there. Not like it actually matters in practical terms since the framerates for either side are already in the hundreds.
 

nonworkingrich

Junior Member
Jul 4, 2013
7
0
0
Hopefully this will lead to other companies thinking twice before ruining their credibility for a fistful of dollars.
 

galego

Golden Member
Apr 10, 2013
1,091
0
0

Thanks by the link. An excerpt:

But a report by EE Times identified that the testing process has some discrepancies. Technical consulting firm BDTI pointed out that Intel's Atom processor wasn’t executing all the instructions to be run during the RAM test. This artificially improved that results in favor of Intel Atom.

After receiving the criticism, AnTuTu revised its benchmarking tool for Android. The revised tests of the Atom Z2580 processor dropped overall scores by 20 percent for the processor.


This brings the ARM-based Samsung Exonys 5 Octa processor on top of Intel Atom Z2580 in terms of performance. Seems, Intel has a lot more to do to touch ARM-based processors in performance.

:whiste:
 

blackened23

Diamond Member
Jul 26, 2011
8,548
2
0
I can't wait for the product to be released to put these arguments to rest. Hopefully Intel will give good reason for the naysayers to initiate radio silence (as they have normally done in recent years)
 
Last edited:

parvadomus

Senior member
Dec 11, 2012
685
14
81
Because ARM chips win Geekbench despite evidence that Geekbench is intentionally crippled on Intel processors. From a recent interview with Silvermont's lead architect,

Now we can safely say that Antutu is intentionally crippled on Intel processors too. :whiste:
 

StrangerGuy

Diamond Member
May 9, 2004
8,443
124
106

krumme

Diamond Member
Oct 9, 2009
5,956
1,595
136
Let's spin it this way: Intel didn't actually care about the mobile market, right? So who cares if they can't win one lousy benchmark.

In other news: PC shipments drop another 11% in Q2 2013. Looks somebody is quaking in their 22nm boots already...

http://www.dailytech.com/PC+Shipmen...enovo+Beats+HP+as+Top+Vendor/article31948.htm

The pc market is in decline. The highend phone market looking to slow down in increase or even stagnate.

What is the perfect response?

Another fab and antutu on tab

(Edit: well "tap" not "tab" lol, beer on tab gives the wrong interpretation)
 
Last edited:

CTho9305

Elite Member
Jul 26, 2000
9,214
1
81
CTho9305 said:
Great post, Exophase. A friend who works on an ARM design has been ranting for a while about Antutu and Geekbench and the quality of code currently coming out of JITs on ARM... I really hate the cross-ISA situation in terms of benchmarking. The worst part is that generally-credible reviewers don't caveat their articles enough, so people actually give credit to these results. It's worse than the 80s IPC comparisons across RISC/CISC because the macroscopic workload characteristics aren't even the same.
That was very interesting.

Now please explain to me like I was a 5 year old.

Someone is trying to improve the performance of a future CPU. CPUs are very complicated, and you have to pick and choose which things you make fast (if you try making everything fast, your CPU will be too expensive and use batteries too quickly). To pick the right parts to make fast, you need to look at many programs and see what they want most from the CPU. Benchmarks are special programs that act like many different programs, so instead of looking at 100 different programs, people just look at a few benchmarks.

It turns out that the benchmarks people use now are terrible. They do a very bad job of measuring what they're supposed to measure; they don't act like the programs they're supposed to act like. It's even worse because they don't measure the same thing on ARM CPUs as they do on x86 CPUs. One of the CPU makers may even be deliberately cheating in how they handle the benchmark programs. That means when somebody uses them to compare an x86 CPU like Jaguar to an ARM CPU like A15, the winner of the benchmark might not run real programs faster. If somebody who cares about running real programs well uses e.g. Anandtech's reviews to help pick a phone, the bad benchmarks might trick them into buying a slower phone.
 

StrangerGuy

Diamond Member
May 9, 2004
8,443
124
106
The pc market is in decline. The highend phone market looking to slow down in increase or even stagnate.

What is the perfect response?

Another fab and antutu on tab

(Edit: well "tap" not "tab" lol, beer on tab gives the wrong interpretation)

Taiwanese makers can make cheaper and better mobos? We can always go back to sell x86 chips.

Consumer SSDs not profitable? Nah, We can always go back to sell x86 chips.

Atheros & Co can make cheaper, good enough Wi-Fi adapters? We can always go back to sell x86 chips.

Nobody except Apple wants Thunderbolt? Nah, we can always go back to sell x86 chips.

Netbooks and consoles going AMD? Nah, We can always go back to sell x86 chips.

Market doesn't want x86 chips anymore? Nah, we can always go back to sell x86 chips. Oh wait-well, we can always sell cable TV, lols.
 
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |