Question Geekbench 6 released and calibrated against Core i7-12700

Page 16 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.
Jul 27, 2020
17,787
11,593
106
All CPU design companies I know are using GB because they know how to exploit its results (and because their customers want to know).
How many of these? https://www.inven.ai/company-lists/top-28-microprocessor-production-companies

I suppose there are more than a few good developers with deep knowledge of microprocessors. What is keeping such developers from creating a GB competitor with popular opensource software as tests? Could it be due to legal reasons? (a CPU performs badly in their benchmark and they get contacted by the CPU manufacturer's lawyers or something).
 

Nothingness

Platinum Member
Jul 3, 2013
2,722
1,357
136
I don't know most of them. And did you notice AMD is missing and Arm is named Armagh
Anyway you'll have to trust my words (yes, I agree it means nothing, as an anonymous I have few credentials).

I suppose there are more than a few good developers with deep knowledge of microprocessors. What is keeping such developers from creating a GB competitor with popular opensource software as tests? Could it be due to legal reasons? (a CPU performs badly in their benchmark and they get contacted by the CPU manufacturer's lawyers or something).
Attending a benchmark committee just once would allow you to answer that
 

TwistedAndy

Member
May 23, 2024
113
89
56
All what you say is applicable to all benchmarks, not only GB. Do you think it makes more sense to compare SPEC FP results of a CPU without AVX against a CPU with AVX?

SPEC results are also far from perfect, considering the test suite. I have covered this problem here.

Another problem is how those sets match the real-life performance. For example, Geekbench 6 includes a Clang compilation test. That's great if you are a developer, write your code in C and use Clang.

But if we take PHP, Python, Ruby, and other interpreters, they are heavily optimized for Intel CPUs. They show much higher performance than Apple Silicon (PHPBench, PyBench, Optcarrot). The difference is up to two times for ST tests.

There's no sense in measuring how fast a CPU renders a scene (Cinebench), for example, because in real life, you will be using a GPU for that.

Geekbench also includes a lot of AI tests. That's great, but why should you use a CPU for AI workloads if the dedicated NPU is nearly 10 times faster while consuming less power? It's the same situation as Cinebench.

The same can be said about the Ray Tracer test in Geekbench. Why should you do lighting calculations on a CPU if a GPU is hundreds of times faster because of the dedicated hardware blocks?

So yes, comparing Geekbench scores may be pretty funny, but it's a useless metric to compare the CPU performance on different platforms. To a certain degree, you can compare those results within one platform, but the test suite should be objective. Geekbench 6 is not an objective test.

Ultimately, the only viable metric is the actual performance of the apps you're using on the fixed power limit.
 

Nothingness

Platinum Member
Jul 3, 2013
2,722
1,357
136
So yes, comparing Geekbench scores may be pretty funny, but it's a useless metric to compare the CPU performance on different platforms. To a certain degree, you can compare those results within one platform, but the test suite should be objective. Geekbench 6 is not an objective test.
It's not only funny, but again it's definitely useful for people developing CPUs. Properly used any benchmark shows you something. It's just not for the typical end-users (where your comment below applies), but, outside of CPU design, it is still useful if you know what you're looking at.

Ultimately, the only viable metric is the actual performance of the apps you're using on the fixed power limit.
On that I definitely agree.
 

TwistedAndy

Member
May 23, 2024
113
89
56
It's not only funny, but again it's definitely useful for people developing CPUs. Properly used any benchmark shows you something. It's just not for the typical end-users (where your comment below applies), but, outside of CPU design, it is still useful if you know what you're looking at.

Yes, benchmarks, in general, may be pretty useful, especially when they are correct.

At the same time, many companies tend to manipulate that data for marketing purposes. Geekbench is the best example here.

Another field of manipulation is power efficiency.
 

Nothingness

Platinum Member
Jul 3, 2013
2,722
1,357
136
Yes, benchmarks, in general, may be pretty useful, especially when they are correct.
OK that becomes more specific What do you consider a correct benchmark? Don't list properties, give me a name.

At the same time, many companies tend to manipulate that data for marketing purposes. Geekbench is the best example here.

Another field of manipulation is power efficiency.
That's another topic that doesn't apply to GB only.
 

poke01

Golden Member
Mar 8, 2022
1,379
1,578
106
Much of what Twisted Andy says makes sense in that the end user does not care about benchmarks but performance in applications.

As for those tests that are included well it’s to test CPU performance and only that and if they applicable to you that’s great.

Benchmarks give us an idea of a CPU performs, that’s their function.

But what is really funny is that Apple never uses Geekbench, SPEC or Cinebench to showcase any CPU improvement in their testing. Intel, AMD, Qualcomm, MediaTek do in their keynotes to hype or either manipulate.

Let’s take an example of Apple’s last keynote and look at what they claimed regarding M4 performance improvement. They said M4 has up to 50% faster multi-threaded performance than M2 in the Affinity Photo 2 using the built in benchmark. They mention up to 50% because there’s various versions of M4.

Now let’s see if that up to 50% multithreaded claim holds up in Geekbench 5.



what do you know?? It does. Of course, this makes sense to compare because Geekbench 5 MT scales well. That’s why use the right benchmark that suits your needs.
 
Reactions: igor_kavinski

TwistedAndy

Member
May 23, 2024
113
89
56
What do you consider a correct benchmark?

I prefer using Cinebench R23 mostly because of its consistency, relatively long benchmark time, and great scalability. Cinebench 2024 is also pretty good.

Geekbench v5 is not that bad, but it's short and OS-dependent. It tends to show higher results on Linux and macOS. As for Geekbench v6, it's pure garbage, which has only one benefit: the base with results.

There are some other benchmark suites, like Puget, which represent real performance in real apps, but they are not so consistent.
 

TwistedAndy

Member
May 23, 2024
113
89
56
I never use R23, becuase it's not good for comparing x86 vs ARM CPUs.

Cinebench R23 has native support for Apple Silicon, and it does not use specific extensions missing on some platforms (SME, AVX512). So, unlike Geekbench v6, you can compare those results between platforms.

As for Cinebench R24, it's less representative from my perspective but also pretty consistent.
 

roger_k

Member
Sep 23, 2021
102
215
86
I prefer using Cinebench R23 mostly because of its consistency, relatively long benchmark time, and great scalability.

Cinebench R23 is a benchmark of SIMD throughtput and L1 cache bandwidth. It's as useful for assessing general CPU performance as someone's love of pizza is indicative of their knowledge of the Italian language. The 2024 version is barely any better. The only reason why Cinebench kind of works is because x86 CPUs historically didn't change much about their INT/SIMD balance, and their L1 is very fast.

By the way, I find it really difficult to understand how some can prefer Cibenench to Geekbench as a general-purpose benchmark. Cinebench-liek workload (Blender) is literally included as one of the workloads in Geekbench. My criticism of GB would be overreliance on CPU-driven ML workloads, which are not the best choice in the advent of the NPUs.

Cinebench 2024 fixes the issues that CB R23 had.

It fixes some issues, which does not make it a good benchmark. It's still represents a very narow workload, and it still uses AVX intrinsics mapped to Neon on ARM platforms.
 

TwistedAndy

Member
May 23, 2024
113
89
56
By the way, I find it really difficult to understand how some can prefer Cibenench to Geekbench as a general-purpose benchmark.

According to Geekbench, 16-core AMD Ryzen 9 7950X is 32% faster in the multi-core test than AMD EPYC 9754 with 128 cores:



Again. 16 cores are faster than 128 cores using the same architecture. Does it match the reality?
 

Nothingness

Platinum Member
Jul 3, 2013
2,722
1,357
136
Cinebench R23 has native support for Apple Silicon, and it does not use specific extensions missing on some platforms (SME, AVX512). So, unlike Geekbench v6, you can compare those results between platforms.

As for Cinebench R24, it's less representative from my perspective but also pretty consistent.
Cinebench R23 used an automatic translation layer to support NEON. It is a joke. Much worse than GB6.

And who cares about rendering performance on a CPU? ;-)
 
Reactions: Hitman928

Nothingness

Platinum Member
Jul 3, 2013
2,722
1,357
136
According to Geekbench, 16-core AMD Ryzen 9 7950X is 32% faster in the multi-core test than AMD EPYC 9754 with 128 cores:

View attachment 100597

Again. 16 cores are faster than 128 cores using the same architecture. Does it match the reality?
I have been saying for months that the only GB6 test that seems to scale is Ray Tracer. Here we see that 7950x scales by about 19, while 9754 scales by 108.
 

gdansk

Platinum Member
Feb 8, 2011
2,478
3,373
136
I know you're kidding, but claiming GB6 is a poor benchmark for MT just proves the point I've made today multiple times: if you don't properly use benchmark results, you get wrong answers. Don't blame the benchmark, blame the people that are clueless
I agree but there's no harm in telling everyone that comparing GB composite MT score is very limited.
 

FlameTail

Diamond Member
Dec 15, 2021
3,139
1,790
106
Let's say hypothetically Primate Labs adds back the GB5 style MT test to GB6...

Then there are 2 MT tests in GB6. How will they be named?

"Professional Multi-Score" and "Consumer Multi-Score"?

Or maybe:

"Embarassingly Parallel Multi Score' and "Dog's Multi Core"
 
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |