Question Geekbench 6 released and calibrated against Core i7-12700

Page 7 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

Det0x

Golden Member
Sep 11, 2014
1,053
3,078
136
P.S. I do think that excluding AVX-512 was a bit of a shame. If a CPU supports the technology and can perform faster on a task, it should be used. At the same time, I do understand the decision — AVX512 isn't really used all that much in desktop software, and basing benchmarks off it might create unrealistic expectations.
Pretty sure AVX512 will be enabled in a patch shorty after sapphire rapids HEDT become widely available
 

roger_k

Member
Sep 23, 2021
102
215
86
You are plain misleading people with statements like this. There is a huge difference between having all cores available and effectively scaling across all cores available. For single workloads you can measure a chips MT performance until the point that workload stops scaling with any more threads. Below that point the chip is the bottleneck, above that point the workload is the bottleneck. The more cores a chip has the more workloads are incapable to effectively scaling across all cores available, the more GB6's MT score is hampered and misleading by its increasing larger part of bottlenecking benchmarks. With GB6's MT score on bigger chips you are not measuring the overall MT performance of the chip but a mix of ST performances limited to the cores effectively used and workloads incapable of scaling across the remaining idling cores.

Nobody is interested in a score representing workloads bottlenecking themselves.

I certainly agree that GB6 MT benchmarks lack transparency. It is not clear how the parallel work is done and where the bottlenecks lie. It would help if the source code for these tests were published so that the community can verify it.

That said, it is not clear to me why what you write has to be true. How do you know that the diminishing returns observed in GB6 are due to cores running idle and not due to the overhead of synchronisation between different cores/clusters/caches?
 
Reactions: Kaluan

Kocicak

Senior member
Jan 17, 2019
982
973
136
Has anybody tried to run the benchmark on AMD 5700G? I tried but it stopped the moment after I started it - is the problem in the benchmark itself or in my system?
 

whoshere

Junior Member
Feb 28, 2020
14
38
91
I don't care much about the choice of ST benchmarks in GB6 but its selection of MT benchmarks looks to me just terribly irrational by basically negating extra cores and introducing one of many MT workloads limitations.

GB5 showed very nice MT scaling, something you'd expect from your CPU, GB6 shows ... what?

I've read John Poole's explanation - sorry, I can't relate.
The new "shared task" approach requires cores to co-operate by sharing information. Given the larger datasets used in Geekbench 6 several workloads are now memory-constrained, rather than CPU-constrained, on most systems.

Here are two Geekbench 6 results for a 5950X system, one result run with single-channel RAM and one result run with dual-channel RAM:


Several workloads, including File Compression, Object Detection, and Photo Filter almost double in performance when moving from single-channel to dual-channel RAM.

For myself I will stick to GB5 testing and results. I don't care if it's "outdated" - its results do translate into real-world performance. GB6 MT test computes God knows what.


Also years ago I reported that the Linux version of the benchmark is considerably faster both in ST and MT modes than the Windows version and this has not been fixed or addressed. This doesn't have to do anything with the Windows process scheduler - this is reproducible in safe mode + real time priority with zero background tasks. [ Artem S. Tashkinov ]
 
Last edited:

Exist50

Platinum Member
Aug 18, 2016
2,452
3,101
136
GB5 showed very nice MT scaling, something you expect from your CPU, GB6 shows ... what?
MT scaling and performance implications for a number of real world use cases. I'd argue that solely having embarrassingly parallel workloads is far less representative of actual MT performance demands.
 
Reactions: Nothingness

Markfw

Moderator Emeritus, Elite Member
May 16, 2002
25,736
14,767
136
MT scaling and performance implications for a number of real world use cases. I'd argue that solely having embarrassingly parallel workloads is far less representative of actual MT performance demands.
For workstations ? They need a lot of cores. For servers, they need a lot of cores. For HEDT or people that do encoding, rendering and the like for desktop, they need a lot of cores. So this benchmark you are telling me is only worth using for the like of 12700k or below ???
 

Exist50

Platinum Member
Aug 18, 2016
2,452
3,101
136
For workstations ? They need a lot of cores. For servers, they need a lot of cores. For HEDT or people that do encoding, rendering and the like for desktop, they need a lot of cores. So this benchmark you are telling me is only worth using for the like of 12700k or below ???
Yeah, for servers or higher end workstations, you'd want a different, better benchmark. But Geekbench was never great for those either, so I think it's an improvement that they've mostly chosen a lane to stick to.

Like, for content creation workstations, PugetBench is pretty great. Severs are trickier because of how much individual use cases vary, but SPEC has a couple that should be a good fit.
 

Markfw

Moderator Emeritus, Elite Member
May 16, 2002
25,736
14,767
136
Yeah, for servers or higher end workstations, you'd want a different, better benchmark. But Geekbench was never great for those either, so I think it's an improvement that they've mostly chosen a lane to stick to.

Like, for content creation workstations, PugetBench is pretty great. Severs are trickier because of how much individual use cases vary, but SPEC has a couple that should be a good fit.
It just seems to me that making a benchmark that is worthless for any more than 20 threads/14 cores is pretty stupid. I could see stopping at 64c/128t, as that includes all kinds of HEDT workstations, etc... and only excludes servers. But this is worthless for all but mid to low end desktop.
 
Reactions: moinmoin

Exist50

Platinum Member
Aug 18, 2016
2,452
3,101
136
It just seems to me that making a benchmark that is worthless for any more than 20 threads/14 cores is pretty stupid. I could see stopping at 64c/128t, as that includes all kinds of HEDT workstations, etc... and only excludes servers. But this is worthless for all but mid to low end desktop.
There's really no "one size fits all" benchmark, and of those two markets, the consumer one (up to mainstream desktops) is by far the bigger, and the one Geekbench has traditionally targeted. Frankly, I don't think much was lost at all. No one actually buying servers used Geekbench to compare them. Barely more useful than the Cinebench dick measuring contests.
 

whoshere

Junior Member
Feb 28, 2020
14
38
91
John Poole has replied:

At the start of Geekbench 6 development in 2020, we collected benchmark results for client and workstation applications across various processors. We found that only some applications scale well past four cores. We also found that some applications exhibit negative scaling (where performance decreased as the number of threads increased). We concluded that, at some point, client applications experience diminishing returns with increased core counts due to the inability to use all available cores effectively. The investigation led us to believe that Geekbench 5 overstated multi-core performance for client applications.

One design goal for Geekbench 6 was to accurately reflect multi-core performance for client applications while not arbitrarily limiting workload scaling. We wanted to ensure the multithreading approaches used were reasonable and representative of how applications use multiple cores. We also wanted to ensure that no workloads exhibited excessive negative scaling.

To achieve this goal, we switched from the "separate task" approach to the "shared task" approach for multithreading in Geekbench 6.

The "separate task" approach parallelizes workloads by treating each thread as separate. Each thread processes a separate independent task. This approach scales well as there is very little thread-to-thread communication, and the available work scales with the number of threads. For example, a four-core system will have four copies, while a 64-core system will have 64 copies.

The "shared task" approach parallelizes workloads by having each thread process a single shared task. Given the increased inter-thread communication required to coordinate the work between threads, this approach may not scale as well as the "separate task" approach.

What the shared task is varies from workload to workload. For example, the Clang workload task compiles 96 source files, while the Horizon Detection workload task adjusts one 24MP image.

For client systems, the "shared task" approach is most representative of how most client (and workstation) applications exploit multiple cores, whereas the "separate task" model is more representative of how most server applications use multiple cores.

Some Geekbench 6 workloads will scale poorly, and others will scale well on high-end workstation systems. These results follow what we observed from real-world applications.
 
Jul 27, 2020
17,741
11,528
106
John Poole has replied:
I like Artem's idea of different multicore scores for different types of multicore tasks. I still think that at the end, they should run a few of the MC tests concurrently and see how the CPU handles multiple tasks at once, since that would lead to cache evictions and it would bring memory controller and cache design strengths into play. This capability needs to be reflected in the final score.
 
Jul 27, 2020
17,741
11,528
106

Also for text processing, the multi-core performance was 137.2 pages/ sec for the 64-core TR Pro while it was 136.1 for the 4-core D-1718T.

What kind of crappy benchmark can't discriminate between 4 cores and 64 cores?

It's broken. Primate Labs need to admit that their research effort was all wasted. They need to go back and learn how to do multicore programming from scratch.
 

Markfw

Moderator Emeritus, Elite Member
May 16, 2002
25,736
14,767
136



What kind of crappy benchmark can't discriminate between 4 cores and 64 cores?

It's broken. Primate Labs need to admit that their research effort was all wasted. They need to go back and learn how to do multicore programming from scratch.
Like when my 128 core Milan came in losing to a 13600k
 

naukkis

Senior member
Jun 5, 2002
768
634
136
Yeah. Their whole point seems to be, most real application developers can't write good multicore code so why should we? Here. Now our benchmark has the same crappy code as the real world!

Multi-core speedup is limited by Amdahl's law. If everything would scale to massive amount of threads we shall not need faster cores at all - just add more slow ones. For most of desktop use cases cores beyond 8 are totally worthless.
 

Markfw

Moderator Emeritus, Elite Member
May 16, 2002
25,736
14,767
136
Multi-core speedup is limited by Amdahl's law. If everything would scale to massive amount of threads we shall not need faster cores at all - just add more slow ones. For most of desktop use cases cores beyond 8 are totally worthless.
So you are saying this benchmark is ONLY for use on desktop PC's ? not even high end ones that CAN use up to 32, or even 64 cores ?
 
Reactions: Kaluan
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |