nvidia tegra K1

Page 4 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

BallaTheFeared

Diamond Member
Nov 15, 2010
8,115
0
71
but just going from 0.8V to 1V would bump that up to ~3.9W. (For comparison, desktop Kepler goes from something like 0.875V to 1.125V I believe it is?)

I don't think you can really go off desktop in this case.

Boost 2.0 shoots for 1.16v.

My 780 will run 900Mhz fixed with 0.94v, its like 50% faster than my 7950 was at 800\1250 0.85v and killed in it perf/w. 174w sys for the 7950 and like 215w for the 780.
 

sontin

Diamond Member
Sep 12, 2011
3,273
149
106
But at max clocks under full load 5W is not going to cut it. That is easily seen in a lot of devices.

4x A15 @1.8 ghz + GPU at full load is over 5W.

That's the reason why Tegra 4 downclocks the cores to 1,4GHz when there is load on all four cores.
And if you combine both the DFVS will downclock the CPU and/or GPU to stay within the power budget.
 
Last edited:

IntelUser2000

Elite Member
Oct 14, 2003
8,686
3,786
136
Tegra 4 had about 3000 multi-thread on geekbench 3. Their whitepaper has K1v1 at 1.4x performance. So, thats 4200+ multi thread. Denver has 2 cores, so we're already at least 2100 points for single-thread (I said near 2,000 in my last post, and assuming better multi-thread scaling).

Don't be so quick to conclude...

The 1.4x is comparison at same power. The maximum performance is up approximately 15%.
 

ams23

Senior member
Feb 18, 2013
907
0
0
Thanks for providing a source - http://www.nvidia.com/content/PDF/tegra_white_papers/Tegra-K1-whitepaper-v1.0.pdf for those who don't want to search for it, first paragraph of page 10. And that definitely explains the discrepancy seeing as how I'd be surprised if an 'average of populate Android games' even gets the GPU up to a 50% load. Especially if that's run on an equivalent to the Tegra Note 7 at only 1280x800 resolution with a 60 fps cap.

The Tegra Note 7 prototype chassis's used to demonstrate Tegra K1 had a 1080p screen resolution.
 

witeken

Diamond Member
Dec 25, 2013
3,899
193
106
Denver is a massive upgrade, I think it will easily be competitive with Airmont/Goldmont CPU wise. It could be competitive with i3-U and Broadwell-i5/i7-Y. Denver will likely have higher multi-thread peformance than K1v1 (A15r3).

Tegra 4 had about 3000 multi-thread on geekbench 3. Their whitepaper has K1v1 at 1.4x performance. So, thats 4200+ multi thread. Denver has 2 cores, so we're already at least 2100 points for single-thread (I said near 2,000 in my last post, and assuming better multi-thread scaling).

i5-4200U is ~2500, 4300U is about ~2700.

And if that one GFXbench result is accurate to true GPU performance, then we're seeing HD4400 level performance with 3x less TDP.

All this on 28nm is pretty incredible, imagine 16nm FinFET (though 20nm BEOL) will give an additional 35% perf (according to TSMC), that is 4500U perf. territory.

A lot of ifs, but Denver is a massive jump in terms of CPU and GPU compared to past iterations.

I'm a bit less optimistic. Denver is Nvidia's first CPU, I do not expect it at 28nm to be a threat to 22nm or 14nm Tri-Gate transistors.

It might be somewhat competitive with Silvermont, but certainly not Airmont, let alone Goldmont which will be Denver's main competitor .
 

witeken

Diamond Member
Dec 25, 2013
3,899
193
106
What makes you say that A7 [Cyclone] CPU consumes "tons of energy"? A7 seems to work fine in the relatively small and thin iphone 5s (albeit with some throttling compared to the variant in ipad Air).
1 Cyclone core consumes about as much energy as 3 Silvermont cores. Single threaded performance is only 40% faster -> performance/watt of Cyclone is 2x worse. It would surprise me if Denver would be any different.
 

Roland00Address

Platinum Member
Dec 17, 2008
2,196
260
126
So the 192 kepler graphics is comparable to an 8800 gts (old g80 version) or a 8800 gs (g92) in gigaflops. Since the architecture is similar I assume you would get similar gaming performance if the software allows... correct?
 

Khato

Golden Member
Jul 15, 2001
1,248
321
136
The Tegra Note 7 prototype chassis's used to demonstrate Tegra K1 had a 1080p screen resolution.

Now that I finally found some coverage that shows that slide, it's actually 1920x1200. But that's beside the point - NVIDIA's statement that K1's GPU consumes an average of 2W playing popular Android games doesn't tell us anything about actual load power consumption. They could have run those tests at 1280x720 just as easily as 2560x1440... and given that the performance levels they're promising should be capable of running the majority of Android games at a constant 60 fps even at 2560x1440 the resolution will directly correlate to power consumption.

Not that it really matters much regardless for exactly that reason. Until games catch up to make use of the feature set and computer power that K1 will offer it can run at a far more efficient frequency/voltage rather than full power.
 

ams23

Senior member
Feb 18, 2013
907
0
0
NVIDIA's statement that K1's GPU consumes an average of 2W playing popular Android games doesn't tell us anything about actual load power consumption.

Well, to fit inside a thin fanless tablet, the TK1 GPU peak power consumption really cannot be much more than ~ 4w (and this level of consumption from the GPU would require an extremely GPU-intensive application to be running).
 

NTMBK

Lifer
Nov 14, 2011
10,322
5,352
136
Well, to fit inside a thin fanless tablet, the TK1 GPU peak power consumption really cannot be much more than ~ 4w (and this level of consumption from the GPU would require an extremely GPU-intensive application to be running).

It would indeed be limited to 4W in a fanless tablet. But is that with the K1 throttling? If so, how badly? And how much power would it consume if running at full speed? That's the question. It's a lot trickier to get a straight answer (from anyone!) these days, with configurable TDPs, SDPs, throttling, turbo...
 

Khato

Golden Member
Jul 15, 2001
1,248
321
136
It would indeed be limited to 4W in a fanless tablet. But is that with the K1 throttling? If so, how badly? And how much power would it consume if running at full speed? That's the question. It's a lot trickier to get a straight answer (from anyone!) these days, with configurable TDPs, SDPs, throttling, turbo...

Precisely. And it tends to be the case that it's far more important to focus on the GPU behavior in this market since gaming is pretty much the only sustained 'heavy' workload that the devices are subjected to. The CPU being capable of spiking up to ~7 watts (roughly how much the A7 appears to draw under full CPU load) doesn't matter much in terms of the chassis' thermal dissipation capabilities because it's not sustained. Even a benchmark that runs for a few minutes may not be enough to get the device perceptibly warm. But when you're playing a game for half an hour or more? Yeah, then you're limited to a certain steady-state thermal limit.

So it could quite possibly be the case that when playing games steady-state the K1 GPU is limited to roughly 2W... while peak consumption to hit that 365 GFLOPS number is 5W or more. Wouldn't it be nice if they just gave all the relevant information? After all, they've already presented enough to make it clear that it beats all the current competition on performance and efficiency.
 

ams23

Senior member
Feb 18, 2013
907
0
0
Most likely the 365 GFLOPS GPU throughput variant (in the Xbox360 and PS3 slide) refers to Shield v2. Based on the current gen Tegra (where Shield has 15% higher GPU performance than Tegra Note 7), I would expect closer to ~ 320 GFLOPS GPU throughput for TK1 in a tablet.

Hope to hear more details at Mobile World Congress 2014 next month and/or GTC 2014 in two months.
 

Bryf50

Golden Member
Nov 11, 2006
1,429
51
91
950mhz on a 192 Cuda Core GPU is way overly optimistic for 28nm. Nvidias 384 core mobile GPUs can't clock that high and they have a 35w tdp to work with. In any application that people will actually buy(phone or tablet) we're never going to see these peak gpu and cpu clocks in practice.
 
Last edited:

BallaTheFeared

Diamond Member
Nov 15, 2010
8,115
0
71
950mhz on a 192 Cuda Core GPU is way overly optimistic for 28nm. Nvidias 384 core mobile GPUs can't clock that high and they have a 35w tdp to work with. In any application that people will actually buy(phone or tablet) we're never going to see these peak clocks in practice.

Sure, if you think everything needed to run two SMX together will be required to run one only.

I think Anand covered this very subject.
 

Bryf50

Golden Member
Nov 11, 2006
1,429
51
91
Sure, if you think everything needed to run two SMX together will be required to run one only.

I think Anand covered this very subject.

If you take a GeForce 740M (2 SMX/384 CUDA core design), you’re looking at roughly a 19W GPU. Of that 19W, around 3W is memory IO, PCIe and other non-GPU things. You can subtract another 6W for leakage, bringing you down to 10W. Now that’s a 2 SMX design, so divide it in half and now you’re down to 5W. Drop the clock from 1GHz down to 900MHz, and the voltage as well, and now we’re talking around 2 - 3W for the GPU core and that’s without any re-architecting. Granted you can’t just subtract out things like leakage like that, but you get the point. Kepler wasn’t a bad starting point for a good mobile GPU design.
19W seems remarkably low. I can't find any concrete info but most search results calling the 740m a 19W part come from the past few days. Prior to that most people were citing a 33W tdp.
 

Enigmoid

Platinum Member
Sep 27, 2012
2,907
31
91
Nonsense. TDP for Tegra 4 is 5W.

19W seems remarkably low. I can't find any concrete info but most search results calling the 740m a 19W part come from the past few days. Prior to that most people were citing a 33W tdp.

I think Nvidia is pulling the wool over our eyes. GK107 @ 1000 mhz w/ DDR3 uses much more than 19W in power. 650 desktop is a 50W+ part. Mobile chips are better binned but still suck down the power. In AT review of the razar edge, anand mentioned that the 640m LE was a 20W part. That chip runs at ~600mhz. I do not believe that nvidia got that much improvement over the last couple years on the same process and without a new stepping.

Note that nvidia has another SKU (GK208) which has a 64 bit bus. With DDR3 the chip is tremendously gimped and will likely use much less power. Still the scaling looks suspicious.

Look at AMD's kabini chips. 128 gcn cores at max 600 mhz on a much larger power envelope. I simply do not see 192 kepler cores at 950 mhz using 4W under load.
 

ams23

Senior member
Feb 18, 2013
907
0
0
The expectation is that TK1 variants that go into tablets will have a GPU clock operating frequency close to ~ 850MHz, while a higher frequency variant ~ 950MHz can go into an actively cooled device such as Shield v2.

TK1 has significant re-architecting (and an improved 28nm HPM fabrication process) to help fit inside a mobile power envelope.
 
Last edited:

VirtualLarry

No Lifer
Aug 25, 2001
56,554
10,171
126
I think Nvidia is pulling the wool over our eyes.
...
Look at AMD's kabini chips. 128 gcn cores at max 600 mhz on a much larger power envelope. I simply do not see 192 kepler cores at 950 mhz using 4W under load.

I agree!
 

joshhedge

Senior member
Nov 19, 2011
601
0
0
950mhz on a 192 Cuda Core GPU is way overly optimistic for 28nm. Nvidias 384 core mobile GPUs can't clock that high and they have a 35w tdp to work with..

Yes they can, I run my 650M at 1100MHz stable with okay thermals..
 

Bryf50

Golden Member
Nov 11, 2006
1,429
51
91
Yes they can, I run my 650M at 1100MHz stable with okay thermals..
I never doubted that a 650m can physically run at 1100MHz if the cooling was sufficient. But at the stock clocks and tdp(a whopping 45W for the 650m) it does not.
 
Last edited:

jdubs03

Senior member
Oct 1, 2013
825
435
136
I'm a bit less optimistic. Denver is Nvidia's first CPU, I do not expect it at 28nm to be a threat to 22nm or 14nm Tri-Gate transistors.

It might be somewhat competitive with Silvermont, but certainly not Airmont, let alone Goldmont which will be Denver's main competitor .

In terms of single-thread performance, Tegra 4 is roughly on-par with Silvermont cpu-wise, at least according to geekbench, though obviously this is just one benchmark, it is the "go-to" benchmark a lot of the time. However, in tests like webxprt, its not even close, my venue 8 pro (Z3740D) did 555 which is 170 pts higher than T4, and beats Apples A7.

I'm thinking that Denver would essentially be the K1v1 x2 per single-thread performance. That may be optimistic, but if multi-thread performance is higher, then I can't see it another way.

If that is optimistic, what about x2 Tegra 4 for single-thread, using geekbench-32 bit that still puts it at 1900, and beats Haswell-Y series chips.

In essence they will be very competitive against the cherry trail, and likely at least broadwell-y.
 

witeken

Diamond Member
Dec 25, 2013
3,899
193
106
Why do you think that Denver will have 2x Tegra 4 single thread performance? What about power consumption?

It seems that Tegra K1 Denver will be made on a 28nm process. Do you really think Nvidia's first CPU will be on par with a 14nm 5th gen Core CPU?
 

sontin

Diamond Member
Sep 12, 2011
3,273
149
106
Look at AMD's kabini chips. 128 gcn cores at max 600 mhz on a much larger power envelope. I simply do not see 192 kepler cores at 950 mhz using 4W under load.

nVidia spent more money than AMD on their GPU tech. Kepler is the first architecture they really focused on efficient. Kepler.M is an evolution of the concept. nVidia has started to develop new architectures with mobile in mind. Maxwell will be the first architecture of this concept.

It is possible to deliver more performance with less power when you focus on this.
 
Last edited:

ams23

Senior member
Feb 18, 2013
907
0
0
unless I missed it ut wasn't the 5W figure an estimate for gles3 android games? or total chip power?

No. According to the whitepaper, the average power consumed by the Kepler.M GPU with some of today's best Android games is < 2w. This makes sense because Kepler.M easily handles most Android games today. Pushed to peak levels, Kepler.M really cannot consume more than ~4w to fit in a tablet form factor.

5w is the rated TDP for the entire TK1 SoC (same as T4).

The perf. per watt (and hence perf.) of the Kepler.M GPU is well beyond any high end ultra mobile GPU used today.
 
Last edited:
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |