Nvidia appears to be the latest chip designer to use Samsung’s and GlobalFoundries’ 14nm node. The company won’t be alone, as the same node will be used by Apple, Qualcomm, AMD and of course Samsung.

According to Business Korea, the new node will go online in the second quarter of 2015. Apple, Qualcomm and Nvidia will be among the first outfits to use it, aside from Samsung. Samsung could lead the way with the first 14nm ARM SoC in the upcoming Galaxy S6, but this is still speculation and nothing is confirmed.

A Samsung official told the publication that 20nm is already a “mainstream technology” and that rivals are trying to compete with 16nm FinFET, but Samsung is in an advantage.

“Our company has already introduced the production process of 14 nm FinFETs,” the unnamed Samsung source told Business Korea.

Qualcomm is using TSMC’s planar 20nm node for its latest Snapdragon 810 flagship SoC and the chip seems to be experiencing some thermal issues, although Qualcomm is playing them down.

As for Nvidia, its first FinFET product should be Parker, a SoC based on the company’s custom 64-bit Denver core and Maxwell GPU.

Apple's new flagship SoC has no trouble taking on 32-bit ARM cores, but when it comes to other custom 64-bit cores, the A8X cannot pull ahead.

Luckily for Apple, there are no high-end 64-bit cores in currently Android products, apart from one. The Nexus 9 features the Tegra K1 64-bit SoC, Nvidia's first mobile chip based on a custom ARM core.

According to Apple Insider, the A8X has no trouble wiping the floor with products based on Snapdragon 800 and Exynos 5 Octa processors. Cupertino's new tablet chip beats them in Geekbench with a wide margin. It has no trouble taking on the Nvidia Shield tablet, which is marginally faster than the Kindle Fire HDX 7 and Galaxy Tab S 10.5.

It should be noted that this is a CPU benchmark, so GPU performance is taken out of the equation. In any case the A8X outperforms Samsung and Amazon tablets by roughly two to one in single core CPU tests.

However, when faced with another 64-bit custom core, the A8X hits a wall, but there's a catch. The A8X and Tegra K1 64-bit end up with the same single core score, but the Apple chip runs at just 1.5GHz, or 700MHz slower than the Nvidia chip. The A8X also pulls ahead in multicore tests, since it has three CPU cores, while the TK1 has just two Denver cores.

In multiple core benches the Denver-based Tegra ends up on a par with the 32-bit Tegra used in the Shield tablet, which means a single Denver core is as fast as two Cortex-A15 cores. It just goes to show that Apple's SoC performance lead can be matched with another custom core. There is no magic pixie dust involved, Apple's custom CPU cores are roughly on a par with Nvidia's design, although Apple still delivers more performance per watt at lower clocks.

Unfortunately there is still no data on GPU performance. We know what to expect from the new Tegra, since it uses the same GPU as the 32-bit variant, but there are no A8X GPU benchmarks yet.

Google will announce both the oversized Nexus 6 and the Nexus 9, its new 4:3 tablet powered by a Tegra K1 64-bit processor, at an event later today. The Nexus 9 has a 8.9-inch, 2048x1440 display (2048x1536 minus UI bars), 2GB of RAM and 16GB to 32GB of internal storage.

Android L will be called Android 5.0 and it will launch on the Nexus 6 and Nexus 9. We remember one industry figure pointing out that Android 4.0 was updated in 0.1 increments because Google was saving 5.0 for something bigger. Android 5.0 sports Android RunTime ART, a 64 bit-runtime and it finally brings 64-bit support to Android devices. Currently Android runs Dalvik on which Android Java code is executed on.

Apple iOS and Windows Mobile run natively compiled software that is optimised for hardware platforms used by their mobile devices. Both Apple and Microsoft have closed hardware environments, as Apple is using its A6, A7, A8 and other in-house processors, while Microsoft uses a range of Qualcomm processors.

Android uses multiple CPUs and architectures, so the majority of Android software is based around a generic code language which is transformed from “byte-code” into native instructions for the hardware on the device itself. Anandtech promises a significant performance uplift with ART in many benchmarks, but I guess we will know soon enough.

We will see if the 64-bit Denver core can excite many vendors and if priced well, the Nexus 9 should have a better of success than the Nexus 7 2013. This one had a Full HD panel that many consumers wanted, but also came with rather weak and obsolete Qualcomm S4 processor and a form factor that is no longer popular.

Nexus 6, a Snapdragon powered phablet is coming too, but we don’t think many people will want it. While phablets are popular in some markets, in others they are not, although that depends on branding, too. Both devices will be announced later today.

For much of the year we were under the impression that the second generation Maxwell will end up as a 20nm chip.

First-generation Maxwell ended up being branded as Geforce GTX 750 and GTX 750 TI and the second generation Maxwell launched a few days ago as the GTX 980 and Geforce GTX 970, with both cards based on the 28nm GM204 GPU.

This is actually quite good news as it turns out that Nvidia managed to optimize power and performance of the chip and make it one of the most efficient chips manufactured in 28nm.

Nvidia 20nm chips coming in 2015

Still, people keep asking about the transition to 20nm and it turns out that the first 20nm chip from Nvidia in 20nm will be a mobile SoC.

Our sources didn’t mention the exact codename, but it turns out that Nvidia wants to launch a mobile chip first and then it plans to expand into 20nm with graphics.

Unfortunately we don’t have any specifics to report.

AMD 20nm SoC in 2015

AMD is doing the same thing as its first 20nm chip, codenamed Nolan, is an entry level APU targeting tablet and detachable markets.

There is a strong possibility that Apple and Qualcomm simply bought a lot of 20nm capacity for their mobile modem chips and what was left was simply too expensive to make economic sense for big GPUs. 20nm will drive the voltage down while it will allow higher clocks, more transistors per square millimeter and it will overall enable better chips.

Just remember Nvidia world's first quad-core Tegra 3 in 40nm was rather hot and making a quad core in 28nm enabled higher performance and significantly better battery life. The same was true of other mobile chips of the era.

We expect similar leap from going down to 20nm in 2015 and Erista might be the first chip to make it to 20nm. A Maxwell derived architecture 20nm will deliver even more efficiency. Needless to say AMD plans to launch 20nm GPUs next year as well.

It looks like Nvidia’s 16nm FinFET Parker processor, based on the Denver CPU architecture and Maxwell graphics won’t appear before 2016.

The HTC Volantis tablet has been rumoured for months and it is supposed to ship with a powerful Nvidia processor, a high resolution screen and HTC’s trademark build quality.

WSJ says the tablet is a 9-incher. Details remain sketchy, but we know it is supposed to launch with 64-bit Android L and an appropriate 64-bit capable chip in tow. The silicon comes from Nvidia and chances are the Nexus 9 will be the launch device for the Denver-based Tegra K1 64-bit.

Although WSJ did not talk about specs, previous leaks point to a 2048x1536 display, aluminium zero-gap construction, 2GB of RAM and 16GB to 32GB of internal storage.

Nvidia has officially introduced the 64-bit variant of the Tegra K1 SoC. Unlike all previous Tegra designs, the TK1 64-bit is based on a custom CPU core.

Nvidia’s Denver 64-bit core has been long in the making and at last we’re about to see what it can do in real world tests.

The Tegra K1 implementation features two ARMv8 CPU cores, but it is pin compatible with the 32-bit version of the chip. Nvidia says this should ease implementation and speed up time-to-market.

First 64-bit ARM processor for Android?

Nvidia describes the 64-bit Tegra K1 as the “world’s first 64-bit ARM processor for Android” which is a bit of a stretch. Another stretch is the console graphics claim, but truth be told Nvidia's 192 Kepler cores should be on par with console GPUs of years goneby.

“Denver is designed for the highest single-core CPU throughput, and also delivers industry-leading dual-core performance. Each of the two Denver cores implements a 7-way superscalar microarchitecture (up to 7 concurrent micro-ops can be executed per clock), and includes a 128KB 4-way L1 instruction cache, a 64KB 4-way L1 data cache, and a 2MB 16-way L2 cache, which services both cores,” Nvidia said.

The chip features 128MB of main memory optimization cache. It reduces the need to re-optimise software routines. Dynamic Code Optimisation optimizes frequently used software routines into dense microcode routines.

“Denver looks across a window of hundreds of instructions and unrolls loops, renames registers, removes unused instructions, and reorders the code in various ways for optimal speed. This effectively doubles the performance of the base-level hardware through the conversion of ARM code to highly optimized microcode routines and increases the execution energy efficiency,” the company said in a blog post.

Nvidia working to implement Android L on Tegra K1

Nvidia promises “amazing” devices based on the new part later this year. It also stressed that it is developing Android L on the 64-bit Tegra K1.

It is still too early to say what Android L and 64-bit support have to offer in terms of sheer performance gains, but the first ARM 64-bit parts are slowly starting to trickle out. This should compel developers to start looking at ARMv8 hardware seriously.

Nvidia says dual-core processors can attain significantly higher performance than existing four- to eight-core mobile SoCs on most workload. Since it was optimized for single –core performance, Denver should capitalize on this fact. On a side note, Nvidia was the first ARM outfit to offer a dual-core Cortex A9 part and it later introduced the first 4+1 core SoC. Now it is going back to dual-core designs, at a time when many competitors are embracing bit.LITTLE designs with six or more small cores.

The first device based on the Tegra K1 will in all likelihood be HTC Volantis tablet, which is Google’s next Nexus tablet. We don't have any information on upcoming Denver-based products other than the HTC/Google tablet at this point. One thing is certain though. Nvidia is not going after smartphone design wins with its first implementation of Denver.

A part of Nvidia's Financial Q1 2015 conference call Q&A session included some questions about micro servers, whether or not the 64-bit Tegra K1 can make it into the GRID market.

Nvidia's CEO Jen-Hsun Huang was a straight shooter saying that Nvidia is “seeing a lot of interest in putting something like Tegra in micro servers,” but he added a caveat: “one step at a time, one step at a time."

Jen-Hsun addressed the importance of the software component, or software stack for this market. He points out that the software stack Nvidia is building for GRID will eventually used on top of Tegra.

Denver could have what it takes

The Denver 64-bit architecture used in the Tegra K1 64-bit might have a shot in the micro server market as it will offer a lot of compute power and the CPU core should end up faster than the Cortex A57. Back at GTC 2014 we saw a demo at the GE booth where a single Jetson board hooked up to a camera over 10Gbit network port managed to trace a multiple targets at once.

The significance of such capabilities is that they can could easily find applications in the security market, drones, self-driving cars and all this can be done with close to 10W of power. Of course, this can be done with existing chips, but the CUDA powered Tegra K1 needs significantly less power to pull it off, which results in a much smaller footprint than say a 100W system that would handle the same task.

Eventually Nvidia could make a move in the traditional server market. The way things are going some 64-bit Cortex A57 servers might hit the market in early 2015 and Denver 64-bit might be the only custom based 64-bit ARM core ready at that time. Qualcomm’s 64-bit Krait is expected in the first half of 2015 and it might give Nvidia, AMD and other players a run for their money.

Does an ARM server push make sense for Nvidia?

In order to be successful in server market you need software, hardware, a great field application engineer network, customers that want to work with you and this is exactly why Calxeda failed. Nvidia has an advantage over Intel as it can do ARM based micro server, but then again, everyone else can. AMD is taking the ARM server market quite seriously and it has a lot more experience in the field than Nvidia.

Should Nvidia choose to proceed with an ARM server push, it will ultimately be a matter of research and development money that company can afford to put behind this risky strategy. Nvidia is making some progress with Tesla and Grid into the server market, but it will take some great products to convince Dell, HP and the rest of the market go your way.

In theory, Nvidia could leverage its compute software stack, but that’s only relevant in a handful of niches. AMD is targeting a much wider micro server market, while Nvidia could go after niche systems that would benefit from its compute tech. However, whether or not this niche is worth the investment remains to be seen. Both Nvidia and AMD can offer unique compute capabilities that could differentiate their ARM parts from the rest of the field (CUDA, Open CL), rendering their ARM server parts more competitive in a number of market segments.

Nvidia is running quite a tight ship when it comes to Tegre related information, so we don't know much other than what CEO Jen-Hsun Huang revealed at the company's recent GPU technology conference. He did mention Parker and the fact that it will come with Maxwell graphics, and he did share quite a bit about Pascal.

However, when you inquired about anything other than that, Nvidia had nothing much to add other than the fact that Spiderman aka Parker and Volta come after Pascal, insisting that they are not delayed. We got confirmation from Nvidia that its first Denver part got an official name. It will be called Tegra K1 64-bit, kind of a straight shooter approach. Graphics fans are familiar with Tegra, K denotes Kepler and the 64-bit moniker helps. The other K1 is of course 32-bit based, as it still employs ARM's Cortex A15 core. The K1 64-bit brand was confirmed by GM of Tegra unit Deepu Talla.

The chip is expected in 2H 2014 and apparently this dual-core should end up faster than Cortex A57 based 64-bit high performance processors. It will be interesting to see how Nvidia plans to handle marketing, as its dual-core parts are supposed to compete with quad-cores.

Apple has managed to do it with the Cyclone core, but since Apple's marketing department looks and feels like a new religious movement, it is not the best comparison.

GM of Nvidia's Tegra business unit Deepu Talla told us that the general expectation is that Denver CPU performance will generally be better than what we can expect from a standard ARM Cortex A57 core. He said that when Nvidia builds an architecture, it usually doesn’t want to match the competition, it tries to make it faster than the reference.

Since A57 will be the highest performing ARM 64-bit core, the general idea is that Nvidia has to beat that to get some traction. Talla said that we can expect the Denver dual-core Tegra K1 64-bit part to end up faster than A57-based CPU.

This is a big deal and if works well for Nvidia might create quite a popularity for Nvidia’s first 64-bit SoC. It will be great for Nvidia if they can deliver on this promise. The shipping date for Denver based Tegra K1 is still the second half of 2014 and we don’t know how Nvidia plans to price or position its 32-bit and 64-bit Tegra K1, at least not yet.

We were told that we should expect Tegra K1 in phones, tablets, consoles, automotive, dev kits such as the Jetson board, and all other markets that Nvidia addressed with previous Tegra processors.

An interesting Tegra K1-based prototype has popped up in Antutu. Unlike the K1 demoed by Nvidia at the Consumer Electronics Show in January, the chip is a dual-core affair and it is a 64-bit part.

The Tegra K1 is supposed to show up in two flavours and the first one to appear will feature four Cortex A15 CPU cores. The 64-bit version, based on Nvidia's custom Denver ARMv8 CPU core, is expected to show up later this year. The Antutu listing leaves us with more questions than answers. The unnamed device runs Android 4.4.2, it has a 1920x1080 screen (no word on size), 2GB of RAM, 32GB of storage, 2MP front facing camera and a 13MP rear camera. The megapixel count on the latter points to a smartphone, although the Tegra K1 should feel more at home in a tablet.

The clock ranges from 510MHz to a whopping 3.0GHz, which sounds rather high. The device scores 43617 in Antutu, slightly less than the A15-based version which scored 43851. However, this is relatively good news, as two 64-bit cores can hold their own against four A15 cores. Both the 64-bit and 32-bit versions of the Tegra K1 end up 20 to 25 percent faster than the old Tegra 4 and Qualcomm's Snapdragon 800. So could this be an all new Cortex A57 version? Probably not, as the A57 would not outperform the A15 by a factor of two to one. There is no word on the GPU either - so we can't say for sure that it has the same 192-core Kepler GPU used in the 32-bit Tegra K1.

To say that benchmark leaks are not a very reliable source of information would be an understatement, but this one makes sense. Whether or not Nvidia can score enough design wins with the K1 remains to be seen, but it is safe to assume that its focus will be on tablets and form factors other than smartphones. It's a big, power hungry chip with no on-die LTE modem, so we don't expect to see many smartphone design wins.