• Hey, guest user. Hope you're enjoying NeoGAF! Have you considered registering for an account? Come join us and add your take to the daily discourse.

NVIDIA Pascal GPU to feature 17B transistors and 32GB HBM2, coming Q2 2016 (earliest)

Hasney

Member
They're talking NVLINK in that article, defo going to build a system with a couple of NVLINK 1000 series. Shame even by then we'll likely be stuck with Skylake, but eh, not like CPU is the issue still.
 

AlStrong

Member
Second thing is, Nvidia should have plenty of headroom on 16nm FF+ after Pascal, after finally moving from 28nm to 16nm FF+ - Volta could be to Pascal what Maxwell was to Kepler. A major refresh, not completely new architecture. It's often been said both Pascal and Volta will make use of NV Link.

Perhaps one difference between Pascal and Volta might be Pascal uses HBM2 and Volta uses HBM3 or the Hybrid Memory Cube they originally planned for the Volta intended for 2016.

Pure speculation of course. Feel free to tell me where I'm wrong, as I probably am on some things.
or wrong on everything

Speculation is fun. :)

I had to wonder if perhaps Pascal is mostly a shrunk Maxwell whose introduction was a result of the 16nmFF delay, and they wanted to play it safe by using a mature design first ( a bit like Intel's tick-tock*) while making the aforementioned changes/additions (NVLink, HBM, mixed precision) instead of jumping straight to a more significant rejigging on an expensive node. **

I know you mentioned kepler <-> maxwell as not being a major architecture shift, but I consider it to be fairly changed up in a number of ways. IMHO, architecture refresh example would be 750Ti vs 9xx series( or even the GCN series &#3232;_&#3232; ).

*always hated that the "tick" was the shrunk version of the previous "tock". :|
** i.e. Take tegra x1 GPU, which is Max-based, make it super phat & add/change up a couple things, and shift from 20nm to 16FF+... ¯_(&#12484;)_/¯
 

Azulsky

Member
Kinda feel like its a long shot that NVLINK will be doable on consumer.

Needs motherboard and CPU to be compliant. Is Nvidia going to go back into the chipset business?I cant see Intel playing nice with them because they are competing companies in the gpgpu space with the Xeon Phi line.


More hoping that they get can get improved multi-gpu scaling as it looks like we are going to need it for Hi-Resolution VR. Seems like that is one of the promises that the silicon interposer offers for multi-die, single card configurations.

I'm just hoping we have a way to do multi-gpu and PCI-E SSD's without compromising bandwidth on the CPU's controller. That might be solvable with Skylake E.

I still feel like my 780TI's run just about everything at 2560x1600. Star Citizen runs flawlessly and that's about the peak of usefully pretty.

Just need to last until Skylake E and the first HBM Titan
 

tuxfool

Banned
I'm still maintaining that NVLink will only be used for servers.

Everything they have said about Nvlink, such as the requirement for special physical connectors makes me think that the protocol requires even stronger signal integrity requirements than PCIE. The physical protocol seems to be be using very high clockspeeds otherwise they wouldn't be worried about the Impedance of connectors used in PCIE. Those preliminary connectors they showed look very much like shielded interfaces used in RF.

This essentially means you cannot just shove in something like a SLI connector unless it is very carefully designed. I could definitely see it being used on Multi Gpu cards though.
 

Kieli

Member
Why would they ever cram 17 billion transistors on a chip even if they could? This article doesn't make a lot of sense.

It's a much, much better business strategy to make incremental improvements (maybe 10%/20% improvements tick over tock) and milk it dry. You can go from 8 bil to 10 to 12 to 14 to 18, and you have 4 generations of products to sell rather than 1. Also prevents the consumer from expecting too much because you gave them too good a deal for Pascal.
 

Renekton

Member
Why would they ever cram 17 billion transistors on a chip even if they could? This article doesn't make a lot of sense.

It's a much, much better business strategy to make incremental improvements (maybe 10%/20% improvements tick over tock) and milk it dry. You can go from 8 bil to 10 to 12 to 14 to 18, and you have 4 generations of products to sell rather than 1. Also prevents the consumer from expecting too much because you gave them too good a deal for Pascal.
They will safely do that once they finished curbstomping AMD into bankruptcy.
 

Volotaire

Member
So at what point will this filter down into a sub-$300 card?

Am I wrong in suggesting there will be a range of cards on market juts like other NVIDIA series? So, you would be able to purchase a 750/760/770 equivalent card that would be sub $300 (770 equivalent more akin to 300-400), and then 780/790 equivalent will be more expensive.
 

Renekton

Member
Am I wrong in suggesting there will be a range of cards on market juts like other NVIDIA series? So, you would be able to purchase a 750/760/770 equivalent card that would be sub $300 (770 equivalent more akin to 300-400), and then 780/790 equivalent will be more expensive.
They may start doing the refresheroo for lower-end cards.
 

AmyS

Member
The Pascal equivalent of GTX 680 (GK104) and GTX 980 (GM204) certainly won't need 17 Billion transistors. That number is obviously for the flagship GPU, 'Big Pascal'. unofficially referred to as GP100. It's totally reasonable to expect a range of Pascal GPU sizes, serving a full range of cards, as Nvidia always has. That's a given.

Nvidia may have taped out Big Pascal first but I'd expect them to launch cards using a mid-sized Pascal GPU first, as they did with both Kepler (680, march 2012) and second gen Maxwell (980, 970, fall 2014). Big Pascal will probably be reserved for Quadro, the next Titan and the successor to 980 Ti. However, just because Nvidia has done similar things for 2 gens in row, doesn't mean they're locked to do things the same a 3rd time around.
 

tuxfool

Banned
The Pascal equivalent of GTX 680 (GK104) and GTX 980 (GM204) certainly won't need 17 Billion transistors. That number is obviously for the flagship GPU, 'Big Pascal'. unofficially referred to as GP100. It's totally reasonable to expect a range of Pascal GPU sizes, serving a full range of cards, as Nvidia always has. That's a given.

Nvidia may have taped out Big Pascal first but I'd expect them to launch cards using a mid-sized Pascal GPU first, as they did with both Kepler (680, march 2012) and second gen Maxwell (980, 970, fall 2014). Big Pascal will probably be reserved for Quadro, the next Titan and the successor to 980 Ti. However, just because Nvidia has done similar things for 2 gens in row, doesn't mean they're locked to do things the same a 3rd time around.

They almost have to. If they plan on launching big Pascal with HBM2, they're going to have to wait for those to be available, which is Q2 at the earliest.They then have to integrate the HBM dies into the interposer etc. You're basically looking at HBM2 equipped products to appear Q3.
 

Pagusas

Elden Member
I've got 1k saved to drop on the best card Nvidia releases in 2016, heres hoping it will last a long time! I'm so ready for 4k gaming
 

Damerman

Member
Dayum.

Got myself a 970 earlier this year, I'll grab a Pascal in 2017 when they're cheaper
This is the best route... Hopefully by then FF XV will be out for PC.

This sounds like its going to be a huge leap. It definitely sounds like they are pushing for a 4k standard. This is good because i settled for a 1080p monitor and a 970, skipping out on 1440p on purpose because i only thought a jump directly to 4k would be worth it.
 

Damerman

Member
Why would they ever cram 17 billion transistors on a chip even if they could? This article doesn't make a lot of sense.

It's a much, much better business strategy to make incremental improvements (maybe 10%/20% improvements tick over tock) and milk it dry. You can go from 8 bil to 10 to 12 to 14 to 18, and you have 4 generations of products to sell rather than 1. Also prevents the consumer from expecting too much because you gave them too good a deal for Pascal.
Whose fucking side are you on?
 

dr_rus

Member
You likely quite wrong if only because they skipped multiple in between processes rather then just reach to the next new node. True 28 nm to 16 nm scaling would make chips nearly 1/4th the size. Power doesn't scale as well anymore but should be enough to give them significant headroom as well. Though I know different foundries have been playing with the numbers so I'm not sure if it's true 16nm anymore at all. There is always the option that they take the destroyer of worlds chips and make them enterprise only for huge margins though if AMD doesn't come out swinging as well.

Too bad that 16FF+ is actually 20nm physically and from a size savings point it's just the next step from 28nm.

Everyone here seen to also constantly forget about prices of the new nodes - even with a decent size savings it's rather unlikely that a chip of the size of GM200 will cost the same to produce on 16FF+ as on 28HP. I would say that it will cost several times more during the first year or so. Thus a smaller chip at a higher price is the most possible outcome.

There was a reason why we got 680 and 7970 first at their respective price points. This reason is even more pronounced with 16FF+. Everyone should keep their expectations in check.
 
The 17 Billion transistors on the Pascal GPU are twice the transistors found on the GM200 Maxwell and the Fiji XT GPU core which is literally insane.

It's literally insane? Sounds scary, are there so many transistors it's going to set my house on fire or something? So the designer was insane to think it was a good idea?
 
It's literally insane? Sounds scary, are there so many transistors it's going to set my house on fire or something? So the designer was insane to think it was a good idea?

143445ee35.png
 

AmyS

Member
Too bad that 16FF+ is actually 20nm physically and from a size savings point it's just the next step from 28nm.

Has it been said anywhere what 10nm FF+ might be, physically: 20nm, 16nm, 14nm or a true 10nm ?

I know there's TSMC, Samsung and GF and they might use different method, but...
 

AlStrong

Member
Has it been said anywhere what 10nm FF+ might be, physically: 20nm, 16nm, 14nm or a true 10nm ?

Well, physically*, it'll be a proper transition. The issue with planar 20nm was that the power & performance characteristics didn't scale as expected.


* node names have been marketing speak at least since 40/45nm anyway
 

grkazan12

Member
So, when is this expected to come out now? I'm waiting on a 980 TI order from Amazon because it's not in stock currently, but the thought of Pascal on the horizon really dampens the mood.

Should I just use the 980 TI till pascal comes out then sell it?

Will the 980 TI hold it's value till then?
 

dr_rus

Member
Has it been said anywhere what 10nm FF+ might be, physically: 20nm, 16nm, 14nm or a true 10nm ?

I know there's TSMC, Samsung and GF and they might use different method, but...

No idea. With how often things are changing around TSMCs nodes I doubt that anyone has. Theoretically 10nm should be a "proper" 10nm as it retains the same FINFET tech used to create "16nm" out of their 20nm.
 

Xyber

Member
So, when is this expected to come out now? I'm waiting on a 980 TI order from Amazon because it's not in stock currently, but the thought of Pascal on the horizon really dampens the mood.

Should I just use the 980 TI till pascal comes out then sell it?

Will the 980 TI hold it's value till then?

It will of course go down a bit, but if you need to performance now you should get it. It's still ways off until we get these new cards.
 

AmyS

Member
AMD readies three new GPUs: Greenland, Baffin and Ellesmere

AMD&#8217;s code-named &#8220;Arctic Islands&#8221; family of products will include three brand-new chips &#8211; &#8220;Greenland&#8221;, &#8220;Baffin&#8221; and &#8220;Ellesmere&#8221; &#8211; a source with knowledge of AMD plans said. The &#8220;Greenland&#8221; will be the new flagship offering for performance enthusiasts, whereas the &#8220;Baffin&#8221; and &#8220;Ellesmere&#8221; will target other market segments, such as high-end and mainstream. It is unclear whether the &#8220;Arctic Islands&#8221; family will take in any existing products, but it is possible that AMD may address certain markets with previous-gen products.

&#8220;Greenland&#8221; will be AMD&#8217;s first graphics processing unit based on the all-new micro-architecture, which development began a little more than two years ago. While the architecture is currently known as another iteration of GCN, the new ISA [instruction set architecture] will be so considerably different compared to the existing GCN that it has every right to be called &#8220;post-GCN&#8221;, the source said. It is likely that the &#8220;Greenland&#8221; will retain layout of the contemporary AMD Radeon graphics processing units, but there will be significant changes in on the deeper level.

The only official thing currently known about the new architecture, which Mark Papermaster, chief technology officer of AMD, calls the next iteration of GCN, is that it is projected to be two times more energy efficient compared to the current GCN. Essentially, this means means major performance enhancements on the ISA level. Thanks to the fact that the &#8220;Greenland&#8221; graphics processing unit will be made using either 14nm or 16nm FinFET process technology, expect it to feature considerably larger number of stream processors than &#8220;Fiji&#8221;.

The &#8220;Greenland&#8221; graphics processor will rely on the second-generation high-bandwidth memory (HBM), so expect ultra-high-end graphics cards and professional solutions with up to 32GB of DRAM onboard with bandwidth of up to 1TB/s. Consumer-class &#8220;Greenland&#8221;-based products will likely come with 8GB &#8211; 16GB of memory. Due to usage of HBM, expect the &#8220;Greenland&#8221; chip and upcoming graphics cards on its base to resemble the currently available AMD Radeon R9 Fury-series adapters.

The number of transistors inside the &#8220;Greenland&#8221; as well as its die size are unknown. Since 14nm/16nm FinFET manufacturing technologies have considerably (up to 90 per cent) higher transistor density than contemporary TSMC&#8217;s 28nm fabrication process, it is logical to expect that the new flagship product will feature 15 &#8211; 18 billion of elements if it retains around 600mm² die size from the &#8220;Fiji&#8221;.

It is believed that AMD has already taped-out its &#8220;Greenland&#8221; graphics processing unit and is about to get the first silicon in the coming weeks.

KitGuru Says: As it traditionally happens, everything looks very good on paper. If AMD manages to release three brand new chips within a reasonable amount of time in 2016 and the chips will be competitive against Nvidia&#8217;s, then it has all the chances to win back market share. However, one should keep in mind that the &#8220;next GCN&#8221; or the &#8220;post-GCN&#8221; will not compete against Nvidia&#8217;s &#8220;Maxwell&#8221;, but will have to compete against Nvidia&#8217;s &#8220;Pascal&#8221;, which promises to be very powerful. As a result, expect 2016 is going to be an interesting year for the GPUs

http://www.kitguru.net/components/g...gpus-for-2016-greenland-baffin-and-ellesmere/
 

dr_rus

Member
the all-new micro-architecture
the next iteration of GCN
Riiiight.

it is logical to expect that the new flagship product will feature 15 – 18 billion of elements if it retains around 600mm² die size from the “Fiji”.
Is it logical to expect AMD to start the next generation with a top end chip which will be as big is the production tech allow it to be?
 

jaypah

Member
I'm buying new GPU in 2016, and most likely AMD will give me best price/performance.

Bring it read team. Make me happy.

I was all set to switch back to team green (I was due for a flip-flop anyway) but I guess I'll wait and see what AMD is doing. It's for VR so I have a little time and my 290x is still kicking.
 

AmyS

Member
Taking a wild guess here, whatever AMD's architecture after Arctic Islands / R-400 is, which is probably in R&D now and intended for PC in 2018 (R-500 series) could very well possibly be the basis for the GPUs in XB4 & PS5 APU by 2020.
 

dr_rus

Member
pascal_02.jpg


This basically proves that Pascal is Maxwell on 16nm + HBM2 + some compute stuff that was cut from Maxwell due to 28nm.
 

belmonkey

Member
I'm hoping there's going to be yet another jump in GPU clock speeds, because Maxwell really got a lot out of GPUs with high clock speeds that otherwise looked like they would perform worse due to their other specs. I wanna hit 2 GHz core on air.
 

Hasney

Member
pascal_02.jpg


This basically proves that Pascal is Maxwell on 16nm + HBM2 + some compute stuff that was cut from Maxwell due to 28nm.

So NVLINK is coming with Pascal as well. I assume if it's up there, it will appear in some high end motherboards.

That's when I'm building my 4k/60 PC.

EDIT: Actually, do they use the same chip names for their pro cards?
 

Kezen

Banned
No idea if they chose Samsung 14nm or TSMC FF+. Also whether maintaining GCN is holding them back or not.

TSMC will manufacture their chips, not Samsung. Obviously we don't know much at all about the new rounds of GPUs but if async compute is going to be as big as some people claim then it's a bit short-sighted on Nvidia's part not to include something akin to AMD's ACEs. Or perhaps async compute can still be fully supported with their Pascal setup, no idea for now.
 

Renekton

Member
TSMC will manufacture their chips, not Samsung. Obviously we don't know much at all about the new rounds of GPUs but if async compute is going to be as big as some people claim then it's a bit short-sighted on Nvidia's part not to include something akin to AMD's ACEs. Or perhaps async compute can still be fully supported with their Pascal setup, no idea for now.
Oops sorry I meant AMD choosing a foundry.

Nvidia has chosen TSMC 16FF+.
 

Marlenus

Member
This is what happens when a node is skipped. Not that surprising really to be honest but I guess the date assumes no respins or other issues with the die, which on a new node using new tech feels a bit optimistic but still doable.

Being stuck with 28nm has really stagnated GPU performance so I am glad we they are finally moving on.
 

Renekton

Member
This is what happens when a node is skipped. Not that surprising really to be honest but I guess the date assumes no respins or other issues with the die, which on a new node using new tech feels a bit optimistic but still doable.

Being stuck with 28nm has really stagnated GPU performance so I am glad we they are finally moving on.
Also optimistic now that Samsung has a competing foundry and smartphone uptake slowed. Two years ago it felt like TSMC was the only viable foundry yet preferred mobile business.
 

Applecot

Member
They will safely do that once they finished curbstomping AMD into bankruptcy.
Finished? You mean that's not the current state of play?

Being stuck with 28nm has really stagnated GPU performance so I am glad we they are finally moving on.
Maxwell was a pretty significant performance enhancement considering it was still on 28nm. By all reports there's nothing to suggest Pascal is going to be revolutionary; just a shrunken down Maxwell.
 

Marlenus

Member
Maxwell was a pretty significant performance enhancement considering it was still on 28nm. By all reports there's nothing to suggest Pascal is going to be revolutionary; just a shrunken down Maxwell.

As good as Maxwell is the improvements are nothing compared to what a die shrink would have provided. Instead of a 7Tflop Titan X / 980Ti a 20nm Maxwell would likely have been a 10-11 Tflop Titan X / 980Ti which is a significant improvement. Along with the fact the increased transistor density would have enabled more ROPS/TMUs and other parts to have had their numbers boosted as well.
 
Top Bottom