At the heart of the Core i7-6700K and Core i5-6600K quad-core processors, which made their debut at Gamescom earlier this month, is Intel's swanky new "Skylake-D" silicon, built on its new 14 nanometer silicon fab process. Intel released technical documents that give us a peek into the die layout of this chip. To begin with, the Skylake silicon is tiny, compared to its 22 nm predecessor, the Haswell-D (i7-4770K, i5-4670K, etc).What also sets this chip apart from its predecessors, going all the way back to "Lynnfield" (and perhaps even "Nehalem,") is that it's a "square" die. The CPU component, made up of four cores based on the "Skylake" micro-architecture, is split into rows of two cores each, sitting across the chip's L3 cache. This is a departure from older layouts, in which a single file of four cores lined one side of the L3 cache. The integrated GPU, Intel's Gen9 iGPU core, takes up nearly as much die area as the CPU component. The uncore component (system agent, IMC, I/O, etc.) takes up the rest of the die. The integrated Gen9 iGPU features 24 execution units (EUs), spread across three EU-subslices of 8 EUs, each. This GPU supports DirectX 12 (feature level 12_1). We'll get you finer micro-architecture details very soon.

82 Comments on Intel "Skylake" Die Layout Detailed

1 to 25 of 82 Go to Page 1234 PreviousNext

#1 GhostRyder

Wow, that's a lot of room dedicated just to the GPU. Bet those Xeons look great without that space taken up so they can cram cores! Posted on Aug 18th 2015, 18:30 Reply

#2 ensabrenoir

.............In before the lame comments....... ITS ALL ABOUT THE GPU!!!!!!!!!!!! Posted on Aug 18th 2015, 19:07 Reply

#3 FordGT90Concept

"I go fast!1!11!1!" I intend to disable the GPU via UEFI. Posted on Aug 18th 2015, 19:12 Reply

#4 ZenZimZaliben

btarunr Look at the size of that GPU. It takes up more then 1/3 of the chip....I do not understand why a K class chip even has a freaking IGP. 90% of the people interested in K series run dedicated gpu. SO basically 35% of the cost of the chip is going towards an IGP that will never be used. I would much rather pay for 35% more cores. With the IGP gone it would be easy to fit in at least 2 more cores with out expanding the die. Look at the size of that GPU. It takes up more then 1/3 of the chip....I do not understand why a K class chip even has a freaking IGP. 90% of the people interested in K series run dedicated gpu. SO basically 35% of the cost of the chip is going towards an IGP that will never be used. I would much rather pay for 35% more cores. With the IGP gone it would be easy to fit in at least 2 more cores with out expanding the die. Posted on Aug 18th 2015, 19:21 Reply

#5 Sony Xperia S

That's unbelievable - 50% for graphics and 50% of the shared die area for the real stuff.



Of course, this sucks, and it sucks even worse when you know that Intel does nothing at least to try to develop software environment to unleash all those wasted transistors.

40% of the die - to be not used. Posted on Aug 18th 2015, 19:25 Reply

#6 n-ster

Is there any downside to having a disabled IGP there in terms of thermal/power usage? Posted on Aug 18th 2015, 19:51 Reply

#7 Hades

n-ster Is there any downside to having a disabled IGP there in terms of thermal/power usage? You just save power. That's it :) You just save power. That's it :) Posted on Aug 18th 2015, 20:21 Reply

#8 n-ster

Hades You just save power. That's it :) I meant is there a negative effect to have that dead-weight IGP there versus a chip that would completely cut out the IGP? I meant is there a negative effect to have that dead-weight IGP there versus a chip that would completely cut out the IGP? Posted on Aug 18th 2015, 20:42 Reply

#9 btarunr

Editor & Senior Moderator The IGP is this big not because Intel hopes you'll play Battlefront with it at 1080p. It's because Intel has to cope with the new wave of >1080p displays, such as 4K and 5K (particularly with high-res notebooks). This IGP has just enough juice to play 4K 60Hz video without stuttering. The API support is up-to-date just so you'll probably do DX12 asymmetric multi-GPU using discrete GPUs in the future, where your display is plugged into the IGP. Posted on Aug 18th 2015, 20:53 Reply

#10 Patriot

n-ster I meant is there a negative effect to have that dead-weight IGP there versus a chip that would completely cut out the IGP? You can use quicksync to accelerate certain tasks that may not be coded in such a way to use your discrete card. You can use quicksync to accelerate certain tasks that may not be coded in such a way to use your discrete card. Posted on Aug 18th 2015, 21:08 Reply

#11 Disparia

Sony Xperia S That's unbelievable - 50% for graphics and 50% of the shared die area for the real stuff.



Of course, this sucks, and it sucks even worse when you know that Intel does nothing at least to try to develop software environment to unleash all those wasted transistors.

40% of the die - to be not used. I know!



I mean, I would know if I didn't know that Intel had provided developer guides since 2011:

software.intel.com/en-us/articles/intel-graphics-developers-guides



Or if I had missed when Intel put support behind OpenCL back in 2014:

streamcomputing.eu/blog/2014-03-25/intel-promotes-opencl-as-the-heterogeneous-compute-solution/



Or didn't attend any of the numerous Intel-sponsored development events every year:

software.intel.com/en-us/ I know!I mean, I would know if I didn't know that Intel had provided developer guides since 2011:Or if I had missed when Intel put support behind OpenCL back in 2014:Or didn't attend any of the numerous Intel-sponsored development events every year: Posted on Aug 18th 2015, 21:40 Reply

#12 peche

Thermaltake fanboy scumbag intel .... Posted on Aug 18th 2015, 21:45 Reply

#13 ZenZimZaliben

Jizzler I know!



I mean, I would know if I didn't know that Intel had provided developer guides since 2011:

software.intel.com/en-us/articles/intel-graphics-developers-guides



Or if I had missed when Intel put support behind OpenCL back in 2014:

streamcomputing.eu/blog/2014-03-25/intel-promotes-opencl-as-the-heterogeneous-compute-solution/



Or didn't attend any of the numerous Intel-sponsored development events every year:

software.intel.com/en-us/ I see what you did there. I see what you did there. Posted on Aug 18th 2015, 21:49 Reply

#14 ppn

Waits for 8-core by intel Posted on Aug 18th 2015, 22:18 Reply

#15 R-T-B

It's no secret Intel wants to go full SOC.



The only thing really missing is decent graphics. This is filling that gap. No, enthusiasts don't like it, but they also make up like 10% of the market if I'm being generous... Posted on Aug 18th 2015, 22:27 Reply

#16 Slizzo

peche scumbag intel .... I'm sorry, Scumbag? For what exactly? Improving performance for users while lowering TDP? I'm sorry, Scumbag? For what exactly? Improving performance for users while lowering TDP? Posted on Aug 18th 2015, 22:30 Reply

#17 ZenZimZaliben

Slizzo I'm sorry, Scumbag? For what exactly? Improving performance for users while lowering TDP? I don't want a lower TDP with 10% performance increase. I want HUGE Performance gains and TDP doesn't matter at all. It's an i7 K chip. It should not be a power sipper and it also shouldn't have an IGP. However this is the first release at 14nm and hopefully the next release will ditch the IGP and go for more cores. I don't want a lower TDP with 10% performance increase. I want HUGE Performance gains and TDP doesn't matter at all. It's an i7 K chip. It should not be a power sipper and it also shouldn't have an IGP. However this is the first release at 14nm and hopefully the next release will ditch the IGP and go for more cores. Posted on Aug 18th 2015, 22:39 Reply

#18 ERazer

still not worth upgrading, intel made sandy bridge such a badass Posted on Aug 18th 2015, 22:50 Reply

#19 peche

Thermaltake fanboy ERazer still not worth upgrading, intel made sandy bridge such a badass no sandies were soldered chips... this crappy new ones still use crappy paste... no sandies were soldered chips... this crappy new ones still use crappy paste... Posted on Aug 18th 2015, 22:51 Reply

#20 Joss

You can always go for AMD... oops, I forgot they don't release a new chip for 3 years.



Joke apart that's exactly the problem: the Red team is not putting up a fight and Intel can do as they please.

Problem is, considering what AMD did with Fury I'm not hoping much for their next FX series (if there is one). Posted on Aug 18th 2015, 23:14 Reply

#21 Yorgos

Hades You just save power. That's it :) Actually you don't.

Using a GPU while browsing, fapping, e.t.c. consumes much more power than having the iGPU enabled and turning on and off the dGPU for the heavy staff.

Also, your dGPU lives longer, unless you use nVidia and nVidia decides when to cripple your h/w. Actually you don't.Using a GPU while browsing, fapping, e.t.c. consumes much more power than having the iGPU enabled and turning on and off the dGPU for the heavy staff.Also, your dGPU lives longer, unless you use nVidia and nVidia decides when to cripple your h/w. Posted on Aug 18th 2015, 23:22 Reply

#22 deemon

Slizzo I'm sorry, Scumbag? For what exactly? Improving performance for users while lowering TDP? Thats the problem... intel does not improve the performance... not by much. Ever anymore. Thats the problem... intel does not improve the performance... not by much. Ever anymore. Posted on Aug 18th 2015, 23:59 Reply

#23 tabascosauz

I'm beating a dead horse here, but the intent of HEDT is to satisfy the exact conditions that most of you seem to expect from a top of the product stack mainstream i5/i7.



The complaints won't stop until Intel gets rid of the GPU, and they really won't stop because Intel is not going to take that GPU off. This die is going to power the other desktop i5s and i7s and those are parts intended for powering a 1080P monitor at work without the assistance of a dGPU. Not throwing money into the water for a pointless GPU-less Skylake design that basically gets them no $$$ at all is actually a pretty smart business plan, believe it or not.



Not a lot of praise where it's deserved, I'm afraid. I wouldn't be showering Intel with praise 24/7, but I didn't see any positive comments about the 5820K when it was released, only "only 28 PCIe lanes? What if I need the extra 12 to get maximum performance while getting off every day?" Tried to appease the enthusiasts with a 6-core HEDT part below $400, well, I guess that didn't work out very well, did it? The 5820K wasn't even a forced hand; it could very well have been a carbon copy of the 4820K, just on Haswell, as there are 4-core E5 V3s a-plenty to prove that.



Give people something better, and they'll find something better to complain about. Posted on Aug 19th 2015, 0:05 Reply

#24 peche

Thermaltake fanboy deemon Thats the problem... intel does not improve the performance... not by much. Ever anymore. thanks didnt saw the message!

you took my words...

point #2: intel knows pretty much the problem and differences about soldering and thermal paste on their CPU die's.....

but they still use that shitty paste... so? thanks didnt saw the message!you took my words...point #2: intel knows pretty much the problem and differences about soldering and thermal paste on their CPU die's.....but they still use that shitty paste... so? Posted on Aug 19th 2015, 0:16 Reply

#25 Uplink10

FordGT90Concept I intend to disable the GPU via UEFI. Why? ZenZimZaliben Look at the size of that GPU. It takes up more then 1/3 of the chip....I do not understand why a K class chip even has a freaking IGP. 90% of the people interested in K series run dedicated gpu. That. K chips are premium beacuse you get the privilege of overclocking even though in the end performance per dollar is lower than if you bought non-K chip. I imagine these people have enough money for dGPU. btarunr This IGP has just enough juice to play 4K 60Hz video without stuttering. The CPU part can already do that without the iGPU hardware acceleration. Why?That. K chips are premium beacuse you get the privilege of overclocking even though in the end performance per dollar is lower than if you bought non-K chip. I imagine these people have enough money for dGPU.The CPU part can already do that without the iGPU hardware acceleration. Posted on Aug 19th 2015, 0:43 Reply