[HN Gopher] The Intel Split ___________________________________________________________________ The Intel Split Author : feross Score : 226 points Date : 2022-01-18 15:07 UTC (7 hours ago) (HTM) web link (stratechery.com) (TXT) w3m dump (stratechery.com) | cletus wrote: | IMHO this is just Intel hiding rather than solving its problems. | It gives the appearance of management doing something but it's | the wrong thing. | | If we've learned nothing else, the big winners are at the cutting | edge and vertically integrated. Splitting design and fab seems | like a good idea: force fab to compete for design business. But | middle management will take over and create perverse incentives | where neither fab nor design radically improves. | | I've heard tales of Intel being rife with politics, fiefdoms and | empire-building. This is classic middle management run amok in | the absence of real leadership. I honestly think the best thing | Intel could do is fire everyone above the director level and | start over. | | Intel dominated when their fab was cutting edge and no one else | had access to it. Splitting this means if their fab improves then | everyone has access to it. | | There's clearly an organizational problem here but this split | isn't going to solve it. | threatripper wrote: | IMHO it's the same pattern of the time when Intel gave up on | their memory business and entered the CPU business. Citing from | memory: "If you were the new CEO what would you do to save the | company?" - and he immediately knew the answer - "Then let's go | back in the boardroom and do that." | | Judging from that it seems like their fab business doesn't have | any long term future, the outside world just doesn't know it | yet. Now they put on a show until they are ready to fully | convert to other fabs. After all, the current fabs are still | cranking out chips that make them billions each month. | JumpCrisscross wrote: | > _their fab business doesn 't have any long term future_ | | It has a fine future, just not in a 60%+ margin business. | Intel should bite the bullet and spin out its fabs. Let it | compete and grow as a separate company from Intel's design | units. Stratchery is spot on, as he's been about Intel from | day one. | notswiley2 wrote: | Isn't this kind of thing part of what killed the PalmPilot? | bluedino wrote: | Imagine what could have been if Intel wouldn't have sold XScale | back in 2006 and kept focusing on ARM | stefan_ wrote: | What exactly would have been? There is no inherent superiority | in the ARM architecture. Hell, 2006 ARM is a trash | architecture. They just happened to be the only vendor that | cared about that niche, and then the niche stopped being niche. | MangoCoffee wrote: | ARM is not end all and be all. x86 still have a big market | share | neogodless wrote: | That's one way to put it! I'm not finding any comprehensive | sources, but here's an example: | | https://www.theregister.com/2021/11/12/apple_arm_m1_intel_x8. | .. | | > Arm's market share in PC chips was about eight per cent | during Q3 this year, climbing steadily from seven per cent in | Q2, and up from only two per cent in Q3 2020, before Arm- | compatible M1 Macs went on sale. | | ARM is very much on the rise, but also still in the single | digits, leaving plenty for x86. | | https://futurecio.tech/intel-losing-share-to-amd-and-arm/ | | > 5% of the servers shipped in the third quarter of 2021 had | an Arm CPU | carlycue wrote: | Intel cannot and will not catch up. The arrival of the M1 chip is | analogous to the arrival of the monolith in 2001 A Space Odyssey. | It's efficiency is frankly alien technology. | pjmlp wrote: | M1 only runs on Apple hardware. Hardly matters to Intel | customers. | foobiekr wrote: | That's true to an extent but no one lives in isolation. | Losing Apple was a pretty massive blow. | | Still the real blow will be if someone successfully enters | the server business. Graviton* are interesting but that's not | a broad threat yet. | sydbarrett74 wrote: | I think it'll be another huge blow if Windows users with | laptops move almost entirely to ARM over the next 3-5 | years. While Intel may have an absolute performance | advantage in the desktop and server arena, most Windows | users these days are using laptops, where ARM's energy | efficiency matters more. | malfist wrote: | Why do you say that? They're already responding to ARM with | their big.Bigger architecture in their latest generation. | | I'm sure people said the same thing about Intel when AMD | introduced 64 bit procs or Sun introduced multicore processors. | Intel has adapted and lead the field many times after being | overtaken. No reason to expect them not to do the same here, or | at least compete. | klelatti wrote: | Not sure 'already' is appropriate here. big.LITTLE was | announced in 2011. Taking 11 years to copy a competitor's | successful feature must be a record of some sort. | Kon-Peki wrote: | Not sure you understand the point of big.LITTLE. It's about | dealing with power and thermal constraints in the quest for | more performance [1]. So was the transition to multicore | processors, BTW [2]. These are the things hardware | companies do when they have no other option. And as [2] | points out, we software folks _still_ don 't have a good | way to deal with it. | | [1] https://armkeil.blob.core.windows.net/developer/Files/p | df/wh... | | [2] https://www2.eecs.berkeley.edu/Pubs/TechRpts/2006/EECS- | 2006-... | | Alder Lake is about dealing with power and thermal | constraints. Intel has finally pushed performance so far | that they need to do this. The chips are benchmarking so | well _because_ of this move, not in spite of it. | klelatti wrote: | I do know what big.LITTLE is about thanks. No idea why | you'd think otherwise from my post. | | Of course Alder Lake is benchmarking well because of it. | | My point was 'already' makes it sound like Intel is | rapidly adopting this technology - not 11 years after | Arm. | Cthulhu_ wrote: | > It's efficiency is frankly alien technology. | | That's probably an exaggeration, and this exaltation of Apple | is a Bad Thing. The M1 had good designers - two of which have | moved on already, one to Microsoft [0], and the lead designer | to Intel [1] only days later. | | [0] https://9to5mac.com/2022/01/12/apple-silicon-microsoft- | poach... | | [1] https://9to5mac.com/2022/01/06/apple-engineering- | director-m1... | atty wrote: | I love my M1, but it is most certainly not alien technology | that can't be replicated, nor does it imply that other | foundries/chip designers can't catch up. In many respects AMD | is already very close (on a previous node, no less), and Intels | 12th Gen, while very power hungry, is very good in terms of | total performance (obviously not competitive on perf/watt, but | it was never going to be, because of the node difference). | tomjen3 wrote: | You can get better performance than an M1, but you are not | getting anything else with the performance and battery life. | Except maybe the M2. | | If you want an actual ultra portable, you are getting a Mac. | kcb wrote: | If you want a Mac. | AnimalMuppet wrote: | Well, if "ultra portable" matters more than "Mac vs non- | Mac", then get a Mac (at the moment). If you want "not | Mac" more than you want "ultra portable", then don't get | a Mac. | criddell wrote: | Are you saying that AMD is very close in perf/watt? If so, | that's pretty cool. | atty wrote: | Sorry I should have been clearer. As far as I am aware, the | M1 is still better at perf/watt (again, comparing TSMC 5 to | 7, so it's not unexpected) - for laptops, the M1 also has | an extremely low idle power draw that I don't think anyone | else can match right now, which significantly improves | battery life. I meant comparable single-threaded | performance in a laptop form factor. | donkarma wrote: | if my understanding is correct then M1 was a one time boost due | to the massive decoder and on chip RAM, it's not going to be | constantly getting faster | selectodude wrote: | The M1 also has old cores from the A14. The A15 isn't hugely | faster than the A14 but it clocks faster and has other | efficiency tweaks. Clock for clock, the M1 is _slower_ than | the iPhone 13 Pro. | [deleted] | foobiekr wrote: | That's basically it. Apple went big because the target | process allowed it. To someone out of the ASIC industry, it | is a set of interestingly scaled but pedestrian design | choices coupled with some exaggeration (realizable memory | bandwidth per core on the Max is a fraction of what you'd | expect based on Apple's aggregate memory bandwidth claim for | the processor as a whole) and a very serious investment in | performance/watt (legacy of the phone use case). | | The Max has barely any improvement core-wise than the first | M1. It's going to be interesting to see what the real next | generation looks like. | GeekyBear wrote: | > The Max has barely any improvement core-wise than the | first M1. | | Isn't that what you would expect in two chips that use the | same core design? | | There is more memory bandwidth to the Max, and the system | level cache is larger, so there are differences outside of | the core, but the core itself didn't change. | GeekyBear wrote: | >one time boost | | That's not what you would expect if you look at the graph of | six years of year-over-year SPEC performance gains on iPhone | cores. Their history shows a pretty reliable 20% gain per | year. | | >Whilst in the past 5 years Intel has managed to increase | their best single-thread performance by about 28%, Apple has | managed to improve their designs by 198%, or 2.98x (let's | call it 3x) the performance of the Apple A9 of late 2015. | | https://www.anandtech.com/show/16226/apple- | silicon-m1-a14-de... | stingtao wrote: | If Intel split the manufacturing unit, it would not be | competitive unless it invested lots more in advanced new | processes. | rkagerer wrote: | The company's willingness to tackle tooling to multiply the | effectiveness of their employees was a key factor in their | success: | | _This incredible growth rate could not be achieved by hiring an | exponentially-growing number of design engineers. It was | fulfilled by adopting new design methodologies and by introducing | innovative design automation software at every processor | generation. These methodologies and tools always applied | principles of raising design abstraction, becoming increasingly | precise in terms of circuit and parasitic modeling while | simultaneously using ever-increasing levels of hierarchy, | regularity, and automatic synthesis. | | As a rule, whenever a task became too painful to perform using | the old methods, a new method and associated tool were conceived | for solving the problem. This way, tools and design practices | were evolving, always addressing the most labor-intensive task at | hand. Naturally, the evolution of tools occurred bottom-up, from | layout tools to circuit, logic, and architecture. Typically, at | each abstraction level the verification problem was most painful, | hence it was addressed first. The synthesis problem at that level | was addressed much later._ | adamc wrote: | Reading this, I had the uneasy feeling that we might be seeing a | company chase a past that cannot be its future. | blinkingled wrote: | I don't think Intel's manufacturing problem is purely that of | incentive which will be fixed by a split. That's massive | oversimplification. | | The approach Intel is taking - outsource cutting edge products to | TSMC while continuing to invest in their fabs making their lower | end stuff and other people's stuff like automotive chips in-house | is the best strategy to buy some time to advance their fabs while | letting them earn money to support R&D investments. | | It's a huge problem and nobody except TSMC has succeeded at it. | Besides there is years of lack of focus, incentives and interest | in specialized education and manufacturing processes that'll take | time for Intel to fix. Meanwhile they will be competitive in | consumer markets by going TSMC 3nm and continue to improve on the | side by taking on outside fab orders. Seems reasonable to me. | ryan93 wrote: | They had profit of 20B last year. The bean counters are the | reason they wont invest in cutting edge fabs not a lack of | capital. | blinkingled wrote: | Sure, but does it sound more challenging to invest in a spin | off company than an integrated arm that will deliver revenue | from inside and outside business? I am sure the shareholders | would be more agreeable to the latter. Especially when there | is enough demand for chip making to be almost certain that it | will not be a huge money loser. | | To invest Intel's money in a spinoff fab company while losing | strategic control over its direction and have little to gain | from its success doesn't feel all that attractive to me. | dannyw wrote: | Intel is already competitive on 10nm (Intel 7) with Alder Lake | on desktop against AMD; I'm frankly impressed to see what | Intel's designs on TSMC 3nm will do. | saberience wrote: | Only competitive if you don't look at energy usage, Intel | relies on cranking the energy usage in order to get their | chips almost to compete with AMD. | khyryk wrote: | Alder Lake is more power efficient under low to moderate | loads, which includes all gaming. It's only less efficient | under loads pushing it at 100% on all cores when it's on a | high power limit. | wtallis wrote: | Ironically, Alder Lake is only efficient when it doesn't | have to use the E-cores. | khyryk wrote: | Intel's (literally) doubling down on the Atom cores in | Raptor Lake so they'll have to get it right for them to | have a Zen 2-esque progression. CPU advances beyond the | old +5% per generation are pretty exciting. | api wrote: | That may depend substantially on process node. 3nm would | use less power. | phkahler wrote: | Sure, but it's not Intel 3nm. AMD is moving to TSMC 5nm | and will go to 3nm in the future as well. It really | bothers me when either side compares tomorrows products | to competitors todays products. | the_duke wrote: | Sure, but so would AMD chips, which aren't even on 5nm | yet. | GeekyBear wrote: | I'm suspicious that reliability issues stemming from an | extreme power draw and level of heat generated when you run | AVX-512 instructions was the reason for those instructions | being disabled recently. | | >One of the big takeaways from our initial Core i7-11700K | review was the power consumption under AVX-512 modes, as | well as the high temperatures. Even with the latest | microcode updates, both of our Core i9 parts draw lots of | power. | | The Core i9-11900K in our test peaks up to 296 W, showing | temperatures of 104oC, before coming back down to ~230 W | and dropping to 4.5 GHz. The Core i7-11700K is still | showing 278 W in our ASUS board, tempeartures of 103oC, and | after the initial spike we see 4.4 GHz at the same ~230 W. | | There are a number of ways to report CPU temperature. We | can either take the instantaneous value of a singular spot | of the silicon while it's currently going through a high- | current density event, like compute, or we can consider the | CPU as a whole with all of its thermal sensors. While the | overall CPU might accept operating temperatures of 105oC, | individual elements of the core might actually reach 125oC | instantaneously. So what is the correct value, and what is | safe? | | https://www.anandtech.com/show/16495/intel-rocket- | lake-14nm-... | celrod wrote: | In that particular benchmark (3d particle movement), the | 11700K performed about 4x better than the AMD 5900X. | Performance/watt clearly wasn't suffering. | | Perhaps it could downlock more to address the wattage | while still coming well ahead in terms of performance, | although some older CPUs doing this gave AVX512 a bad | reputation. | | Few workloads are as well optimized to take advantage of | AVX512 as the 3d particle movement benchmark, so both the | increases in performance and wattage seen in that | benchmark are atypical. If they were typical, then AVX512 | would be much more popular. | | FWIW, I'm a big fan of wide CPU vector instructions. The | real reason it was disabled is probably to push people | like me to buy Saphire Rapids, which I would've favored | anyway for the extra FMA unit. Although I'll also be | waiting to see if Zen4 brings AVX512, which some rumors | have claimed (and others have contradicted). | GeekyBear wrote: | One of the failure modes for chips as they move to | smaller process nodes is electromigration. | | >Electromigration is the movement of atoms based on the | flow of current through a material. If the current | density is high enough, the heat dissipated within the | material will repeatedly break atoms from the structure | and move them. This will create both 'vacancies' and | 'deposits'. The vacancies can grow and eventually break | circuit connections resulting in open-circuits, while the | deposits can grow and eventually close circuit | connections resulting in short-circuit. | | Chips that get very hot are expected to be the first to | show this sort of failure. | | >In Black's equation, which is used to compute the mean | time to failure of metal lines, the temperature of the | conductor appears in the exponent | | https://www.synopsys.com/glossary/what-is- | electromigration.h... | PixyMisa wrote: | The high-end Alder Lake parts - particularly the 12900K - | are terrible energy hogs. But as you move down the stack, | efficiency improves a lot while performance remains very | competitive with AMD parts in the same price bracket. | jeffbee wrote: | My 12700K has core power usage under 1W and package power | under 4W when it's just sitting there, which is does a | heck of a lot of. When it's running flat out compiling | C++ projects on all performance cores, package power is | ~90W. Single-threaded performance is much better than any | Ryzen and even beats my M1 Pro. I'm not really seeing the | energy argument for desktop/workstation usage. For | datacenter TCO AMD is probably still the champ. | monocasa wrote: | I guess I weight datacenter workloads much higher for | perf/watt because the higher margins there is what funds | next gen's R&D. Cranking up power to get perf under load | is a move that cuts off funding streams in one of the | most capital intensive industries. | jeffbee wrote: | Intel still sells 85% of the server market including 75% | of the megascale cloud market, so at this point it does | not appear to me that Intel has been strategically | wounded. I'm sure they have more than enough cash to fund | R&D at this time. | monocasa wrote: | It was 98% of the datacenter market in 2017, and | apparently the rate of decrease is accelerating, and | that's even before taking into account that the | datacenter chip equivalent to your 12700k doesn't come | out until later this year and that's where you'd expect | to see the real fall off in DC marketshare. | | That money can dry up very quickly. | jeffbee wrote: | By no means do I imagine that Intel has everything right | where they want it. Clearly, they'd prefer to be the | "machine of the day" at Google, instead of AMD having | that honor. But, it's also not the first time they've | been in this position. I would argue they were in much | more serious trouble in 2008 or so, when it was Itanium | and NetBurst vs. Opteron, and everyone bought Opteron. | monocasa wrote: | Intel still had leading edge node supremacy at that | point, and everyone go caught with their pants down with | the end of Dennard scaling. AMD simply lucked out that | they weren't designing a core to hit 10Ghz in the first | place (because what would become GloFo didn't have the | node chops to even pretend that they could hit that in | the first place even ignoring Dennard scaling issues). | AMD therefore didn't have to pull slower mobile core | designs off of the back burner like Intel had to which | took time to readjust. | | Intel's in a much worse position now. Their process woes | aren't issues that the rest of the industry will hit in a | year or so like the end of Dennard scaling was. | dannyw wrote: | Not really: https://www.reddit.com/r/hardware/comments/qwn1 | j9/core_i9129... | | The holistic efficiency comparison definitely favours Intel | on desktop, as Ryzen has heavy idle power usage, due to its | 12nm I/O die from GloFo. | thereddaikon wrote: | idle power draw on desktop isn't a major concern. When | people talk about power draw in a desktop context its | really an indirect way to measure heat. Managing thermals | at heavy load is the most important consideration. And if | your chip manages to win benchmarks by creating massively | more heat in the process then that is worth noting. | | In laptops idle power draw and heat are far more | important because they effect system endurance. In fact | most gains in mobile device power management are with | lowering idle consumption not load consumption. | marricks wrote: | Clearly Intel has huge issues or they wouldn't be making such | a drastic turn. For some people whatever happens to intel the | picture is always extremely rosey. | cs702 wrote: | In short, Gelsinger and his team are configuring the company so | Intel's internal manufacturing operations _must earn Intel 's | business, in competition against every other fab-as-a-service | provider in the world, including TSMC_. | | The risk is obvious: The company's internal manufacturing | operations can survive only if they figure out how to outperform | competitors. The upside is that all other parts of Intel stand a | good chance of surviving and thriving -- but that remains to be | seen. | | Only one thing is certain: Gelsinger and his team have a lot of | _cojones_. | lmilcin wrote: | As an ex-Intel employee it is interesting to see. | | This seems to be sound long-term though it may cause | significant shrink in the business in next couple of years to | cut lines of business that are not above water. | | From the point of view of infinite game it makes a lot of sense | to diversify and then cut any branch of business that is not | profitable, before it drags entire company. | | While I think it is difficult to predict results, one thing is | sure -- this will force a lot of change in Intel and we would | likely not recognise the company ten years from now. | threatripper wrote: | This means that Intel manufacturing is used where it makes | sense and vice versa isn't used where it doesn't make sense. | | I can see those general outcomes: | | * Intel manufacturing is soon again better than other | manufacturers in all aspects as it has been a long time. All | Intel business goes to Intel manufacturing and things return to | what they were. | | * Intel manufacturing is better in some aspects and continues | to outcompete other manufacturers in some areas. Most Intel | business goes to Intel manufacturing but some special business | goes to other manufacturers. Some other companies use Intel | where they have an advantage. Everybody wins. | | * Intel manufacturing cannot keep up with the market leader. | Most Intel business goes to the leading Fab. Intel | manufacturing cannot catch up as they don't earn enough money | to do so. They start cutting costs and compete downmarket with | lower tier fabs. In the long run Intel becomes fabless. | opportune wrote: | I think there is a pretty clear outcome, Intel will go | fabless for highest end chips and continue as normal with | everything else. They will have many customers (USG) who will | require domestically manufactured chips. But they also need | to compete with AMD, Apple, and friends and this will allow | them to fab out to TSMC while, for lack of a better term, | saving face. | paulpan wrote: | Agreed except there's only 2 outcomes: #1 (win) or #3 (lose). | The chips business is a winner-take-all market where if your | product isn't the best (in efficiency, performance, whatever | metric), then it can't command a price premium nor the | attention of OEMs. Also the core business isn't consumer but | for enterprise/datacenter market where efficiency/performance | is paramount. | | If #1, then everything is rosy and Intel will regain its | dominance from the days of Sandy Bridge. Its chips will be | fabricated internally and they get to enjoy the profits from | vertical integration. | | If #3, then Intel will certainly spin off or sell off its | manufacturing/fabrication business very similar to AMD more | than a decade ago. | | The caveat is that I don't think Intel's problem right now is | simply a manufacturing issue. Sure their Alder Lake is | competitive but it's not superior to AMD's offerings or even | measured against Apple's SOCs. Remember that unlike the last | iteration (Rocket Lake), Alder Lake doesn't suffer from a | mismatch in design vs. manufacturing cadence - it's arrived | as expected. | Spooky23 wrote: | Another possibility is that the market is changing. The age | old Xeon/Core/"binned low end" model may not be viable | anymore as hyperscale cloud providers start to rule and | people don't care about CPUs as a differentiator. Add in | COVID chaos and the answers are unclear. | | Hell I just replaced thousands of older desktops with slower | laptop chips -- nobody noticed. Frankly, I only bought new | devices because of Microsoft's bullshit Windows 11 | requirements. | | The guaranteed cash flow train is slowing down, which makes | setting billions of dollars on fire with high risk fab | processes a big deal. | | Intel had to adapt like this in the early 90s when memory | became a commodity. We'll see if they make it. | xt00 wrote: | Yes, the market is changing -- on all fronts for Intel. | Many people talk about this being a tech problem, really | its more like the logic at companies when you hold a close | to monopoly position -- why invest tons of R&D dollars when | you can just increment more slowly and make the same | money.. So they end up with a loss of tech leadership when | they easily could have had it. The biggest mistake that | Intel appears to have made was their failure to bring up | the new 13nm EUV processes at the same rate as TSMC -- | probably they were hoping it would fail, but it did not. | Now TSMC has done a couple of nodes with EUV and they have | some customers that are spending tons of money right now | like Apple, AMD, nvidia, qualcomm, etc and now TSMC is | flush with cash and have a strong runway in front of them | -- its like TSMC is already in execution mode while Intel | is still trying to get things working. Intel basically | thought, hey lets see how long we can milk our big cloud | customers with $2000 processors and only minor | improvements. Those customers realized they could hire a | chip team themselves and build their own processor for the | amount of money they were paying Intel each year, so they | did that. So they got attacked on the enterprise side by | their own customers making their own chips like Amazon and | Google, the consumer side by AMD just throwing lots of | cores at the problem -- and Apple dumping x86 in their | laptops -- and soon to be others, and they essentially have | nothing in the mobile space. So they went from being | dominant 10 years ago to being 2nd place in multiple of | their core businesses. Clear failure of leadership to | realize they need to maintain technical supremacy otherwise | they will not be able to charge the prices they want to | charge. Now their competitors are shipping chips with 64 | cores and Intel's plan to uber slowly release 16 cores, | then 20 cores, 24 cores, etc over a 8 year period or | whatever their plan was is blown up.. | seanp2k2 wrote: | I'm still bullish on $AMD to break much further into the | enterprise space. So many AWS instances still run on | Xeons. That's slowly changing, especially now with stuff | like Graviton , but I think AMD can go much higher just | on enterprise. | rjzzleep wrote: | > Hell I just replaced thousands of older desktops with | slower laptop chips -- nobody noticed. Frankly, I only | bought new devices because of Microsoft's bullshit Windows | 11 requirements. | | Big enterprises think like that yeah, because they can't | get their crappy Dell Inspirons to new employees anymore. | | Personally I just had to get a Ryzen laptop replacement | temporarily because Lenovo took too long to repair my | thinkpad. And the result was delightful. Better sound, | fantastic compile times, beautiful screen for half the | price of my thinkpad. Next time I'll get a cheap Ryzen and | if it breaks just buy a replacement device instead of | relying on a pro intel enterprise device with crappy | overpriced enterprise service. | Spooky23 wrote: | Agreed - but that's a problem for Intel, as they need | enterprises to lap up those shitty Dell laptops to keep | the throughput going! | | NYC Public Schools buy like 200,000 computers a year. A | decade ago, it was probably 50% higher due to faster | refresh cycles. There are a lot more big dumb enterprise | endpoints than you might think. When I sold computers in | college to commercial customers, the average refresh was | 30-40 months. Now it's closer to 60. | acdha wrote: | > Hell I just replaced thousands of older desktops with | slower laptop chips -- nobody noticed. | | This is certainly plausible if these were mostly Office | users but how confident are you that you'd know if they did | notice? Most of the large IT departments I've interacted | with would post glowing self-evaluations while the users | had plenty of complaints and/or were doing their real work | on personal devices. Depending on how the business is | structured and the relationships, this can be hard to | measure without a fair amount of effort -- I've heard | people say they weren't going to report something because | they didn't think it was worth the effort or they were sure | that the IT people would notice and fix it. | chasil wrote: | The problem that the industry faces is that the economics and | reliability of these chips have been undermined on the recent | nodes. | | Sophie Wilson has said that cost-per-transistor is rising | since we shrank below 22nm, and that a 7nm die cannot run | more than half of its transistors at the same time without | melting. | | Sophie addresses the cost rise at 22:00. | | https://m.youtube.com/watch?v=zX4ZNfvw1cw | | Heat is more of a problem because finfet can't dissipate as | well as planar, and reliability is declining at smaller | nodes. | | "With a planar device, you do not have to bother about self- | heating. There are a lot of ways to dissipate heat with a | planar device, but with finFETs that is not the case. The | heat gets trapped and there are few chances for that heat to | get dissipated." | | https://semiengineering.com/chip-aging-becomes-design- | proble... | | https://news.ycombinator.com/item?id=29889951 | | If I needed a 20-year service life, I would not choose 7nm. | | Edit: I did not realize that ARM1 had fewer transistors | (~25k) than the 8086 (29k), and over ten times less than the | 80386 (275k). Intel should have bought ARM in the 80s; | instead Olivetti got them. | bentcorner wrote: | Anecdotally, I recently had a CPU fail for the first time | and it was a 7nm one. Sent it to AMD, they verified the | failure and sent a new one back. Meanwhile I have had | assorted 22nm/14nm processors around the house chugging | along for years without any issues. | monocasa wrote: | > Sophie Wilson has said that cost-per-transistor is rising | since we shrank below 22nm | | The data has been saying otherwise. 5nm is the only node | that increased $/transistor beyond it's previous node | (7nm), and that's at a time when Apple payed out the ass | for a monopoly on the node except testing shuttle runs from | competitors but isn't a sign of a fundamental increase in | cost. | | https://cdn.mos.cms.futurecdn.net/Unwdy4CoCC6A6Gn4JE38Hc-97 | 0... | | > and that a 7nm die cannot run more than half of its | transistors at the same time without melting. | | That's true for trying to run multiGhz designs in classic | plastic package BGAs like most cell phone chips, but that's | been true for a while, hence why flip chip packages are a | thing. Actually having a heat spreader connected to the die | goes a long way. | | Wilson's comments aren't incorrect from a certain point of | view, but tend to get extrapolated out of her niche to the | greater industry in a way that pushes the statements into | inaccuracy. | chasil wrote: | Thanks, the cost per wafer does look convincing. I wonder | where Wilson's figures emerged. | [deleted] | monocasa wrote: | I think what's happening is that initially a node is more | expensive, because of consolidation of the market and | supply/demand amongst all of the fabless customers until | capacity fully comes up. Once we're back into steady | state we see the traditional economics of $/transistor | falling. | | That sort of coincides with TSMC having competitive, | close to leading edge nodes (so the 28nm timeframe) which | would line up with the rumor. The information simply | hasn't been updated over the timeframe of the node. | Previous to that the cost of the node was pretty fixed as | long as someone like ARM cared about, now there's a lot | more economic effects from the increased buyer | competition that heavily changes final cost over time. | chasil wrote: | I believe her talk was from late 2017, so 7nm would have | been expensive. | | At the same time, AFAIK Intel was doing quite well at | 14nm finfet even then (likely better than any other | foundry?), but that production capacity was not available | to ARM, so I guess it didn't count. | monocasa wrote: | Yeah exactly. I want to be clear, I've got a tremendous | amount of respect for Sophie Wilson; she's a much better | engineer and more connected to how the industry is going. | Her statements simply require a lot more caveats than | they are normally given. It's more about the much the | changing position of ARM and TSMC in the market place | than anything else. | | > At the same time, AFAIK Intel was doing quite well at | 14nm finfet even then (likely better than any other | foundry?), but that production capacity was not available | to ARM, so I guess it didn't count. | | Yeah, and Intel was right in the middle of their biggest | misstep. Intel 10nm was in risk production for Cannon | Lake with awful yields and therefore a huge $/transistor. | It got shipped anyway as one SKU in May of 2018 that | didn't make financial sense (there's rumors that | management bonuses were tied to that release), before | being relegated to process improvements for years until | just recently. | | It would have been fair for her to extrapolate a trend | there that actually ended up being more complex in | hindsight. | seanp2k2 wrote: | We're going to need stuff like processors with built-in | microfluidic cooling layers. Why not have club sandwich | construction with every other layer working on heat | extraction, power, or both? I see a future with cube-shaped | processors with hundreds of layers. | threatripper wrote: | This problem is not specific to Intel. Many chips are not | using the smallest node sizes and old fabs get a second | life producing those chips instead of CPUs. That could soon | be the future for Intel fabs. | monocasa wrote: | That's where a fab goes to die. It needs to be fully | capitalized at that point or else you don't have the | money for R&D on leading edge. Intel going this direction | is the direction of GloFo and no longer even attempting | leading edge node production in the US anymore. | me_me_me wrote: | Fabless scenario will not happen, ever. | | One simple reason that similarly to aeronautics chip | manufactures are critical industries for US. | | US government will not let either one die. | credit_guy wrote: | Does an Arizona fab of TSMC count as an American fab? | threatripper wrote: | We don't need small nodes for military/aero chips. It would | suffice to keep some old fabs running to secure supply. The | US needs fabs but they don't need Intel as a company. | | Splitting out the fab business and letting it run into the | red however could incentivize the government to save the | fabs and the jobs with billions of bailout money while the | other Intel continues to rake in money with TSMC chips. | alrs wrote: | This assumes that US military has no need for servers, | desktops, or mobile devices. | threatripper wrote: | Which will be covered by fabs built on US soil which | presumably are contractually obliged to manufacture chips | for the military on demand. (And obliged to take the | precautions to be able to do so.) Some of them can be | (ex) Intel fabs, some TSMC or Samsung. I don't see any | hard necessity for Intel to keep their fabs in house when | there are other solutions. | me_me_me wrote: | Literally every sentence is out of touch and outright | wrong. I don't even know where to start. | | > We don't need small nodes for military/aero chips. It | would suffice to keep some old fabs running to secure | supply. | | What? How do you think the military works, via smoke | signals? They need computer, a LOT of computers to | operate day to day. And that PCs for just boring clerk | work. Not to mention the idea that old chips would | somehow be good enough to keep up technologically, both | military and economically. | | > The US needs fabs but they don't need Intel as a | company. | | Oh yeah? They will just a walk a engineer corps to a fab | and have them press big red manufacture button few | million times to make enough chips right? Can't be more | difficult than that... | threatripper wrote: | What is the problem with an Intel design manufactured in | a TSMC fab built on US soil? There will be fabs on US | soil in the future but they might not be owned by Intel. | me_me_me wrote: | hmmm I dunno... maybe China taking Taiwan and banning all | export of electronics to US, leaving US with pants down | monocasa wrote: | If China invades Taiwan, the rumor is that Taiwan's | strategy is to scuttle everything at the last second so | China isn't additionally incentivized to invade just to | get the electronics manufacturing infrastructure. | | So there won't be any electronics manufacturing to ban at | all, but it leaves the US in the same place. | Kon-Peki wrote: | It probably depends on whether the chip is a commodity | chip or something special | | https://en.wikipedia.org/wiki/International_Traffic_in_Ar | ms_... | thereddaikon wrote: | The military doesn't just need chips for cruise missiles | and radars. There has actually been a big push the last | 20 years to move a lot of systems to COTS architectures. | An Arleigh Burke destroyer uses Xeon powered blade | servers to run its Aegis combat system not some weird | proprietary chip from the 80s. | | The level of computerization and networking is going up | in the military so the needs will only increase. Intel's | CONUS fabs are a national security concern. | monocasa wrote: | Increasingly we do. The military is very much embracing | edge compute currently. | JumpCrisscross wrote: | > _US government will not let either one die_ | | This isn't Intel's shareholders' problem. Split the | businesses. If the U.S. wants to bail out Intel's fabs, let | them. (Though a better strategy would be a general subsidy | for anyone building fabs in America, which should be a | thing.) | SavantIdiot wrote: | > Intel's internal manufacturing operations must earn Intel's | business | | This is putting lots of high-level employee's future earnings | on the line in a far more direct manner. It will be interesting | to see if they accept this challenge, or fight it in order to | accept the slow decay that will still ensure at least a longer- | term personal financial gain (i.e., instead of failing in 2 | years, failing slowly over 10). | skynetv2 wrote: | This was always the case. If anyone tells you otherwise, they | do not know. In fact, every team at Intel operates as a P&L | ruthlessly, not just manufacturing. | strikelaserclaw wrote: | decoupling | blihp wrote: | I don't see it as that much of a risk. One of the reasons Intel | has so thoroughly dominated for so long was that they were at | least at parity to slightly ahead on process nodes. If they | don't get back there fast, they are in real trouble. Intel | likes to command a price premium and you can't do that in | second place. | threatripper wrote: | I think some of the advantage comes from synergies of having | microarchitecture development and process development under | one roof. You can fine tune them together to get the best | performance/power ratio. Even if both alone are just on par | with the competition, together they are still ahead by a few | iterations. Also they get out new products a bit faster. | | The problem is that by switching to another fab they lose | these advantages. | DCKing wrote: | > The company's internal manufacturing operations can survive | only if they figure out how to outperform competitors. | | Is outperforming competitors necessary for Intel's survival? | There's plenty of fabs in the world doing quite alright well | behind TSMC, the vast majority of which can't even come close | to Intel's current capabilities [1]. Even if Intel never | succeeds with their existing process roadmap - which is _not_ | on pace to beat TSMC - they still possess some of the most | advanced chip manufacturing processes in a world that 's all | the more dependant on chip manufacturing. | | GlobalFoundries got out of the race on 14nm - a 14nm not as | good as Intel's 14nm and far behind Intel's 10nm - and is still | posting YoY revenue growth, despite losing much of the volume | they were producing for AMD over the last few years. | | In addition to that, I suppose that even if Intel merely | succeeds in roughly keeping up with TSMC and Samsung (their | current 10nm being on par with Samsung's and TSMC's 7nm would | classify as "roughly keeping up", I think) there's American | national security interests at play. _Especially_ so if Intel | 's manufacturing capabilities are accessible to (American) | third parties. No way the powers that be would let Intel's | manufacturing plants go under. | | It's a pretty bold strategy at face value, but I think it's | actually a pretty straightforward choice and the risks aren't | all that existential. | | [1]: | https://en.wikipedia.org/wiki/List_of_semiconductor_fabricat... | seanp2k2 wrote: | >No way the powers that be would let Intel's manufacturing | plants go under. | | Perhaps that's part of the calculus too; if they know that | Uncle Sam will backstop any failures, are they not also | making the same play as their competitors were in the 80s? | The problem is that by the mid-1980s Japanese competitors | were producing more reliable memory at lower costs | (allegedly) backed by unlimited funding from the | Japanese government, and Intel was struggling to compete... | ignoramous wrote: | > _Intel 's internal manufacturing operations must earn Intel's | business, in competition against every other fab-as-a-service | provider in the world, including TSMC._ | | I think this might be an interesting retrospective application | of _The Innovator 's Solution_'s 'be willing to cannibalize | revenue streams in-order to avoid disruption.' | kelp wrote: | It seems like echos of, and a less drastic version of the | AMD/Global Foundries split. | | And then Global Foundries couldn't keep up, cancelled their 7nm | work, and AMD has been sending more and more business to TSMC. | bsder wrote: | > In short, Gelsinger and his team are configuring the company | so Intel's internal manufacturing operations must earn Intel's | business, in competition against every other fab-as-a-service | provider in the world, including TSMC. | | This is a monumentally stupid idea and, even worse, _we have | seen it before_. Every company that has done this in the past | is _gone_. | | I would say that every VLSI designer in Intel should be | printing their resumes, but I'm pretty sure that the decent | ones left long ago. | | In addition, this _completely_ throws away the advantage that | having a fab brings at a time when it 's finally important. | | Fabless works great when you can push a synthesis button every | 18 months and your chip doubles in speed. Smart design can't | keep up with that. | | However, once Moore's Law stops and you can't run every | transistor without melting your chip, suddenly design matters | again. You want to be able to do creative things with a bit of | tweak from your fab. | ksec wrote: | >Only one thing is certain: Gelsinger and his team have a lot | of cojones. | | Trained and mentored by Andy Grove. Disciple of the old school | "Only the paranoid survive". I expect nothing less. | | I hope someday he could do the next unthinkable, Once their | foundry could at least compete against Samsung, open up the x86 | ISA. Or clean up the ISA and call it something like AE86. In | the very long term, x86 is pretty much dead. By 2030 you would | expect all hyperscaler to be on ARM with x86 serving some minor | legacy clients. You have PC manufacture eager to compete with | Apple and ChromeBook. So x86 could loss up to 50% of its | current volume by 2030. Then a long tail of 10 - 15 years | decline. | alfor wrote: | - The PC is in decline (bad for intel) | | - Intel lost the mobile platform | | - Apple is moving to ARM, PC laptops to follow soon. | | - AMD is eating intel lunch in the high performance x86 | | - Servers are starting to move to ARM | | Last piece of the puzzle - Intel is not doing great at the fab | level. | | It doesn't look good for Intel, they need to do a radical | transformation fast. | | Innovate now or die. | neogodless wrote: | I agree with many of your points, however... | | > - Apple is moving to ARM, PC laptops to follow soon. | | > - AMD is eating intel lunch in the high performance x86 | | That first one really needs... a source? It's a guess, but far | from a certainty! | | The second one is partially true? AMD has greatly increased | their market share and is much more competitive with Intel than | they were 5 years earlier. They are not _beating_ Intel in | market share, but they are gaining on them. They recently had | leads in performance, but it 's not a one-sided race. | silvestrov wrote: | The first is true to a degree if "PC laptops" includes the | laptops kids use in schools: a lot of schools have switched | to ChromeBooks due to price and ease of management. | pclmulqdq wrote: | The more I have seen of big tech companies, the more I think they | need to split up from time to time. Complacency and ossification | emerge as a product of scale. If you want to avoid them, reduce | your scale. If your manufacturing engineers (really, the | managers) need to be at the cutting edge of manufacturing or be | out of a job, they will figure out how to get there. | | Intel has had a number of embarrassing technology misses that | have put them behind: the first I remember was through-silicon | vias for 3D stacking and it was only a matter of time before they | missed something necessary for an advanced fab, and they missed | on EUV. | | Their foundry engineers at the same time were recruiting more | engineers out of school on the basis that they were the "best in | the world." They thought their position was unassailable because | they had high yield at 22 nm, so they rested on their laurels and | missed technology after technology before it bit them. | airstrike wrote: | > The more I have seen of big tech companies, the more I think | they need to split up from time to time. | | This is somewhat true of every industry. Industrial companies | in particular are always playing the game of conglomerization | vs. specialization. But I agree it will definitely be | interesting to see this play out in Tech in the near future | threatripper wrote: | There is little information available on what was actually | going on inside Intel. All big companies have big problems and | each single division or working group is not indicative of the | company as a whole. You can find several people with good and | bad experience painting all kinds of pictures. A single voice | won't tell you the whole picture. Probable even the leaders of | Intel don't really know exactly where they are standing. | | Of course all companies go through cycles of explosive growth, | ossification and renewal. Small companies often die because too | much of the company goes sour at the same time. In big | companies this can be more spread out over time and space and | there's always enough left to rebuild the leadership from the | inside and push out new profitable products in time. | | That being said I don't have any indication that Intel right | now is in a particularly good position. Nobody in the | engineering area seems to feel too honored for having worked at | Intel. Yet, they got 10nm finally working and continue to rake | in money. Of great concern to me is that much of that money is | paid out to investors instead of reinvesting it into the | company. There doesn't seem to be a convincing plan of growing | the business in the future. Also they did not admit that they | have a serious problem at their hands for a long time. You just | can't trust anything that Intel releases; it's just a rosy | projection of the future when everything goes as planned. | | Unless a group of senior Ex-Intel get together and tell their | war-stories we won't know what actually was going on inside. | pm90 wrote: | How to fix Intel: | | 1. Lobby the US Government and public to increase public spending | on manufacturing semiconductors in the US | | 2. Corner most of those subsidies/funding | | 3. Done | | Only partially joking; Why has Intel not made the over reliance | of overseas fabs a natsec issue? | khyryk wrote: | It has, now it's waiting for money from the CHIPS Act, which is | now in the House. | awill wrote: | I'm always confused when I hear reviewers say that Intel is back | on top. | | They beat AMD's benchmarks at ~double the power. I get it's | winning, but it's not a fair fight. I certainly don't want my | computer to be 10% faster at double the wattage. | magila wrote: | The whole "double the power of AMD" meme is an | oversimplification at best. It's all based on a 12900K running | a heavy multi-core load with an increased power limit. This is | the ultimate "I don't care about power efficiency" setup so | it's kind of silly to knock it for drawing lots of power. | | If you care about power efficiency you can drop the power limit | and still be very competitive with AMD. The only reason you | don't see Ryzen CPUs drawing the same power as a maxed out | 12900K is because if you pumped that much power into a Ryzen | CPU it would explode (at least without LN2 being involved). | pradn wrote: | For home use, that's fine. I don't worry about the power peak | usage of my computer since it rarely gets to that level. For | servers, that's a no-go. | neogodless wrote: | I think there's another way to word this or spin this... | | For home use, it's going to depend on the consumer. Some | people want a chip that's easy to power and cool, and still | gets 80/90/95% of the performance. Some people want the | _absolute best performer_ (for their specific uses) with less | regard to cooling and power. | pradn wrote: | I think peak performance, even if it gets thermo-throttled | quickly, is important. Web site JS parsing during loading | is like that, and that's an essential function of my | computer these days. And perhaps for bursty sections of AAA | games as well. | | But yes, it depends on the user. | acomjean wrote: | Reviewers like a good comeback story. And intel is doing much | better than it was doing before (I own machines with both types | of CPUs). And intel did some different things with the new | chips (2 different core types.), so interesting. | | AMD hadn't released there new chips when some of that stuff was | written. but its actually a little bit exciting that | competition has returned and I suspect it will be good for | consumers as long as both chip makers have competitive | products. | wmf wrote: | Alder Lake is efficient at the low end; the 12100 and 12400 are | really good against AMD. | mwcampbell wrote: | > if tomorrow morning the world's great chip companies were to | agree to stop advancing the technology, Moore's Law would be | repealed by tomorrow evening, leaving the next few decades with | the task of mopping up all of its implications. | | On the one hand, if this happened, it could be good for putting | an end to the routine obsolescence of electronics. Then maybe we | could get to the future that bunnie predicted roughly 10 years | ago [1], complete with heirloom laptops. | | On the other hand, ongoing advances in semiconductor technology | let us solve real problems -- not just in automation that makes | the rich richer, enables mass surveillance, and possibly takes | away jobs, but in areas that actually make people's lives better, | such as accessibility. If Moore's Law had stopped before the SoC | in the iPhone 3GS had been introduced, would we have smartphones | with built-in screen readers for blind people? If it had stopped | before the Apple A12 in 2018, would the iOS VoiceOver screen | reader be able to use on-device machine learning to provide | access to apps that weren't designed to be accessible? (Edit: | A9-based devices could run iOS 14, but not with this feature.) | What new problems will be solved by further advances in | semiconductors? I don't know, but I know I shouldn't wish for an | end to those advances. I just wish we could have them without | ever-increasing software bloat that leads to obsolete hardware. | | [1]:' https://www.bunniestudios.com/blog/?page_id=1927 | lvl100 wrote: | I am pretty bearish on the the semi space especially after TSMC | comments last week. I really think they're expanding at the worst | possible time. Understandably, this is largely due to politics. | [deleted] | altcognito wrote: | I can't quite grasp why anyone would be bearish on the semi | space. It's been non-stop growth for the past four decades and | only appears to be accelerating. | | What in the world is there to be "bearish" about? | acomjean wrote: | I'm not the original poster. But I'm guessing that the | reasoning is: there might be more capacity than needed so | that would drive profits down. Cheap chips seem like a great | thing to me, but I'm not selling them. | | But at this point when there putting chips in everything its | hard not to see the market expanding to meet increased | capacity. (I think some greating cards have chips in them to | control music/lights....) | mrintellectual wrote: | Having worked at Intel before, I can safely say that it's | difficult to boil down Intel's problems to a single issue (e.g. | chip manufacturing). Stagnant company culture, limited growth | opportunities within the company, and a fanatical focus on | selling CPUs instead of innovating are all problems that Intel | currently faces. | hedgehog wrote: | The conversation is also muddied by varying definitions of | success and failure. Intel isn't going to go broke in any | foreseeable future but they may lose the premium margins they | used to earn for having the best performing product. To the | extent fab advantage drove that advantage (likely a big part) | and that advantage is not coming back (because they no longer | have a scale advantage) then maybe the premium margins will | never come back. That's what investors worry about. | anonporridge wrote: | Limited growth opportunities seems like a big problem that can | kill any organization. | | Specifically if leadership gets ossified and there's no | realistic path for ambitious young people to rise into | positions of power and responsibility. | | When this happens, the ambitious young people are more likely | to vote with the feet and build a competing organization that | they will actually have power running. | | This is especially problematic when it happens in government. | What are young people in America supposed to do when the | political conversation can't break out of asking which of near | geriatric octogenarian who had a childhood with rotary phones | should be our next president? | | Our culture seems to have lost the very important feature that | there comes a time when you've had your turn at the helm, and | now you need to step down and support new leadership. | gumby wrote: | I think Thompson says this (while talking about a series of | lackluster CEOs) and says the topic of the "split" is a | consequence of that, which has to be fixed (and is merely one | of the problems) | klelatti wrote: | Problem is the less product Intel pushes through its own fabs the | less investment they will get and the more they will fall behind. | | This feels like a defensive / short term value maximisation | strategy - and might be the right one in terms of total value. It | doesn't feel like a long term vote of confidence. | MangoCoffee wrote: | >less product Intel pushes through its own fabs | | that's the problem the article mentioned. Intel is an IDM. they | developed their own tools/process and it gave Intel an edge | during the early years of the semis industry. | | however, it look like the industry catch up to Intel. Intel's | own tools/process is no longer enough and in fact it became an | technical debt. | | Intel split into two allowed Intel's foundry to use standard | equipment/software thus no needs to waste money on building its | own tools. | klelatti wrote: | Why can't they migrate to industry standard tools whilst | keeping their business in house? | MangoCoffee wrote: | no conflict of interest with Intel's foundry customers. | klelatti wrote: | Not sure I understand your point here. They will need to | migrate to industry standard tools for foundry customers | anyway. | | Seems to me that planning to place more business with | TSMC is essentially an admission that they don't expect | to be competitive with TSMC in the near future. And | Gelsinger has more visibility on this than anyone. | MangoCoffee wrote: | >They will need to migrate to industry standard tools for | foundry customers anyway. | | foundry is about trust. lets say AMD want to use Intel | foundry, how can you be sure the Intel's design side | won't take a peek at AMD's x86 design. your fabless | customers spend million on design and their design is | their life blood. how they going to trust you if you also | sell semis. | | example: Samsung used to make chip for iphone. Apple | ditched them as soon as Samsung compete with Apple w/its | own smartphone. | | TSMC have over 50% of the market share because they only | do one thing and one thing only. | prewett wrote: | Intel can already peak at AMD's designs, just take the | cover off and dissolve off the layers one by one. It's | probably about as helpful getting your competitor's | source code; figuring out what's going on probably takes | longer than figuring out how to do it yourself. Maybe | worse, since everything is so low-level in hardware, it'd | make assembly look like a high-level language. I'm no | hardware designer, but I expect that the results of tape- | out are roughly the same as the results of etching off | the layers: they need to create masks for each layer, | specify the composition of the layer, etc. And then after | understanding it, you still have to implement it and get | started fabbing it. So I'm not sure that fabbing your | competitor's product is a huge risk. | | I think Apple stopping using Samsung is more related to | Apple's higher-level issues: why do business with someone | you accuse of violating your design patents. Not because | you think they'll copy your IP, but out of principle. | There's no IP-related reason Apple needed to stop buying | screens from Samsung; Apple has never manufactured | screens. | benreesman wrote: | Intel blew EUV, which fair enough, it's hard as hell. Are there | other problems, sure, but when you go from leading-edge process | technology to a two node lag, you're fucked either way. | | This article is rambling and over-complicated. | | For a much more insightful and compelling view into Intel at its | greatest I recommend "The Pentium Chronicles". | [deleted] | throwaway4good wrote: | I don't get it - who would use intel as a fab? How would they | competitive? And trustworthy in the sense, not doing competing | designs in house? | wbsss4412 wrote: | People who need to make chips? | | It's not like there is an excess of capacity right now, and | there are only a small number of players that even offer what | intel has available. | Cthulhu_ wrote: | Well first off, Intel themselves; they sell millions of chips a | year. Given the current chip crisis, after their own capacity | they would sell off the rest to the highest bidder. I don't | think they would be shy of customers, except for direct | competitors. But keep in mind that CPUs represent only part of | the semiconductor market. | intev wrote: | Microsoft? Maybe another phone maker who can't secure | manufacturing with TSMC? They will clearly be tier II for a | while, but I think there's enough chip design companies out | there vying for manufacturing capacity. | dannyw wrote: | Apple buys screens and other components from Samsung | (potentially giving them an early look into the next iPhone's | form factor, visual design, and specs); the supply chain is | built around trust. | jacobr1 wrote: | Those who want to manufacture their chips locally (US or EU). | MangoCoffee wrote: | Apple will ditch TSMC if someone else offer better node and | cheaper price. | | there is no loyalty. Apple will ditch you if they can get | cheaper price or you are competing with them like Samsung. | UncleOxidant wrote: | > who would use intel as a fab? | | We've got a very serious shortage of fab capacity right now | that will probably last for at least 2 or 3 more years. The | answer would be anyone who needs fab capacity of the sort Intel | can provide (not likely AMD of course, though I could see Apple | giving it a try). | pwthornton wrote: | I think that's the wrong question. Who would use Intel as a | cutting edge fab? Unclear. But plenty of people would love | access to a fab for their non-cutting edge needs. | | The risk with this is that Intel will become a fab for older, | less expensive chips, while TSMC and others gobble up the top | tier stuff. | | But the world needs more, not less, fab capacity, and there are | plenty of people would gladly use Intel for their needs if the | opportunity was there. | ac29 wrote: | Additionally, only 23% of TSMCs revenue comes from their most | cutting edge 5nm process as of the most recent quarter [0]. | Everything else is 7nm or larger, which Intel is capable of | doing today. In the scenario where Intel remains a major node | behind TSMC, there is still a lot of fab business they could | pick up. | | [0] https://twitter.com/IanCutress/status/1481581119740989440 | /ph... | MangoCoffee wrote: | bottom line: Intel's edge is an technical debt. | | Intel's own tools/process used to give them an edge over everyone | else during the early years in the semis industry. Overtime, tool | suppliers have catch up with Intel and Intel is falling behind | due to their own tools delay/insufficient. | | Intel's own tools is losing out to commodity and standardized | tools ___________________________________________________________________ (page generated 2022-01-18 23:00 UTC)