[HN Gopher] Intel plans spinoff of FPGA unit ___________________________________________________________________ Intel plans spinoff of FPGA unit Author : ChuckMcM Score : 109 points Date : 2023-12-21 18:17 UTC (4 hours ago) (HTM) web link (www.networkworld.com) (TXT) w3m dump (www.networkworld.com) | ChuckMcM wrote: | Apparently they announced this in October. They are kicking | Altera, which they acquired for 16.7 billion back out into the | world to live or die on its own. Which feels a bit weird given | what feels like AMD's relative success with Xilinx. | | It also dampens their "US developed" technology pitch (which they | had been pushing pretty hard vs other solutions that were fabbed | at TSMC) I wonder if they will also give up on their third party | use of their fabs. | | Intel was the first "real" job I ever had and it was during Andy | Grove's tenure as CEO and his "paranoid" attitude was really | pervasive. I had been working on graphics chip which were the | next big thing until the chip recession and then they weren't. | And it followed a long series of Intel sending out tendrils | outside its core microcomputer (eventually x86/x51 only) winners | only to snap them back at the first sign of challenges. | | I wonder if we can get better open source tool support from the | new entity. That would be a win. | AceJohnny2 wrote: | > _only to snap them back at the first sign of challenges._ | | It's really quite impressive how bad Intel's track record has | been. Would you say it's a broad cultural problem within Intel? | Bad incentives? | | (I'b be interested if someone could come up with a list of all | of Intel's failed ventures. StrongARM, Optane, Altera... | Obviously some of these are par for the course for such a large | company as Intel, but Intel still seems to stand out) | | (I know one reason Apple is so secretive about its ventures is | because it knows many of them won't pan out, and it knows that | due to its size anything it does has huge repercussions, and | wants to avoid that) | 1-6 wrote: | When Intel does something like this, it reveals how tight | their operating margin is. This is not a good sign. Intel has | never been able to push away from the fact that they'll | always be a hardware first company. | | It's a very tight market out there. TSMC was able to make it | with very tight partnerships with OEMs and I doubt Intel can | pull off something like that. | ChuckMcM wrote: | I agree with 1-6 that they have always been really focused on | their margins, they were never a company that wanted to go | into debt for something that was 'unproven.' | | They have had lots of things they stopped doing, the graphics | chip I was involved in (82786) was one, the speech chips, the | iAPX431, the modem chips, their digital watch was an early | one. | | I always saw it as an exceptionally low tolerance for risk. | When I was their the 386 had just fabbed and I was one of the | people watching when the very first chip was powered up and | it actually worked (which amazed pretty much everyone). Sun | Microsystems was a little startup in Mountain View using the | 68000 and getting some traction I suggested to Intel that if | they used the graphic chip I was working on and the 386 they | could build an Intel branded "Workstation" and both replace | the VAXStations they were using for CAD work internally and | sell them as part of Intel's initiative to become a "systems | company." I got a response to my memo[1] from Dave House who | was later CEO but at the time just the head of Microprocessor | Operation (MIPO) Marketing explaining to this lowly NCG (new | college grad) that Workstations were not much of a market and | Intel was only interested in pursuing the "big" | opportunities. Six months later I joined Sun (just after they | IPO'd sadly) and didn't look back. :-) | | [1] Yes memo, as in typed up and printed out and put in a | manilla envelope where on the last line of the front had it | coming from Dave and being addressed to me. And yes, it | started as a memo I had printed out and given to my manager | because while Intel had "heard" of electronic mail they | preferred the "durability" of actual mail and forcing people | to use actual mail cut down on "frivolous" communications. | FirmwareBurner wrote: | _> I always saw it as an exceptionally low tolerance for | risk._ | | Big laugh at the claim of intel having a exceptionally low | tolerance to risk when they're the only semi company left | still running their own fabs, decades after AMD sold off | their own for being too risky of a venture to keep running | by themselves. And every fab company out there will vouch | for how risky that line of business is. Even more so to | pour billions into a business like this while knowing | Samsung/TSMC might beat you. | | More correct would be that Intel has a low tolerance to | risk for things that are way outside of their core | competencies which have traditionally been designing X86 | CPUs and fabbing them, and that's mostly it. | | I commend them for not giving up on advancing their fab | nodes despite the massive issues and setbacks(they could | have spun it off like AMD did), and for building the ARC | GPUs and keep improving them despite low sales. | Grazester wrote: | I thought AMD had to spin off their fabs due to cost and | cash flow issue because Intel was too busy playing dirty? | FirmwareBurner wrote: | _" Real men have fabs!"_ - AMD founder and CEO at the | time, Jerry Sanders [1] | | He said this some time in the 80's to early 90's as a | defence for their expensive commitment to still be | running their own fabs when most big semi companies like | Motorola were selling off their fabs and going fabless | because it was eating their resources and it was | impossible to compete with the likes of TSMC on node | shrinks and value. | | So even without Intel screwing with them in anti- | competitive ways, they could not have kept their fabs | competitive much longer. The writing was already on the | wall way back then but Jerry Sanders was just being | stubborn. | | [1] https://archive.is/igC4r | kimixa wrote: | I mean... Maybe? But Intel proved their success of owning | fabs throughout the 90s and 2000s, only 20-30 years after | that statement were Intel even considered to be possibly | dis-advantaged by owning it's fabs. But that was also a | very different market. | | So I don't think it's as direct a link as you seem to | claim. Maybe it was the right decision even if AMD had | the money to keep it going. But it will likely forever be | one of those questions that has no solid answer either | way. | phkahler wrote: | >> I thought AMD had to spin off their fabs due to cost | and cash flow issue because Intel was too busy playing | dirty? | | IMHO Ruiz cut back on engineering and increased | marketing. Then purchased ATI which I though was an odd | or even dumb move (it was not). But that expensive | purchase along with the failure of bulldozer almost | killed the company. That's why they spun off Global | Foundries - to stay alive. | oumua_don17 wrote: | >> when they're the only semi company left still running | their own fabs | | Yes and even within that they are risk averse, no need to | look beyond 10nm and their reluctance to EUV. | | So they are really risk averse and they just want their | success to continue infinitely. In fact, contrary to Pat | G claiming that Nvidia is lucky; it's really Intel which | was lucky. | | >> commend them for not giving up on advancing their fab | nodes despite the massive issues and setbacks(they could | have spun it off like AMD did), and for building the ARC | GPUs | | even if the turn around time in semi market is longer, | the way they faltered on their Arc timelines and no | encouraging signs of their new fab nodes is not | commendable! | FirmwareBurner wrote: | _> even if the turn around time in semi market is longer, | the way they faltered on their Arc timelines and no | encouraging signs of their new fab nodes is not | commendable!_ | | Sure, Intel fucked up along the way, and it's easy to | point fingers and laugh from the comfort of the armchair | at Intel tripping over massively complex engineering | challenges, but if designing and building successful GPUs | and sub-7nm fabs was easy, everyone would be doing it. | ethbr1 wrote: | > _Yes and even within [fab] they are risk averse, no | need to look beyond 10nm and their reluctance to EUV._ | | Wasn't there an article on HN 2 days ago about Intel | pushing 2 new technologies simultaneously in their next | node? (ribbon gate and backside power?) | quercusa wrote: | > _They have had lots of things they stopped doing_ | | I have wondered if leadership didn't understand the Gartner | Hype Cycle, because a lot of projects seem to get killed | after the investments but before significant success was | even possible (the Trough of Disillusionment). | thijson wrote: | I worked at Intel for 10 years, mainly the Otellini | years. To myself I called it Corporate Attention Deficit | Disorder. I think it's a symptom of bad management, | always moving on to the next shiny rock, no long term | vision. | | During my tenure I witnessed several failed initiatives. | Itanium, Wimax, Digital Home, x86 phone (android), LTE | modem. Imagine the billions wasted. | senderista wrote: | Surely Itanium was a massive risk (that obviously didn't | pay off)? | drjasonharrison wrote: | Wikipedia has a partial list: https://en.wikipedia.org/wiki/L | ist_of_mergers_and_acquisitio... what it is missing is the | eventual disposition of the acquisition (product brought to | market, unit pushed out, unit closed down, ...). | AceJohnny2 wrote: | > _I wonder if we can get better open source tool support from | the new entity. That would be a win._ | | Tragically, I wouldn't bet on it. In fact, I'd be shocked if | that were to happen. | raverbashing wrote: | I think Intel is living their Nokia moment, except their lunch | will be eaten much more slowly by the competition | burnte wrote: | They do this so often, buy a company, sell it a few years | later. So weird. | lisper wrote: | Sometimes they don't even bother to sell it. In the case of | Barefoot Networks, they bought it, spent three years | developing the next-generation product, got it to within | 80-90% of being ready for tapeout, and then just pulled the | plug. | andrewia wrote: | One of my relatives is a high level Barefoot employee who | was integrated into Intel leadership after the acquisition. | I was told that it was killed because of worries about | internal competition with Intel IPUs. It still seems like a | ridiculously foolish idea, considering how much money Intel | spent acquiring Barefoot. Not that the founders and CEO | care, they got their money! | | The whole thing definitely seems symptomatic of Intel's | extreme caution. I wonder why they didn't apply the same | caution when buying Barefoot. Intel obviously didn't have a | great idea on how to leverage them. | | This is all in stark contrast to AMD buying Pensando, a | company that my SO works for. It makes sense for AMD to | acquire Pensando to expand their data center offerings and | compete directly with Intel. I think AMD made a smart buy. | lisper wrote: | I was a part-time contractor for Barefoot, and came along | for the ride during the acquisition, so I have some | first-hand knowledge of this. I am very much out of the | managerial loop so I have no insight into the actual | motives for killing the project, but I can tell you that | at least one of the founders cared very much and fought | tooth and nail to keep it alive. | | > It still seems like a ridiculously foolish idea, | considering how much money Intel spent acquiring Barefoot | | Acquiring companies in order to kill them is a horrible | business practice but not unusual. The thing that makes | no sense to me is why they kept it going for three years | before killing it. If they bought it in order to kill it, | they should have done that _before_ spending another | hundred million on it. | ithkuil wrote: | Perhaps having plausible deniability that they didn't | acquire a competitor just to kill it was worth that extra | hundred million? | lisper wrote: | Why on earth should Intel care about having plausible | deniability about that? Buying a company in order to kill | it is not illegal. It isn't even considered unethical | except by a few utopian idealists. It's a common and | accepted practice in the business world. | crote wrote: | Antitrust laws are a thing, you know. Buying out your | competition in order to get rid of them is indeed | illegal, once you get big enough. | touisteur wrote: | They might have discouraged alternatives to pop-up. 'Oh | Intel has this in the pipeline, it's hopeless to compete, | let's invest in something else'. I've heard this so many | times, just to see Intel kill said product, product-line. | .. | | At this point, if it's not about x86 cpus, listening to | Intel's roadmaps seems foolish. | ethbr1 wrote: | > _The thing that makes no sense to me is why they kept | it going for three years before killing it._ | | Internal politics? | | Presumably the internal team(s) with overlap were against | the acquisition. | | But it was likely easier to see if the integration failed | on its own before spending the political capital to kill | it. | | Folks forget executives at large companies usually | optimize for "my career" over "the company." | 7speter wrote: | I don't know I don't have the credentials to be an engineer at | Intel, but I think they know its a bad idea to end or spin off | their foundry service, especially now that they have a new | client that will know the ins and outs of their fabrication | process. | UncleOxidant wrote: | > I wonder if we can get better open source tool support from | the new entity. That would be a win. | | Unfortunately, this seems unlikely. The FPGA vendor tools are a | complete shit show and open source tools would be greatly | welcomed - they're doing really well in the Lattice/ICE40 space | but those are small FPGAs. But Altera and Xilinx don't seem at | all inclined to encourage the development of open source | alternatives. | chrsw wrote: | That's not where the money is. They don't care about | hobbyists, startups or schools. In FPGA world is about big | contracts, big design wins. Networking, defense, space, | industrial control etc. It's why the design tools are so bad | too: there's no reason for the FPGA vendors to invest in | those tools when the design wins come from size and | performance. In fact, giving customers an escape hatch from | vendor lock in would be a Very Bad Idea. | UncleOxidant wrote: | Indeed. I was working at a startup doing some FPGA work | about 10 years ago. We ran into several bugs in the Xilinx | software. Initially we could submit bug reports, but about | 6 months in Xilinx suddenly changed their policy and | decreed that only tier 1 customers (definitely not us) | could submit bugs - everyone else had to look for help on | the forums. We spent a lot of our time just working around | the bugs in their tools. At that point I determined that, | though FPGA development was kind of fun and interesting, I | would not go into a field where I was going to be dependent | on such buggy software and went back to the software side | (where pretty much all the development tools are open | source and if you run into trouble you're going to be able | to get help). | crote wrote: | It does beg the question whether this is simply a chicken- | and-egg problem. | | When it is almost impossible to develop for, you only get | big contracts because nobody else has the resources to | design products for it. On the other hand, with good and | free design tools the toy projects done by hobbyists and | schools can serve as the catalyst for using it in medium- | scale projects. | | There are plenty of applications imaginable for something | like Intel's SmartNIC platform - but you're not going to | see any of them unless tinkerers can get their hands on | them. | mook wrote: | Of course, that's kind of also why nVidia is trouncing AMD | in ML/GPGPU. AMD chased after the big iron, but nVidia got | stuff working on consumer hardware... (their early lead | definitely helped too, of course) | phkahler wrote: | >> The FPGA vendor tools are a complete shit show and open | source tools would be greatly welcomed - they're doing really | well in the Lattice/ICE40 space but those are small FPGAs. | | Thanks for the stock tip. The size of a chip can change, and | will. Stupidity really tends not to. Great dev tools are | incredibly important, and the people using them actually | _can_ influence decisions on which parts to use. | wslh wrote: | Side question, how do you see the future of Intel? Do you share | the doom perspective or think that they continue to have | opportunities ot catch up? Thanks! | ethbr1 wrote: | The answer to that is the intersection of thousands of state | of the art fab technical issues and management decisions. | | Recent hindsight with TSMC makes it look "easy", but it's | incredibly uncertain simply due to the complexity and number | of pieces that have to align. | vachina wrote: | So what're they gonna name it this time? Altera? Hope they kept | the original documentations. | shrewm wrote: | I'm thinking Ctrlera(tm). | krallja wrote: | I like how you think. Could tie into VR/metaverse hype by | naming it after the Meta key, though. | 1-6 wrote: | This was actually a bright-spot in Intel's lineup that had a | chance to be a moonshot. I don't know why Intel is giving up this | early especially when AMD acquired Xilinx recently. Is intel | trying to emulate Nvidia? I don't think they should try to become | another Nvidia. They should be building programmable AI chips | with FPGA. Heck, LLMs are able to code Verilog. There are many | many possibilities. | tester756 wrote: | Why "giving up"? | | It's just spinoff and their people are at executive levels | brucethemoose2 wrote: | > They should be building programmable AI chips with FPGA | | Eh, people have been saying this for over a decade, and I think | that opportunity has passed. | | FPGAs are not going to beat GPUs anytime soon due to the | software ecosystem (among other things), and ultimately they | are not going to outrun ASICs that are now economically viable | (especially in the embedded space). | crotchfire wrote: | > FPGAs are not going to beat GPUs anytime soon due to the | software ecosystem | | Completely true as long as the chip companies stonewall open- | source toolchains. | | If Xilinx or Altera published the bitstream format for their | highest-volume high-end chip (i.e. one or two generations | back from the bleeding edge) you'd see the effect on the | entire AI space within twelve months. I'm not kidding. | | The interest in this kind of access from developers is | enormous, and the problems in this space are extremely | regular. There are massive opportunities for FPGAs to avoid | spilling to DRAM _or even to SRAM_ -- you have an ocean of | tiny register-speed memories in LUTRAM mode. | | But it will never happen. And so FPGAs will continue to be | trinkets for weapons manufacturers and not much else. | MBCook wrote: | They gave up on Xscale right before mobile demand shot to the | moon. | contrarian1234 wrote: | They acquired Altera in 2015, which in the tech world is when | the dinosaurs roamed the earth. How much more time would they | need to see a profitable synergy? Bearing in mind they can | probably see what's in the pipeline for the next couple of | years. | | I'm not an expert... But in my naiive opinion it seems entirely | reasonable to expect an acquisition to start paying off within | ten years? | | Would love to hear counterexamples | crote wrote: | A big issue is that sticking an FPGA onto the _CPU die | itself_ doesn 't really make sense. Those are some really | expensive transistors, and you're essentially wasting them | when they are not being used for cores or cache. Besides, | you'd be spending significant amounts of money building a | niche product. | | It would be a lot more viable to use a chiplet approach: bake | the FPGAs on a slightly-cheaper node, and just glue them | right next to the CPU itself. Unfortunately Intel is still | _miles_ behind AMD when it comes to chiplets, and their tech | is only just now barely starting to ship. | adrr wrote: | Because intel messes up everything with their bureaucracy. Look | at the NUC, they made it extremely hard to get because of their | convoluted distribution strategy. They could have easily done | direct to consumer and made it easy to get. | | Till they fix their culture problem, they should be unloading | these business units and profit by holding majority stakes in | them. | JoshTriplett wrote: | This is sad news; I was hoping one day we'd see chips with | substantial on-die FPGA fabrics, ideally in ways that we could | program with open tools. This announcement makes that less | likely. | rwmj wrote: | Red Hat supported a lot of research into this and there's some | really interesting stuff, but nothing that is very compelling | for commercial use. What uses do you have in mind? | | To my mind the more interesting stuff are the PCIe FPGA boards | like https://www.xilinx.com/products/boards-and- | kits/alveo/u200.h... | | One particularly interesting research project was using the | FPGA fabric to remap addresses, allowing database tables to be | "virtually" rearranged (eg. making a row-major data source into | a column-major source for easier searching). | https://disc.bu.edu/papers/edbt23-relational-memory | https://arxiv.org/pdf/2109.14349.pdf | thereisnospork wrote: | As (very much) a layman I've hoped to see something like | dynamic hardware acceleration. Eg for a game that has | advanced hair simulation, or 3d sound simulation: reconfigure | the fpga and offload the calculations. Maybe even more | pedestrian things like fpga driven bullet tragectory physics | might be implemented at the game engine level. | | More optimistically something along the lines of hands-off | offloading, where if the scheduler sees enough of the same | type of calculation (e.g sparse matrix multiplication) it can | reconfigure the fpga and offload. | magicalhippo wrote: | I was thinking it would be a good fit for crypto. New | algorithms could be implemented with better-than-software | performance, constant-time algorithms could be ensured etc. | adrian_b wrote: | True. | | For instance, SHA-3 is quite slow on x86 CPUs, which, | unlike the recent ARM CPUs, do not have any hardware | instructions to accelerate it. | | If there had been an included FPGA, it would have been | easy to implement a very fast SHA-3, as its hardware cost | is very low. | rwmj wrote: | I think the problem is this misunderstands what FPGAs are | good at. They're actually very bad (or at least painfully | slow) at calculations like your examples. GPUs are good at | that. | | FPGAs excel in parallelizing very simple operations. One | example where FPGAs are good is where you might want to | look for a specific string in a network packet. Because | FPGAs are electronic circuits you can replicate the "match | a byte" logic thousands of times and have those comparisons | all run in parallel and the results combined with AND & OR | gates into a yes/no decision. (I think the HFT crowd do | this sort of thing to preclassify network packets before | forwarding candidate packets up to software layers to do | the full decision making). | soulbadguy wrote: | > FPGAs excel in parallelizing very simple operations. | | I don't think that's a good characterization. FPGA are | good at what they end up being programmed for. And in the | final analysis, everything a chip does is broken down to | simple operation. | | The FPGA selling point has always been around perf/W and | efficiency with regard to a "set of task". An ASIC will | always be faster for a specific task, and CPU will always | be faster on average on everything. However, when | considering say "compression" or "check-sum" as general | class of algorithm, and FPGA with a set of predefined | configuration could be better and cheaper. | 70rd wrote: | Good at what they're being programmed for is a bit of | tautology. | | FPGAs being selected for performance per watt was only a | fairly recent phenomenon, when they were deployed on semi | large scale as password crackers/cryptocurrency miners. | | Their real strength is ultimately real time processing | (DSP or networking), with reconfigurability often quite | valuable for networking applications. For DSP | applications it's usually because a MOQ of custom silicon | can't be justified. | JoshTriplett wrote: | > What uses do you have in mind? | | Acceleration of new hashing and cryptographic algorithms, new | compression algorithms, new codecs, and many other similar | things, without adding special-purpose instructions for them. | | Implementation of fast virtual peripherals. Accurate | emulation of special-purpose hardware. | | Implementation of acceleration modules for well-established | software. Imagine if popular libraries or databases or other | engines didn't just come with acceleration using SIMD and | other instruction sets, they also came with modules loadable | on an attached FPGA if you have one. | vlovich123 wrote: | The problem is that these kind of applications suck for | multi-tenant clouds which FPGAs turn out to be poorly | suited for (high costs to switch out programs) and are too | expensive for the consumer. So the applications are quite | niche and limited to traditional use-cases of prototyping | ASIC designs rather than actual algorithm accelerators | which benefit more from dedicated circuitry/instructions in | terms of adoption. | rkagerer wrote: | Anything where you need cycle-perfect timing. | Laaas wrote: | The new Ryzen chips already have this as "XDNA" AFAIU | wmf wrote: | XDNA is not an FPGA. | wtallis wrote: | I think that's just a neural accelerator, not an FPGA. The IP | block happens to be from the Xilinx side of the company. | blackguardx wrote: | This is pretty funny. Intel bought Altera, which forced AMD to | buy Xilinx with all the zero interest rate money floating around. | AMD's purchase of Xilinx made less sense because AMD is fabless, | but Intel didn't end up doing anything with Altera. Its not clear | if Altera even started using Intel fabs for its chips. AMD's | Xilinx has been comparatively more successful, but I don't think | that had anything to do with AMD. | | Maybe we can look forward to all the ZIRP semiconductor | consolidations to unwind. | pclmulqdq wrote: | The only "synergy" that has come from AMD-Xilinx is that AMD | took a (relatively simple) DSP for machine learning that Xilinx | had built and put it into their newer CPU lines. That's still | better than Intel-Altera, which basically didn't integrate at | all, despite having grandiose plans. | kjs3 wrote: | _AMD 's purchase of Xilinx made less sense because AMD is | fabless_ | | Xilinx was fabless before the acquisition. I'm missing how that | made less sense. | | _with all the zero interest rate money floating around_ | | Hehe...as one of my finance-world pals said: "everyone's doing | M&As like drunken sailors". | 0x457 wrote: | Well, it "made" sense for intel because this allowed Altera | to switch to Intel fabs. Not saying it's a huge strategic | value or anything, just saying that's the difference. | kjs3 wrote: | Altera switched to Intel fabs from TSMC in 2013, so yeah it | makes sense for Intel to pick up the folks you don't have | to do much to integrate into your manufacturing process. | But AMD and Xilinx both already being fabless meant it was | a wash...no process integration to be done. | 1-6 wrote: | I hope Nvidia comes along and scoops up Altera. That will show | Intel a lesson. | ak217 wrote: | Yeah, it's interesting to compare Nvidia's strategy to Intel's. | I'm sure there are quite a few Nvidia projects that have been | cancelled or even acquisitions liquidated, but they all seem to | be small. Every significant part of Nvidia that I can remember | is something they are committed to, sometimes over multiple | decades, even when the market is not there and sales are near | zero. This seems to come from actually having a consistent, | stable long-term vision and buy-in all the way up to Jensen | Huang serving as a driving force behind acquisitions and | projects, unlike Intel where the driving force seems to be bean | counting and market domination related. | | To give credit to Pat Gelsinger, his stated goal is to shed | non-essential units and refocus on the fundamentals. But I'm | not sure how well that's going. | no_wizard wrote: | Part of this I'm sure is management style and culture (Nvidia | famously has a very strong get-it-done-at-all-costs culture | for example) but I think the reason these ventures were | persisted is because the founder (Jensen Huang) is still CEO | and controls the majority of the company. | | Intel now, without Andy Grove and the founders steering the | culture[0] it seems there is no-one who is able to steer this | ship past quarterly results anymore. | | AMD had suffered decades of mishaps and near bankruptcy to | finally find competent leadership in recent years which | turned the company. | | [0]: After all, one core virtue that Andy Grove had was "only | the paranoid survive", which in context, was all about never | getting too comfortable with your market position. Granted, | they also engaged in many illegal practices as essentially a | monopoly in the 90s. I guess they lost his second virtue of | _Competitive Mindset_ which as Grove saw it _viewed | competition as the key driver of innovation and progress_. | This was out the window by then, though I don 't know if he | was effectively leading the company by 96. He stepped down in | 98, but he had health problems before that - eventually | diagnosed with prostate cancer. This is all to say that | Intel's culture is a mixed bag of unhealthy paranoia and | being used to being #1 in their categories, which is all | incoming executives and major shareholders see when they | think Intel. There's no-one to faithfully steer the ship back | to clear waters it seems | soulbadguy wrote: | I think you might be attributing too much agency to the CEO | and management of both companies. IMO there are a lot of | external factors too. | | It might be the case that Andy Grove "only the paranoid | survive" was really a factor. Or simply the fact that under | Andy, the tech sector was really different : X86-64 market | was the dominant and growing platform of compute, they | pretty much had a strong set of patent around X86-64. The | initial military and gov contract give them enough run away | money to build a moat around having a fab. | | Also let's not forget despite all those advantage, intel | was also fine 3 or time, for more than a billion each time | for very anti competitive practice. It's unclear to me that | intel was never the great company they think they are. | | Fast forward to now, X86-64 is no longer growing as fast, | having a fab is no longer a moat with TSCM around (might | even be seen as a liability), the WinTel monopoly doesn't | have strength he had...etc... etc... The game is just much | harder now. | | For nvidia, they had a triple boom market with first PC | gaming, then crypto,the now AI... Hard to distinguish CEO | performance from market performance. | no_wizard wrote: | Companies tend to graft their culture from the founders | establishing it with early employees and outgrowth from | there, it does say _something_. | | Now the market moving to mobile ARM chipsets definitely | damaged Intel, though Intel had _years_ to come up with a | viable alternative and ways to re-capitalize its fab | infrastructure etc. They let their own moat run dry. | Theories abound as to why, and since I 'm not CEO nor do | I have access to Intels internal records, I can only | speculate, however from available information I've been | able to read, it seems that they became _more_ risk | adverse as time has gone on, due in part, because there | is immense pressure for any _new thing_ to quickly | produce margins similar to that of their main CPU | business. Which of course, is a conundrum, as barring any | major breakthrough that puts them well and above any | alternative, these sorts of things to diversify into | strong businesses often take alot of time. | | Compare that to Nvidia: the market dynamics is what is | allowing them to capitalize on 2 booms that ultimately | leveraged GPUs for compute over traditional processors, | but the _foundation_ for all this was laid in ~2007 (CUDA | first came out then) and Nvidia had to both improve CUDA | over time and invested in research around parallelized | computing. If were not for their long horizon investment | in this, they would not have been primed to take | advantage of both crypto and AI. | | I'd argue that gaming while lucrative no doubt, is | comparatively modest to what they are raking in over | crypto and AI, all the while still investing in CPU / GPU | integrated chipsets[0] and CPUs tailored to massive data | center requirements around AI specifically[1], all of | which they wouldn't be able to really take advantage of | it not for their long term investment in things like CUDA | and parallelized computing. I suspect quite strongly that | without the founder at the helm, there would have been | little internal cover to continue investing in these | things at the rate Nvidia was prior to the boom times. | | [0]: https://www.reuters.com/technology/nvidia-make-arm- | based-pc-... | | [1]: https://www.nvidia.com/en-us/data-center/grace-cpu/ | jonnycoder wrote: | Every year Intel announces a new failure, and every year I feel | more ashamed for having Intel as the bulk of my software | engineering experience on my resume. I saw the signs when I was | still there and attending quarterly updates. It felt like in one | breath they admitted to missing the mobile boat and in another | they said only gamers need dedicated/discrete powerful gpus. | jvanderbot wrote: | Intel can and does produce phenomenal technical products. Their | compilers and HPC toolsets have done enormous good. Their chips | are second to none (or perhaps tied for first, depending on who | you ask). They are not a shameful company. They are just | _yesterday 's giant_, and as such, are due for a Microsoft-like | reinvention. | | Intel is my daily driver for gaming, due to it's incredible | (but sadly cancelled) Extreme Compute Element form factor with | its incredible customization. | | What they're struggling with is adapting to a changing world | and to find new business centers. But this is a | hardware/software world built on Intel, even if it changes | daily. | hedora wrote: | Which intel chips are second to none, and by what metric? | jvanderbot wrote: | Trying to be measured here: Depending on who you ask, | you'll find that for gaming, AMD or Intel alternate pretty | regularly on technical benchmarks. At the moment, AMD is | most power efficient, on average, and can eek out higher | overall benchmark performance (as of late 2023). Throughout | 2023, AMD and Intel were both listed as "best" on Toms | Hardware, PC Mag, etc etc | | https://www.pcmag.com/picks/the-best-cpus-for-gaming | | For server marketshare, intel vastly dominates (80%?). For | gaming marketshare, Intel still has 60-70% the marketshare. | | Even if it's a mixed bag, if given an Intel (or AMD) | processor, there's no way I would throw it in the trash and | if a current-gen chip, you can find things that either are | better than the other with. | | Both are top-tier accomplishments. AMD has a signifigant | fan-following though, which tends to distort the story a | little. | greenknight wrote: | > Their chips are second to none (or perhaps tied for | first, depending on who you ask). | | > For server marketshare, intel vastly dominates (80%?). | For gaming marketshare, Intel still has 60-70% the | marketshare. | | Marketshare is one thing, but saying they are the best | but the marketshare is 60-70% is a totally different | metric. | | Its saying oh Toyota has the best cars because they ship | the most... but they cant be as good as a Lambo / Ferrari | because they dont ship the most. | | In terms of server cpus, if you are looking at raw | performance per socket, intel is no where near amd -- | https://www.phoronix.com/review/intel-xeon- | platinum-8592/10 . This is based on Emerald Rapids / | Bergamo. | | In terms of gaming cpus, they definetly are neck and neck | with performance. | | But yes you are right, Intels chips arent something you | would throw away in the trash. They are good chips. They | need to stick around. Its just that you cant say | marketshare makes them the best. | jvanderbot wrote: | I didn't say marketshare makes them the best. I said | there were lots of metrics, that on any given metric, AMD | and Intel trade first place spots frequently, and | marketshare implies there's something special still about | Intel. | kouteiheika wrote: | > Their chips are second to none (or perhaps tied for first, | depending on who you ask). | | No. They're definitely behind AMD. AMD has significantly | better power efficiency[1], better performance in gaming for | most of the titles[2], and vastly more cores and better | multithreaded performance (with Threadrippers[3]). | | [1]: https://gamersnexus.net/megacharts/cpu-power | | [2]: https://gamersnexus.net/cpus/intels-300w-core-i9-14900k- | cpu-... | | [3]: https://gamersnexus.net/cpus/best-cpus-2023-intel-vs- | amd-gam... | kyrra wrote: | I would argue that AMD is not the reason they are more | power efficient, it's TSMC. | | Intel bet on a fabrication process that did not succeed. I | will bet in the next 2 years Intel will be on par with AMD | again. | soulbadguy wrote: | > I would argue that AMD is not the reason they are more | power efficient, it's TSMC. | | Why ? | allie1 wrote: | Because the areas where Intel stumbled wasn't design, it | was their manufacturing being behind, something tsmc does | for AMD. | jvanderbot wrote: | > Based on the numbers we've given you, from our data, and | the prices we have today, the decisions for the most part | are pretty clear: The best gaming CPU is the 7800X3D | (that's an objective fact), the most efficient part is the | 7980X, the 5800X3D is the best upgrade path, and Intel | makes the strongest showing in the i5-13600K or 14600K | (whichever is cheaper) for a balanced build, or the 12100F | for an ultra-budget build. | | Compare to toms' hardware 2023 listing earlier in the year. | | https://www.tomshardware.com/reviews/best-cpus,3986.html | Category | Winner | Alternate Overall Best CPU for | Gaming: Intel Core i5-13400 (Buy) [More] AMD Ryzen 5 7600 | (Buy) | Ryzen 5 5600X3D High Performance Value Best | CPU for Gaming: AMD Ryzen 7 7800X3D (Buy) [More] Intel Core | i7-14700K (Buy) | Ryzen 7 5800X3D (Buy) Highest | Performance Best CPU for Gaming: AMD Ryzen 9 7950X3D (Buy) | [More] Intel Core i9-13900K (Buy) Mid-Range Best | CPU for Gaming: Intel Core i5-13600K (Buy) [More] AMD Ryzen | 5 7600X (Buy) Budget Best CPU for Gaming: Intel | Core i3-12100F (Buy) [More] AMD Ryzen 5 5600 (Buy) | Entry-Level Best CPU for Gaming: AMD Ryzen 5 5600G (Buy) | | (summarizing: Intel best CPU, Best mid and budget entries, | so winner in 3/5 categories. And what does it mean to give | AMD "Best high performance" and "Highest performance?"). | | This is a neck-and-neck race, with AMD a fan favorite, and | year over year changes in leadership. No way is Intel out | of this as a viable competitor of equal-ish standing. I | don't mean to denegrate AMD in my original post, but I do | mean to say that Intel is still producing top-tier results. | | https://www.tomshardware.com/news/amd-and-intel-cpu- | market-s... In particular, Intel still has a 5x lead on AMD | in server market, which, of course, is a pretty big market. | soulbadguy wrote: | > This is a neck-and-neck race, with AMD a fan favorite, | and year over year changes in leadership. | | > (summarizing: Intel best CPU, Best mid and budget | entries, so winner in 3/5 categories. And what does it | mean to give AMD "Best high performance" and "Highest | performance?"). | | Intel as much more brand recognition than AMD. So calling | AMD fan favorite is kinda strange to me. There are only | two x86-64 chip designer right now, and from a pure | market strategy it doesn't make sense for AMD to offer | much more value to customer than what they can get from | the only competition in town. If you want to really | understand how intel is strugling you have to did into | the execution speed of both companies, the profit margin | on each SKU etc... etc... | | Intel is only neck and neck with AMD if you don't look at | things like power efficiency and how much money they are | actually making per chip. | | > In particular, Intel still has a 5x lead on AMD in | server market, which, of course, is a pretty big market. | | This a function of market inertia more than anything. | From a technical perspective AMD Zen 4 workstation and | server offering seem to be much better these days. | smolder wrote: | Toms has always had a _slight_ bias towards Intel in | their assessments, I think. Their picks are pretty | reasonable, but the "highest performance for gaming" | category with the 7950X3D and 13900k doesn't make a lot | of sense to me, when the 7800X3D which won the "High | perf" category beats the 13900k significantly in most | games benchmarks, while drawing much less power. Intel | has some great value offerings, but the 7800X3D is the | real champ in gaming right now. The fact they have Intel | winning 3/5 of their categories seems like a | demonstration of their subtle bias. Other sources like | Anandtech have historically done a better job of neutral | reporting. | J_Shelby_J wrote: | I've been on the extreme of maxing out fps in games for a | decade. | | Gamersnexus is great! But they and the other YouTube | benchmarkers don't tell the whole story. when you are | playing games that are bottlenecked by single thread | performance single core performance and therefore memory | speed becomes important. This is very difficult to | benchmark as most of those games where it matters (CoD | warzone) don't have an easy way to benchmark. When I was | testing my OC I literally would change the ram frequency in | the bios, and drop into a pubg match, and screen record my | fps and manually add it to a spreadsheet. | | So take those benchmarks and then add some performance | gains from memory overclocking that I (and others) been | able to get meaningful gains in fps beyond just | overclocking the cpu. 5-10% more fps is not amazing, but | it's definitely worth it to me for games like warzone or | valorant. | | AMD's support for high end ram has always lagged. Happy to | use amd for non-gaming, but if you're trying to build "the | best gaming rig" I'm not sure AMD can take that crown. And | until people are willing to benchmark single threaded | bottleneck games like warzone with different ram speeds, I | doubt this argument will be settled. | bee_rider wrote: | Intel's engineering is pretty solid, I think anyone who looks | down on Intel's engineers because they "only" managed to | overcome their management's wasteful dithering for... like... | 30 years is not worth working for. | kjs3 wrote: | That's a pretty personal take; don't take that shame on | yourself. Stick to the usual, comfortingly impersonal | "Obviously, Intel is dead and buried and all their management | sucks and they ruined the world" whenever Intel has announced a | setback any time in the last 50 years. No one has a blanket "we | don't hire losers from Intel" policy; it's not like you worked | at CA or Oracle. | boshalfoshal wrote: | These sound largely like management/product direction issues | and is not really indicitive of bad engineering quality. I | wouldn't be concerned about the quality of your resume. | trynumber9 wrote: | Intel bought Altera in 2015 when it still thought 10nm would be | on time and an advanced node. That did not work out. The idea was | to get a better FPGA and have more customers to justify fab build | out expenses. Gelsinger more recently said he does not want to | force products to be on Intel fabrication. Use Intel processes | where it makes sense. No reason to push Altera FPGA to Intel | 10/7/4. No reason to push NICs to Intel 10/7/4. And so on. | jasoneckert wrote: | The first thing this reminded me of was when Intel got rid of | StrongARM/XScale because they didn't think it would amount to | much in the long run. Hopefully they don't regret this particular | spinoff in the future. | reachableceo wrote: | One would presume Intel will get a decent chunk of the stock in | any IPO and capture the upside value. | | That does seem to be how these kind of deals are usually | structured. Spinco is 60% owned by the parent or wherever. | lawlessone wrote: | Ten years from now.. | | "Intel should have dominated this space but Xilinix etc got | lucky" | somethoughts wrote: | It'd be interesting if some of the funds from the sale will be | used for AI software development to provide a better coordinated | response to CUDA. | mardifoufs wrote: | How's the FPGA market at the moment? Has Altera been able to keep | up with Xilinx (or vice versa) under Intel ownership? | dboreham wrote: | FPGAs have never made sense. They're way too expensive to use in | volume. There's no practical use case for "cool, I can reprogram | the chip in the field to implement different functionality". | Nobody has figured out how to usefully integrate them with a CPU | to make a low-volume SOC. CPUs became so fast that most | applications don't need customer hardware. Regular gate arrays | are cheaper and faster above minimal volume. | | They seem to only have been useful for prototyping and military | applications (low volume and infinite budget). | vatys wrote: | I see them used in pro/prosumer audio equipment, synthesizers, | and effects, which is relatively low volume and medium-to-high | budget. FPGAs (and CPLDs, uC+AFE, etc) are great for these | applications because they have great capabilities you might | otherwise need a pile of discrete components or a custom chip | for, but it doesn't make sense to design fully custom silicon | if you're only ever going to sell about 50-500 of something. | | So sure, prototyping and military, but there are other uses as | well. But none of them are super high-volume because once | you're selling millions of something you should be designing | your own chips. | aleph_minus_one wrote: | > CPUs became so fast that most applications don't need | customer hardware. | | When complicated realtime signal processing is to be done, | FPGAs shine - in particular if there exists no DSP that is | competitive for the task. | Bluebirt wrote: | Consumer application and FPGAs are an oxymoron in itself. FPGAs | are used in applications requiring special interfaces, special | computing units or other custom requirements. If there is | enough demand, SoCs are developed for these applications, but | this is only useful in mid to high volume production. Areas | like the ones you gave and many more are making heavy use of | FPGAs. I work in medical for example. We are using custom | designed chips for special detection purposes. But when it | comes to data processing and interfacing with computers, we use | FPGAs. | crotchfire wrote: | The problem is that FPGA companies are really CAD tool | companies who see their chips as copy-protection/payment- | assurance schemes for their software. | | Unfortunately their CAD tools suck, but that's beside the | point. | soulbadguy wrote: | Large acquisition rarely seems to pan out well in the tech | sector. Especially when big companies try to acquire their way | into an adjacent market. | | Also some company seems to be significantly worst than other, | MSFT/microsoft/Dell come to mind. My suspicion is those type of | acquisition are mainly driven by C/executive level employee as | way to hide the real struggle of the company. | | Is there a report analyzing bit tech acquisition say for the last | 30 years, and they economical impact ? That would be an | interesting read. | | Maybe it's time for a new form of regulation around acquisitions | Kon-Peki wrote: | I take exception to the usage of the word "spinoff". Intel is | selling a portion of Altera. If this was a true spinoff, Intel | shareholders would get shares in the new entity. | | Intel needs the cash, so this is understandable. ___________________________________________________________________ (page generated 2023-12-21 23:00 UTC)