[HN Gopher] Will Intel's AXG division survive Pat Gelsinger's axe?
       ___________________________________________________________________
        
       Will Intel's AXG division survive Pat Gelsinger's axe?
        
       Author : oumua_don17
       Score  : 54 points
       Date   : 2022-08-09 18:05 UTC (4 hours ago)
        
 (HTM) web link (www.jonpeddie.com)
 (TXT) w3m dump (www.jonpeddie.com)
        
       | rektide wrote:
       | None of the other items Intel has let go of are an existential
       | must have. GPUs are a must have in any computer big or small. Raw
       | number crunching is core to the very profitable HPC market.
       | Players like Nvidia have fully custom interlinks that threaten to
       | dry up even the CPU sales Intel still gets in this market. GPU is
       | a focus on mobile where expectations have been much raised.
       | 
       | This group is supposedly 6 years old. But reciprocally, the other
       | players have been pouring money in over decades. GPUs are
       | fantastically complicated systems. Getting started here is
       | enormously challenging, with vast demands. Just shipping is a
       | huge accomplishment. Time to grow into it & adjust is necessary.
       | It's such a huge challenge, and I really hope AXG is given the
       | time, resources, iterations, & access to fabs it'll take to get
       | up to speed.
        
         | khitchdee wrote:
         | Intel should double down on its core business which is selling
         | x86 chips for Windows PCs. If it loses to AMD+TSMC here,
         | there's nowhere left to hide. IFS is very long term and a new
         | business opportunity in a market where Intel is not the leader.
         | x86 for Windows is their bread and butter. They need to throw
         | everything else out and downsize to just excelling at what
         | they've always done. My guess is AXG is a separate division
         | than the division that makes the graphics for their core
         | microprocessors.
        
           | mschuster91 wrote:
           | > Intel should double down on its core business which is
           | selling x86 chips for Windows PCs. If it loses to AMD+TSMC
           | here, there's nowhere left to hide. IFS is very long term and
           | a new business opportunity in a market where Intel is not the
           | leader.
           | 
           | The very second someone else but Apple brings a competitive
           | ARM desktop CPU to the market, it's game over for Intel.
           | x86_64 literally _cannot_ compete with modern ARM designs
           | simply because of how much utter garbage from about thirty
           | years worth of history it has accumulated and absolutely
           | needs to support in the future because even the boot process
           | still requires all that crap, whereas ARM was never shy about
           | cutting out stuff and breaking backwards compatibility to
           | stay performant.
           | 
           | The only luck that Intel has at the moment is that Samsung
           | and Qualcomm are dumpster fires - Samsung has enough problems
           | getting a phone to run at a halfway decent performance with
           | their Exynos line and Qualcomm managed to _completely_ botch
           | their exclusive deal with Microsoft [1] (hardly surprising to
           | anyone who has ever had the misfortune to have to work with
           | their crap). A small startup that is not bound by ages of
           | legacy and corporate red tape should be able to complete such
           | a project - Annapurna Labs have proven it 's possible to
           | break into the server ARM CPU market well enough to get
           | acquired by Amazon.
           | 
           | [1] https://www.xda-developers.com/qualcomm-exclusivity-deal-
           | mic...
        
             | khitchdee wrote:
             | Apple will always be niche because they are too expensive.
             | ARM based PC client processors accounted for about 9% of
             | the total market. But I don't expect that number will grow
             | a lot higher due to cost of Macs. Also, you can't bring out
             | a desktop PC based on Linux because that a very slowly
             | growing market segment and MacOS is proprietary. Don't
             | think Apple will start another PowerComputing type scenario
             | where they open their platform to others. It's not a matter
             | of performance for this reason (price). So, on Windows,
             | assuming Macs don't take away any more market share from
             | Windows, they need to beat AMD+TSMC.
        
               | mschuster91 wrote:
               | > So, on Windows, assuming Macs don't take away any more
               | market share from Windows, they need to beat AMD+TSMC.
               | 
               | No. All it needs is
               | 
               | - Microsoft and Qualcomm breaking their unholy and IMHO
               | questionably legal alliance
               | 
               | - an ARM CPU vendor willing to do the same as Apple did
               | and add support for accelerating translation of x86 code
               | (IIRC, memory access models/barriers are done differently
               | between x86 and ARM, and Apple simply extended their
               | cores to be able to use the same memory access/barrier
               | model as x86 on translated-x86 threads)
               | 
               | - an ARM CPU vendor willing to implement basic
               | functionality like PCIe actually according to spec - even
               | the Raspberry Pi which is the closest you can get to a
               | mass market general-purpose ARM computer has that broken
               | [1]
               | 
               | - someone (tm) willing to define a common standard of
               | bootup sequence/standard feature set. Might be possible
               | that UEFI fills the role; the current ARM bootloaders are
               | a hot mess compared to the old and tried BIOS/boot sector
               | x86 approach, and most (!) ARM CPUs/BSPs aren't exactly
               | built with "the hardware attached to the chips may change
               | at will" in mind.
               | 
               | Rosetta isn't patented to my knowledge, absolutely
               | nothing is stopping Microsoft from doing the same as part
               | of Windows.
               | 
               | [1] https://www.hackster.io/news/jeff-geerling-shows-off-
               | an-amd-...
        
               | khitchdee wrote:
               | If you translate, you do face performance issues. Apple
               | or some other vendor cannot make an ARM chip so fast at a
               | competitive cost that beats an x86 chip in emulation
               | mode. The underlying acceleration techniques for both ARM
               | and x86 are the same.
        
               | mschuster91 wrote:
               | > Apple or some other vendor cannot make an ARM chip so
               | fast at a competitive cost that beats an x86 chip in
               | emulation mode.
               | 
               | This reminds me of Iron Man 1... "Tony Stark was able to
               | build this in a cave! With a box of scraps! - Well, I'm
               | sorry. I'm not Tony Stark."
               | 
               | Apple has managed to pull it off so well that the M1
               | blasted an _i9_ to pieces [1]. The M1 is just so damn
               | well more performant than an Intel i9 that the 20%
               | performance loss compared to native code didn 't matter.
               | 
               | [1] https://www.macrumors.com/2020/11/15/m1-chip-
               | emulating-x86-b...
        
               | WkndTriathlete wrote:
               | And for 99.9999% of users the M1 performance benchmarks
               | vs. i9 _don't matter one bit._
               | 
               | The use case for the vast majority of laptops include
               | I/O- and memory-bound applications. Very few CPU-bound
               | applications are run on consumer laptops, or even
               | corporate laptops, for the most part. CPU-bound
               | applications should be getting run on ARM or GPU clusters
               | in the cloud.
               | 
               | The use case for an M1 in laptops is the _power_
               | benchmarks vs. an i9.
        
               | mschuster91 wrote:
               | > The use case for the vast majority of laptops include
               | I/O- and memory-bound applications.
               | 
               | Where the M1 just blows anything desktop-Intel out of the
               | water, partially because they integrate a lot of stuff
               | directly on the SoC, partially because they place stuff
               | like RAM or persistent storage _extremely_ close to the
               | SoC whereas on desktop-Intel RAM, storage and peripheral
               | controllers are all dedicated chips.
               | 
               | The downside is obviously that you can't get more than
               | 16GB RAM with an M1 and 24GB RAM with the new M2's and
               | you cannot upgrade either memory at all without a high-
               | risk soldering job [1]... but given that Apple has the
               | persistent storage so closely attached to the SoC to swap
               | around, it doesn't matter all that much.
               | 
               | [1] https://www.macrumors.com/2021/04/06/m1-mac-ram-and-
               | ssd-upgr...
        
               | khitchdee wrote:
               | That performance difference is not due to architecture
               | but process technology. Intel is on Intel' best node and
               | Apple is on TSMC's.
        
               | spookie wrote:
               | No. The architecture is the biggest factor. Damn, even
               | Jim Keller talks about how most programs use a very small
               | subset of instructions. It isn't like RISC makes
               | miracles, but sure helps them when your power budget is
               | small.
        
               | frostburg wrote:
               | Windows runs on ARM, too (with issues, but those are also
               | related to the questionable performance of the SoCs).
        
               | spookie wrote:
               | > Also, you can't bring out a desktop PC based on Linux
               | because that a very slowly growing market segment
               | 
               | Wouldn't bet on that one
        
             | acomjean wrote:
             | The fact you can run old software on any PC still is a
             | pretty tremendous feature. Apple kicked of 32 bit apps from
             | Mac and iPhones, its user accept that, business PC users
             | might not. I have a number of IOS games I enjoyed that are
             | no longer usable (yeah I could have stopped upgrading my i
             | device, but the security problems make that not a viable
             | option)
             | 
             | Conventional wisdom in the mid 90s was powerpc (RISC) would
             | eventually be better the x86, but it never happened. They
             | worked around the issues. And Microsoft eventually made an
             | OS that wasn't a crash fest (I'm looking at you windows ME)
             | 
             | Also no one can afford to be on TSMCs best node when Apple
             | buys all the production. Apple's been exclusive on the best
             | node for at least a couple years now. Even AMD isn't using
             | TSMCs best node yet.
        
               | mschuster91 wrote:
               | > Apple kicked of 32 bit apps from Mac and iPhones, its
               | user accept that, business PC users might not.
               | 
               | Yeah, but that was (at least on the Mac) not a technical
               | requirement, they just didn't want to carry around the
               | kernel-side support any more. IIRC it didn't take long
               | until WINE/Crossover figured out a workaround to run old
               | 32-bit Windows apps on modern Macs.
               | 
               | > Also no one can afford to be on TSMCs best node when
               | Apple buys all the production. Apple's been exclusive on
               | the best node for at least a couple years now. Even AMD
               | isn't using TSMCs best node yet.
               | 
               | Samsung has their own competitive fab process, but they
               | still have yield issues [1]. It's not like TSMC has a
               | monopoly by default.
               | 
               | [1] https://www.gsmarena.com/samsung_claims_that_yields_f
               | rom_its...
        
             | PaulHoule wrote:
             | There is just one serious problem with the x86 architecture
             | and that is the difficulty of decoding instructions which
             | could be anywhere between 1 and 14 bytes. It's not hard to
             | design a quick instruction decoder but hard to design a
             | quick instruction decoder that is power efficient.
             | 
             | ... Oh yeah, and there are the junkware features like SGX
             | and TSX and also the long legacy of pernicious segmentation
             | that means Intel is always playing with a hand tied behind
             | its back, for instance, the new laptop chips that should
             | support AVX512 but don't because they just had to add
             | additional low performance cores.
        
               | Miraste wrote:
               | I may be showing ignorance of some important use case
               | here, but I'd put AVX512 in the "junkware feature"
               | category. It can run AI/HPC code faster than the main
               | processor, but still so much slower than a GPU as to be
               | pointless; when it does run, it lowers the frequency and
               | grinds the rest of the processor to a halt until it
               | eventually overheats anyway. Maybe, maybe it has a place
               | in their big desktop cores, but why would anyone want it
               | in a laptop? You can't even use it in a modern ultrabook-
               | style casing without a thermal shutdown and probably
               | first-degree burns, without even mentioning the battery
               | drain.
        
               | PaulHoule wrote:
               | Yes and no.
               | 
               | If you have all the cores running hard the power
               | consumption goes up a lot but if it is just one core it
               | won't go up too much. If worse come to worse you can
               | throttle the clock.
               | 
               | In general it's a big problem with SIMD instructions that
               | they aren't compatible across generations of
               | microprocessors. It's not a problem for a company like
               | Facebook that buys 20,000 of the same server but normal
               | firms avoid using SIMD entirely or they use SIMD that is
               | many years out of date. You see strange things like
               | Safari not supporting WebP images on a 2013 Mac while
               | Firefox supports them just fine because Apple wants to
               | use SIMD acceleration and they'd be happier if you
               | replaced you 2013 Mac with a new one.
               | 
               | I worked on a semantic search engine that used an
               | autoencoder neural network that was made just before GPU
               | neural networks hit it big and we wrote the core of our
               | implementation in assembly language using one particular
               | version of AVX. We had to do all the derivatives by hand
               | and code them up in assembly language.
               | 
               | By the time the product shipped we bought new servers
               | that supported a new version of AVX that might have run
               | twice as fast but we had no intention of rewriting that
               | code and testing it.
               | 
               | Most organizations don't want to go through the hassle of
               | keeping up with the latest SIMD flavor of the month so a
               | lot of performance is just left on the table. Intel is
               | happy because their marketing materials can tell you how
               | awesome the processor is but people in real life don't
               | experience that performance.
        
             | spookie wrote:
             | China is gonna be that second someone, with RISC-V.
             | Honestly, it's kind of amazing how bad it can get for them.
        
         | ChuckNorris89 wrote:
         | _> None of the other items Intel has let go of are an
         | existential must have._
         | 
         | Except mobile ARM chips and mobile LTE modems, both of which
         | Intel sold off, and those are some of the most desirable things
         | to make right now. Just ask Quallcomm.
        
           | mrandish wrote:
           | Yep. Competent mobile ARM and LTE modem cores would be _very_
           | nice to have in their tech stack right about now. I think a
           | _credible_ (if not fully competent), GPU stack is pretty
           | essential for strategic relevance going forward. This seems
           | like something they _have_ to make work.
        
         | pjmlp wrote:
         | Intel famously used to lie on their OpenGL drivers, asserting
         | features were supported, but were actually software rendered.
         | 
         | Larrabe was shown at GDCE 2009 as if going to render the
         | competition useless, then faded away.
         | 
         | Their GPU debugger marketing sessions used to be focused on how
         | to optimise games for integrated GPUs.
         | 
         | I really don't get how they keep missing the mark versus AMD
         | and NVidia in both GPU design and developer tooling.
        
           | pinewurst wrote:
           | Don't forget Intel's world-beating i740!
        
             | Miraste wrote:
             | Ah, the i740. Because VRAM is a conspiracy cooked up by
             | 3DFX and Nvidia to take your money, and real gamers don't
             | need any.
        
             | muro wrote:
             | Anyone who bought one will not forgive them that one.
        
       | csense wrote:
       | What the heck about GPU development costs $2 billion? That seems
       | _obscenely_ expensive. Anyone who has detailed knowledge of the
       | industry care to weigh in on what they might actually spending
       | that money on?
        
         | killingtime74 wrote:
         | I think the article aludes to it being massive payroll
        
         | terafo wrote:
         | Drivers and couple attempts that they shot behind the barn(but
         | made it into silicon).
        
         | ChuckNorris89 wrote:
         | Try hiring GPU experts and see how "easy" it is. It's a very
         | small talent pool.
        
       | ksec wrote:
       | >Started in 2016 the dGPU group snatched showboater Raja Koduri
       | away from AMD
       | 
       | We are in 2022 I still dont understand why Raja is popular. One
       | of the reason why I have been extremely sceptical of Intel's GPU
       | since the very beginning. To the point I got a lot of bashing on
       | Anandtech and HN.
       | 
       | And I have been mentioning drivers as the major concern since
       | 2016. Citing PowerVR Kyro on Desktop as an example. Even pointing
       | that out on Twitter. With one of the Intel Engineers on the GPU
       | team replied something as "GPU Drivers is a solved problem".
       | 
       | I do love to be wrong. But it is increasingly looking like
       | another item to be added to my book of prediction that came true.
        
         | philjohn wrote:
         | To be fair, in the consumer space, he seemed to be holding them
         | back. In the data centre acceleration space, they're still
         | essentially using a descendant of the Vega uArch.
        
         | slimginz wrote:
         | Honestly, I'm in the same boat as you, until I realized I don't
         | know anything he's done before he became head of Radeon Group.
         | According to his Wikipedia entry:
         | 
         | > He became the director of advanced technology development at
         | ATI Technologies in 2001.[3] Following Advanced Micro Devices's
         | 2006 acquisition of ATI, he served as chief technology officer
         | for graphics at AMD until 2009. At S3 and ATI he made key
         | contributions to several generations of GPU architectures that
         | evolved from DirectX Ver 3 till Ver 11.[4] He then went to
         | Apple Inc., where he worked with graphics hardware, which
         | allowed Apple to transition to high-resolution Retina displays
         | for its Mac computers.[5]
         | 
         | So he has a history of launching some really good products but
         | I'd say it's been at least a decade since he's been involved in
         | anything industry leading.
        
         | PolCPP wrote:
         | And ironically after he left AMD started improving on the
         | drivers
        
           | allie1 wrote:
           | Very telling... the guy hasn't had a success in the last 10
           | years and still gets top roles. Seems like he can sell
           | himself well.
           | 
           | There needs to be accountability for these failures of
           | execution.
        
         | Miraste wrote:
         | I'm with you here. At AMD, Koduri released GPU after GPU that
         | was slow, terribly hot and inefficient, plagued by driver
         | problems, and overpriced. He left on terrible terms for a
         | number of reasons, but one of them was his complaint that
         | Radeon was underfunded and deprioritized, so he couldn't
         | compete with Nvidia. Here we are, six years of practically
         | unlimited funding and support from a company with a greater R&D
         | budget than AMD's entire 2016 gross...
         | 
         | and he's done the exact same thing.
        
           | ChuckNorris89 wrote:
           | _> and he's done the exact same thing._
           | 
           | That's how failing upwards works in this racket. As long as
           | you have a great looking resume at a few giants in the
           | industry with some fancy titles to boot, you're set for life
           | regardless of how incompetent you are.
        
       | Scramblejams wrote:
       | _Intel is now facing a much stronger AMD and Nvidia, plus six
       | start-ups_
       | 
       | Who are the six startups? It mentions four are in China, and two
       | in the US.
        
         | killingtime74 wrote:
         | Vaporware unless we can buy their cards anyway
        
       | Kon-Peki wrote:
       | Apple Silicon has absolutely proven that an integrated gpGPU is a
       | must on modern consumer chips.
       | 
       | Intel has to slog through this. Forget about CUDA support - Apple
       | is going to kill that monopoly for them.
        
       | Aromasin wrote:
       | I'm still convinced by Intel's and Pat's XPU strategy. Having a
       | well rounded portfolio of CPUs, GPUs, SoCs, and FPGAs that all
       | work flawlessly, alongside a comprehensive software environment
       | to make cross-platform development easier (their OneAPI stack) is
       | a dream for any OEM/ODM; provided it works.
       | 
       | Their issue at the moment isn't strategy. It's execution. Axing
       | their GPU division would only hurt their current plan, and do
       | nothing to fix the the systematic problem that they're missing
       | deadline's and shipping incomplete products. From the outside
       | looking in, it seems like there's some fat that needs trimming
       | and people aren't pulling their weight. If they can scale back to
       | efficient team and org sizes, cut the side projects, and focus on
       | excellent software and hardware validation, I can see them
       | pulling this off and Pat being lauded as a hero.
        
         | klelatti wrote:
         | You may well be right but if each of the individual components
         | of that portfolio is worse than the competition then it will be
         | an uphill struggle.
        
         | PaulHoule wrote:
         | People I know who develop high performance systems are so
         | indifferent to OpenCL it's hard for me to picture what could
         | make people develop for OpenCL.
         | 
         | There is such a long term history of failure here that somebody
         | has to make a very strong case that the next time is going to
         | be different and I've never seen anyone at Intel try that or
         | even recognize that history of failure.
        
           | Miraste wrote:
           | It's unfortunate, but the reality is it would take an act of
           | god to unseat CUDA at this point.
        
             | nicoburns wrote:
             | Could Intel just implement CUDA support? That would
             | certainly be a huge task, but Intel aren't have the
             | resources for that.
        
               | paulmd wrote:
               | The GPU Ocelot project did exactly that at one point.
               | It's probably a legal minefield for an organized
               | corporate entity though.
               | 
               | https://gpuocelot.gatech.edu/
        
               | Aromasin wrote:
               | Intel doesn't implement CUDA, but they're working on a
               | CUDA to DPC++ (their parallel programming language)
               | migration tool to make the jump less painful. I've tested
               | some of the NVIDIA samples with it and it seems to do the
               | vast majority of the grunt work, but there are still some
               | sections where refactoring is required. You can find it
               | in the OneAPI base toolkit.
        
               | TomVDB wrote:
               | One of the big attractions of CUDA is not just the core
               | (programming, API, compiler, development tool) but also
               | the libraries (cuDNN, FFT, cuSignal, ... It's a very long
               | list). These libraries are usually not open source, and
               | can thus not be recompiled even if you have core CUDA
               | support.
        
             | VHRanger wrote:
             | Vulkan compute might, over time, take root.
             | 
             | The fact that a similar API can be used for training on
             | servers, inference from laptops to phones, is an appealing
             | proposition
             | 
             | Best part: most devices have decent vulkan drivers. Unlike
             | openCL.
        
             | bobajeff wrote:
             | That or they could just open their GPU cores to direct
             | programming instead of locking it behind some API.
        
               | amluto wrote:
               | This depends on the degree to which the underlying
               | hardware remains compatible across generations. nVidia
               | can sell monster GPU systems today to run 5-year-old CUDA
               | code faster without any recompilation or other mandatory
               | engineering work. I don't know whether this is possible
               | with direct hardware access.
        
               | PaulHoule wrote:
               | Intel's OneAPI (rebranded OpenCL) can allegedly do this.
               | 
               | It also claims to do this across radically different
               | architectures like FPGA and all I can say to that is "I
               | find that very hard to believe"
        
               | buildbot wrote:
               | Yeah it is in practice not really true. Maybe after
               | hacking the code you can watch it compile for _days_ and
               | then perform probably worse than a GPU without extreme
               | tuning, which removes the point anyway.
        
               | wmf wrote:
               | That's a "did you just tell me to go fuck myself?"
               | solution. It didn't work for AMD and it won't work for
               | anyone else. Developing a CUDA-to-GPU compiler and
               | runtime stack is an immense amount of work and customers
               | or "the community" can't afford it.
               | 
               | Although for completeness I'll note that Intel's GPU
               | architecture is documented:
               | https://01.org/linuxgraphics/documentation/hardware-
               | specific...
        
             | tambourine_man wrote:
             | Unless someone beats them in the hardware front. GPUs that
             | are 30-50% faster (or just as fast but 30-50% cheaper).
             | 
             | I could see Apple and AMD, working with TSMCs latest node,
             | stepping up to the challenge.
        
               | ceeplusplus wrote:
               | No way 30-50% is enough. You need 2x at least, if not
               | more. There is a deep rooted ecosystem of CUDA apps. You
               | would need to have at least 2-3 generations of 2x lead to
               | get people to switch.
        
           | pinewurst wrote:
           | And could care even less about "oneAPI"(tm).
        
         | mrtweetyhack wrote:
        
         | khitchdee wrote:
         | Execution is about engineers. Intel can't retain their
         | engineers. Without good engineers you can't compete. So you
         | downsize and try compete only on one thing -- better x86. Pat
         | Gs return as CEO does help with engineer retention though.
        
           | ChuckNorris89 wrote:
           | _> Execution is about engineers._
           | 
           | Meh, not really. Google has some of the best engineers in the
           | world yet they fail at nearly everything that isn't related
           | to search, android, ads and e-mail.
           | 
           | Even Jensen Huang said that it's more about vision and not
           | all about execution. He said in the early days of Nvidia all
           | of their competitors at the time had engineers good enough to
           | execute, yet only they had the winning vision that enabled
           | them to make and sell exactly what the market wanted, nothing
           | more, nothing less.
        
             | R0b0t1 wrote:
             | Necessary but not sufficient is the phrase.
        
             | random314 wrote:
             | > everything that isn't related to search, android, ads and
             | e-mail.
             | 
             | And YouTube.
             | 
             | No wonder they are 10X as valuable as Intel
        
             | khitchdee wrote:
             | Google used to be search only. Now they have Android.
             | That's pretty good.
        
               | ChuckNorris89 wrote:
               | And how many other dozens of products did Google build
               | then kill off due to failure? Similar to Intel.
        
               | khitchdee wrote:
               | Intel does not have a number 2. They only own x86. They
               | have failed to diversify.
        
               | ChuckNorris89 wrote:
               | _> Intel does not have a number 2_
               | 
               | Yeah they do, Intel has the FPGA division they bought
               | from Altera. That's a huge business on its own.
               | 
               | And my point still stands: having great engineers is not
               | enough for great execution. You need great leadership
               | with a vision a-la Steve Jobs or Jensen Huang.
        
               | __d wrote:
               | They have _tried_ to diversify:
               | 
               | They have competitive NICs, although they don't seem to
               | be maintaining the lead there they once had.
               | 
               | They bought competitive network switches. These have
               | largely languished, in part because they sat on the IP
               | and then targeted it at weird niches.
               | 
               | They bought Altera. I feel like it had lost some momentum
               | vs. Xilinx, but with AMD acquiring the latter, it's
               | probably going to end up a wash.
               | 
               | The AI chips are kinda too early to tell, but at least
               | they're playing the game.
               | 
               | Overall, I think they have squandered the massive
               | advantage they had in CPUs for the last 3 decades.
        
               | touisteur wrote:
               | Oh yeah I'm waiting for their 200G NICs while NVIDIA is
               | ramping up connectx 7 with 400G on PCIe5...
               | 
               | The low end AI chips are a mess. Myriad-X can be used
               | only through openvino, and they've been closing details
               | about the internals... Keembay is... Years late?
               | 
               | Is there anything coming out of Intel these days?
               | 
               | And what happened to nervana systems, became plaidml then
               | disappeared after bought by Intel? Maxas was really great
               | and now, crickets.
               | 
               | Even profiling tools, which they used to be on top of,
               | don't seem to work well on Linux w/ the TigerLake gpu.
               | It's so painful to debug and program, they might as well
               | have not put it in...
        
               | tambourine_man wrote:
               | And that's what, 15 years old?
        
               | AnimalMuppet wrote:
               | The degree of market penetration that Android has
               | maintained for that 15 years is a pretty impressive feat,
               | both strategically and technically.
        
               | pinewurst wrote:
               | Android is still basically ads - in that it's another
               | source of info fodder for their ad machine along with
               | search.
        
             | time0ut wrote:
             | My impression is that most of Google's failures are
             | strategic. Stadia, for example, has been executed very well
             | in the technical sense. It just doesn't make a lot of sense
             | strategically. I feel every failure I can think of fits
             | this mold of great technology solving the wrong problem or
             | held back from solving the right problem.
        
           | PaulHoule wrote:
           | For the last 20 years there have always been stories in the
           | press about how Intel is laying off 10,000 engineers. It
           | never struck me as a very psychologically safe place to work.
        
             | AnimalMuppet wrote:
             | IIRC, they had (maybe still have?) a policy of laying off
             | the bottom 5%. Every year. For a company that size, it's a
             | lot of engineers.
             | 
             | But yes, that probably does not make it a psychologically
             | safe place to work...
        
           | alfalfasprout wrote:
           | It's a bit of everything but yeah, what I've seen is
           | basically that Intel has a combination of absolutely insane
           | talent working on some really hard problems and a lot of
           | mediocre talent that they churn through.
           | 
           | And the reality is that yeah, with their comp for engineers
           | so abysmally bad why would anyone really go there? Especially
           | when it's not like they're getting great WLB.
        
         | AceJohnny2 wrote:
         | > _provided it works_
         | 
         | Ay, there's the rub!
        
         | khitchdee wrote:
         | XPU is too generic a term and harmful looking at it from a
         | business perspective.
        
         | mugivarra69 wrote:
         | isnt that amd rn
        
       | tadfisher wrote:
       | > Since Q1'21 when Intel started reporting on its dGPU group,
       | known as AXG or accelerated graphics, the company has lost a
       | staggering $2.1 billion and has very little to show for it.
       | 
       | You know, except for a competitive desktop GPU. I'm actually
       | impressed that it didn't take much longer and much more money to
       | catch up with AMD and NVIDIA, given that those two were in
       | business when 3dfx was still around.
        
         | ChuckNorris89 wrote:
         | I don't know why your painting Intel as some new entry GPU
         | beginner here, but FYI, Intel had been making and selling
         | integrated GPUs for over 20 years now, to the point they pretty
         | much dominate the GPU market share in sheer numbers (most
         | desktops and laptops, even older Macs, sold in the last 10-15
         | years are most likely gonna have an Intel chip with integrated
         | graphics, regardless if it's used or not in favor of a discrete
         | GPU).
         | 
         | Sure, their iGPU offerings were never competitive for gaming or
         | complex tasks, but given it came for "free" with your CPU, it
         | was good enough for most businesses and consumers and was a
         | also a major boon during the GPU shortage where gamers who
         | build systems with AMD chips were left unable to use their PC
         | while those who went Intel could at least use their PC for some
         | productivity and entertainment until they could buy a dGPU.
         | 
         | So it's not like they had to start absolutely from scratch
         | here. In fact, Intel's latest integrated GPU architecture, Xe,
         | was so good, it was beating the integrated Vega graphics AMD
         | was shipping till the 6xxx series with RDNA2 in 2022, while
         | also killing the demand for Nvidia's low end dGPUs for desktops
         | (GT 1050) and mobile (MX350). Xe was also the first GPU on the
         | market with AV1 decode support.
         | 
         | So given this, Intel is definitely not a failure in the GPU
         | space, they're definitely doing some things right but they just
         | can't box in the ring with "Ali" yet. Anyone thinking they can
         | leapfrog Radeon and Nvidia at their first attempt would be
         | foolish. Intel should take on the losses on the GPU division
         | for a few more years and push through.
        
           | hexadec wrote:
           | > during the GPU shortage where gamers who build systems with
           | AMD chips were left unable to use their PC
           | 
           | This comment baffles me, both AMD and Intel have CPUs with
           | onboard graphics and those without. You even noted the
           | integrated graphics a sentence later.
           | 
           | If anything, this is more evidence that AMD is following the
           | Intel playbook by having that integrated CPU/ GPU
           | architecture plan.
        
             | ChuckNorris89 wrote:
             | _> This comment baffles me, both AMD and Intel have CPUs
             | with onboard graphics and those without. You even noted the
             | integrated graphics a sentence later_
             | 
             | Why does it baffle you? AMD has only been selling desktop
             | chips with integrated GPUs only for a few years now (they
             | called them APUs), and their APUs were not that stellar at
             | either the GPU or CPU part due to compromises on both
             | parts.
             | 
             | Most of the successful Ryzen chips AMD was selling for the
             | desktop were exclusively without integrated GPUs, to save
             | die space and cost, which hurt PC builders during the GPU
             | scalpocalipse, while on the other hand, Intel's almost
             | entire CPU product range for desktops had integrated GPUs
             | for over 10 years now, enabling PC builders to at least use
             | their PCs until a dGPU could be available.
             | 
             | Sure, Intel sold some CPUs without iGPUs but those were
             | very few SKUs in comparison. Similarly, but in reverse, AMD
             | also sold some Ryzen CPUs with iGPUs(APUs), but those were
             | very few SKUs as their CPUs were weaker than the non-iGPU
             | SKUs, and their outdated Vega iGPUs were pretty weak even
             | compared to Intel's Xe.
             | 
             | So that's the major difference between Intel and AMD that
             | was a game changer for many: Intel shipped most of its
             | chips with iGPUs for over a decade while AMD did not,
             | meaning you always needed to buy a dGPU, and if you
             | couldn't, like in the past ~2 years, well ... good luck,
             | your new tower PC is now an expensive door stop.
             | 
             | Still baffled?
        
             | formerly_proven wrote:
             | AMD CPUs with iGPUs are a very different product from their
             | CPUs without, regardless of what the nomenclature might
             | imply.
        
           | paulmd wrote:
           | > So it's not like they had to start absolutely from scratch
           | here
           | 
           | No, and actually in some respects that's not a good thing
           | either. Their existing iGPU driver was designed with the
           | assumption of GPU and CPU memory being pretty much fungible,
           | and with the CPU being pretty "close" in terms of latency,
           | just across the ringbus. It wasn't even PCIe attached, like
           | how AMD does it, it was directly on the ringbus like another
           | core.
           | 
           | Now you need to take that legacy codebase and refactor it to
           | have a conception of where data lives and how computation
           | needs to proceed in order to feed the GPU with the minimum
           | number of trips across the bus. Is that easier than writing a
           | clean driver from scratch, and pulling in specific bits that
           | you need? ....
           | 
           | One of their recent bugs in raytracing was literally due to
           | one line of code in an allocator that was missing a flag to
           | allocate the space in GPU memory instead of CPU, a one-line
           | change produced a 100x speedup in the raytracing performance.
           | 
           | https://www.phoronix.com/news/Intel-Vulkan-RT-100x-Improve
           | 
           | It is most likely much easier to do what AMD did and go from
           | discrete to integrated than the other way around... again,
           | they don't have a tightly-coupled design like Intel did,
           | their iGPU is literally just a pcie client that happens to be
           | on the same die.
           | 
           | (also, AMD paid the penalty like 15 years ago... there were
           | _terascale based APUs_ , and the GCN driver was developed as
           | a dGPU/iGPU hybrid architecture from day 1, they never had to
           | backport GCN itself from dGPU, that was all done in the
           | terascale days.)
        
           | buran77 wrote:
           | I'd have to agree with GP here. It wasn't at all obvious for
           | almost anyone that Intel can get so quickly to the level of
           | performance Arc offers. After decades of bottom of the barrel
           | iGPUs in a span of 2-3 years they went to decent iGPUs, and
           | then decent dGPUs (according to those who touched them).
           | That's quite an achievement in a market that all but
           | solidified around the 1.5 incumbents.
        
             | jjoonathan wrote:
             | > they went from decades of bottom of the barrel iGPUs
             | 
             | Hah, I remember the time I spent a week dodging the
             | software renderer to figure out why my shader wasn't
             | working on an intel iGPU -- turns out that someone at intel
             | decided to implement "float" with an 8 bit float. Not 8
             | byte, 8 bit.
             | 
             | "Sure, we support floating point, what's the problem?"
        
         | bacro wrote:
         | I have not seen any desktop GPU that is competitive with Nvidia
         | or AMD yet. What surprises me is how naive Intel seems to be.
         | It is extremely hard to enter the GPU market right now, so it
         | should expect a lot of years of losses. Hopefully they will
         | continue and fix their drivers and hardware problems so we can
         | have more competition. For Alchemist, if they do not price it
         | extremely aggressively ( like 200$ for a A770 ) they will not
         | get much marketshare, I am afraid.
        
       | windowsrookie wrote:
       | Intel quit making ARM cpus right before smartphones took off.
       | They sold off their LTE modem division essentially giving
       | Qualcomm a monopoly. Now they might give up on GPUs.
       | 
       | Intel is way too quick to give up. Imagine if they had continued
       | to make ARM CPUs. Apple may have never pursued making their own
       | chips. Intel CPU's + modems could be competing with Qualcomm for
       | every mobile device sold today.
        
         | UncleOxidant wrote:
         | I think it was more hubris that caused them to stop making ARM
         | chips as opposed to "giving up". They were still in the "x86
         | everywhere" mindset back then. Remember when they were trying
         | to get x86 into phones & tablets?
        
       | spookie wrote:
       | >The drumroll never stopped, even to the point of talking about
       | code improvements in a Linux driver--Linux? That's Intel's first
       | choice?
       | 
       | This phrase does it, this guy has no clue whatsoever. Baffling.
        
       | rossdavidh wrote:
       | "Of the groups Gelsinger got rid of was Optane (started in 2017,
       | never made a profit), sold McAfee (bought in 2010, never made a
       | profit), and shut down the drone group (started in 2015, never
       | made a profit). Last year, Intel sold off its NAND business to
       | Hynix, giving up its only Chinese fab, and sold off its money-
       | losing sports group; and this year, the company shut down its
       | Russian operations. Since Gelsinger's return, Intel has dumped
       | six businesses, saving $1.5 billion in operating costs and
       | loses."
       | 
       | None of these groups are remotely as close to Intel's core (pun
       | intended) business, as this one. It is certainly true that this
       | division needs to produce better results, but that's true of the
       | company as well. Chopping this one, unless there is some plan for
       | a replacement strategy in the GPU space, would be a bad sign, as
       | it would suggest a company circling the drain/milking the
       | existing winners, unable to make any new winners.
        
       | bearjaws wrote:
       | I really don't understand how Intels drivers for Alchemist are so
       | flawed, they have integrated GPUs that run esports titles at
       | 100+fps.
       | 
       | It seemed to me scaling the existing architecture to have more
       | processing and bandwidth should have yielded a very competitive
       | GPU while reusing their existing talent pool.
       | 
       | Instead we end up with a very buggy, very specialized (needs
       | resizable bar?, DX12 works well but not other runtimes?) that
       | aren't required today for Intel graphics.
        
         | wtallis wrote:
         | One pattern that has emerged is many of the performance
         | problems come from VRAM no longer being carved out of main
         | system RAM. The drivers now have to be careful about which pool
         | of memory they allocate from, and for CPU access to VRAM
         | without ReBAR the access patterns matter a lot more than they
         | used to.
        
           | wmf wrote:
           | It's amazing that fixing that one thing takes over three
           | years.
        
         | terafo wrote:
         | DX12 works well cause it was designed to make driver
         | development easier. Last gen APIs required specific
         | optimizations in drivers for every single game in order for it
         | to run well. And as far as I heard the problem isn't in
         | drivers, problem is in hardware scheduler that have some kind
         | of limitations that become more apparent with scale. It might
         | be impossible to fix this without retape.
        
       | khitchdee wrote:
       | If I were Pat Gelsinger, I would keep PC Client, DataCenter and
       | FPGA(Altera). Get rid of everything else.
        
       | PaulsWallet wrote:
       | I have to believe that Intel's GPU division will be fine just
       | because I refuse to believe that any executive over at Intel is
       | short sighted enough to believe that Intel was going to leap frog
       | AMD and Nvidia or even make a profit within the next 3-5 years.
       | It took AMD years to get it right with Ryzen and AMD still hasn't
       | had a "Ryzen moment" with their GPU. The fact that Intel's GPU
       | can even compete with AMD and Nvidia is a feat of engineering
       | magic. Intel should just take the L and make it a loss leader and
       | grind it out for the long haul.
        
         | Miraste wrote:
         | I would say RDNA/RDNA2 has delivered Ryzen-level improvements.
         | It didn't result in Ryzen-level fanfare because Nvidia improves
         | their GPUs every product cycle instead of sitting still like
         | Intel was doing with processors. I don't think Intel has the
         | leadership or, after their Boeingization (getting rid of senior
         | engineers) the engineering ability to compete with Nvidia.
        
         | [deleted]
        
       | pclmulqdq wrote:
       | Intel missed a huge opportunity by killing Larrabee and the Xeon
       | Phi. They aren't going to be able to beat CUDA with the "oneAPI"
       | software layer they are trying to offer: an outgrowth of OpenCL
       | is not going to be popular with people who care about the
       | performance of their devices. The only programming API that Intel
       | has and CUDA can't beat is x86 assembly.
       | 
       | In my opinion, when they wanted to get back into graphics, they
       | should have brought back Xeon Phi, maybe doubled the vector width
       | (and added some special units), and hired some engineers from HFT
       | firms to figure out how to make it pretend to be a fast GPU.
        
       | fancyfredbot wrote:
       | Intel can make money from deep learning accelerators even if they
       | fail to get the gaming market this generation.
        
       | protomyth wrote:
       | I'm not sure how you can give up the GPU market when Intel
       | aspires to be a performance leader and must have some plan to get
       | back into mobile. They have a GPU with bad drivers right now, but
       | instead of axing the division, it would seem more appropriate to
       | get some folks who know software. The duopoly of NVIDIA and AMD
       | are a lot more vulnerable than other players in other markets
       | Intel could expand into.
        
         | CameronNemo wrote:
         | Aren't mobile devices usually iGPU only? I can't think of a
         | single mobile device with a discrete GPU. Maybe some
         | convertible tablets?
        
       | keepquestioning wrote:
       | It's clear Raja Koduri is the wrong choice.
        
         | tibbydudeza wrote:
         | Was the Arc disaster his making ???.
        
           | arcanus wrote:
           | Yes. As the executive in charge of this business unit (since
           | 2016!) he owns execution.
        
             | tibbydudeza wrote:
             | Hope it does not end up like Larrabee - really don't get
             | why Intel is competing in this market segment - Integrated
             | graphics yes but surely there is bigger and more important
             | fish to fry (Amd Zen) elsewhere ???.
        
               | PolCPP wrote:
               | They don't want to depend on third party companies to
               | complete their stack offering on the datacenter. It's the
               | same way as why Nvidia wanted ARM, but on the other
               | direction (CPU/GPU)
        
       | pyrolistical wrote:
       | Short term is may seem like a smart move to cut AXG but long term
       | if it's a must have.
       | 
       | All three AMD, nvidia and Apple are unifying general purpose
       | compute with graphics. That is the direction the world is going.
       | Unless intel had a new arch trick up their sleeves they will be
       | soon made redundant
        
         | barkingcat wrote:
         | since everyone is on the hybrid computing bandwagon, might as
         | well bring the axg on core and make a xeon with axg on the same
         | package. that's what apple is doing, might as well give that a
         | try.
        
         | nicoburns wrote:
         | Yes, the insight being that "graphics" is actually mostly just
         | highly parallel compute. So if you don't have a graphics
         | solution then you're unlikely to have a competitive compute
         | solution either. It's not quite there yet, but I think it will
         | be soon.
        
         | paulmd wrote:
         | And actually not just graphics but the world is moving towards
         | these highly unified solutions in general, the hardware world
         | is going "full stack" and companies need to be able to offer
         | top-to-bottom solutions that solve the whole problem without
         | relying on someone else's silicon.
         | 
         | Same reason NVIDIA wanted to buy ARM. Intel and NVIDIA are in
         | trouble unless they can make that leap.
         | 
         | Long-term it's the same reason AMD bought ATI too, that vision
         | just took a long time to come to fruition. Remember, they
         | "acquired" their way to success as well, RTG wasn't something
         | that AMD indigeneously developed themselves either... just like
         | AMD bought Xilinx and Intel bought Altera.
        
       | klelatti wrote:
       | To be successful as a new entrant against powerful incumbent
       | players you need to have some sort of competitive angle /
       | advantage. Intel used to have a fabrication lead but chose not to
       | make dGPUs whilst they had that and allowed Nvidia and AMD to own
       | the market.
       | 
       | Really struggling to see this leading to a successful outcome -
       | even if they produce a reasonable product it's likely to be third
       | placed which is not a comfortable place to be.
        
       ___________________________________________________________________
       (page generated 2022-08-09 23:00 UTC)