[HN Gopher] Nvidia in the Valley
       ___________________________________________________________________
        
       Nvidia in the Valley
        
       Author : mfiguiere
       Score  : 175 points
       Date   : 2022-09-26 14:09 UTC (8 hours ago)
        
 (HTM) web link (stratechery.com)
 (TXT) w3m dump (stratechery.com)
        
       | digdugdirk wrote:
       | This article lays out a solid overview as to how Nvidia got to
       | where they are now. I'm curious (as someone without a deep
       | knowledge of the technology or the industry) as to why AMD can't
       | create a similar dedicated ray-tracing functionality in their
       | chips? It seems to be where the industry is going, and the
       | article later goes on to point out how Meta and Google are doing
       | this themselves.
       | 
       | Why has AMD ceded this market? Is it a patent issue? Capability
       | issue? Something else?
        
         | samstave wrote:
         | amd seed was intel having to lit give them chip designs... in
         | the 90s.
         | 
         | google was secretly making their own motherboards in ~2004 that
         | were multi proc
         | 
         | "meta" (fucking idiot name) was building a crap ton of secret
         | datacenter hardware and architecture in the 2010s...
         | 
         | and they are part of the secret sauce...
         | 
         | (source ;;worked for all in diff capacities)
        
         | BudaDude wrote:
         | I'm also curious why AMD hasn't implemented something like CUDA
         | cores for ML. AMD cards are basically useless for doing any
         | intensive ML task.
        
           | smoldesu wrote:
           | They did try. A few years ago AMD and Apple collaborated to
           | make OpenCL, which was a pretty half-hearted attempt at
           | building a platform-agnostic GPGPU library. Their heart was
           | in the right place, but that was part of the problem.
           | Nvidia's vertical control over their hardware and software
           | stack gave them insane leveraging power for lots of dedicated
           | use cases (video editing, gaming, 3D rendering, machine
           | learning, etc.)
           | 
           | In the end, even after years of development, OpenCL was just
           | really slow. There wasn't a whole lot of adoption in the
           | backend market, and Apple was getting ready to kick them to
           | the curb anyways. It's a little bit of a shame that AMD got
           | their teeth kicked in for playing Mr. Nice Guy, but they
           | should have know that Nvidia and Apple race for pinks.
        
             | pjmlp wrote:
             | Intel and AMD never bothered to move OpenCL beyond bare
             | bones C source, while NVidia not only moved CUDO into a
             | polyglot ecosystem early on, they doubled down on IDE for
             | GPGPU computing and first class libraries.
             | 
             | Google never bothered with OpenCL on Android, pushing their
             | C99 Renderscript dialect instead.
             | 
             | Apple repented themselves of offering OpenGL to Khronos and
             | the direction not going into the way they wanted to.
             | 
             | Those that blame NVidia for their "practices" should rather
             | look into how bad the competition has been from day one.
        
             | paulmd wrote:
             | OpenCL had a bit of a "second-mover curse" where instead of
             | trying to solve one problem (GPGPU acceleration) it tried
             | to solve _everything_ (a generalized framework for
             | heterogeneous dispatch) and it just kinda sucks to actually
             | use. It 's not that it's slower or faster, in principle it
             | should be the same speed when dispatched to the hardware
             | (+/- any C/C++ optimization gotchas of course), but it just
             | requires an obscene amount of boilerplate to "draw the
             | first triangle" (or, launch the first kernel), much like
             | Vulkan, and their own solution is still a pretty clos
             | 
             | HIP was supposed to rectify this, but now you're buying
             | into _AMD 's_ custom language and its limitations... and
             | there are limitations, things that CUDA can do that HIP
             | can't (texture unit access was an early one - and texture
             | units aren't just for texturing, they're for coalescing all
             | kinds of 2d/3d/higher-dimensional memory access). And AMD
             | has a history of abandoning these projects after a couple
             | years and leaving them behind and unsupported... like their
             | Thrust framework counterpart, Bolt, which hasn't been
             | updated in 8 years now.
             | 
             | https://github.com/HSA-Libraries/Bolt
             | 
             | The old bit about "Vendor B" leaving behind a "trail of
             | projects designed to pad resumes and show progress to
             | middle managers" still reigns absolutely true with AMD. AMD
             | has a big uphill climb _in general_ to shake this
             | reputation about being completely unserious with their
             | software... and I 'm not even talking about drivers here.
             | This is _even more_ the widespread community perception
             | with their GPGPU /ML efforts than with their drivers.
             | 
             | http://richg42.blogspot.com/2014/05/the-truth-on-opengl-
             | driv...
             | 
             | AMD doesn't have a library of warp-level/kernel-
             | level/global "software primitives" like Cuda Unbound or
             | Thrust either. So instead of writing your application, you
             | are writing the primitives library, or writing your own
             | poor implementation of them.
             | 
             | https://github.com/NVIDIA/cub
             | 
             | https://github.com/NVIDIA/thrust
             | 
             | It's just a fractal problem of "the software doesn't exist
             | and AMD would really rather you write it for them" all the
             | way down and nobody wants to do that instead of doing their
             | own work. AMD is the one who benefits from the rewrite, for
             | everyone else it's a "best case scenario it works the same
             | as what we've already got", so if AMD isn't gonna do it
             | then pretty much nobody else is gonna leap on it. And then
             | AMD has poor adoption and no software and the cycle
             | continues.
             | 
             | AMD really really just needs to get serious and hire a half
             | dozen engineers to sit there and write this software, cause
             | it's just not going to happen otherwise. It's a drop in the
             | bucket vs the sales to be realized here even in the medium
             | term, like one big ML sale would probably more than pay
             | those salaries. They're not doing it because they're cheap
             | or they're doing it because they're not really serious,
             | take your pick, but, AMD is no longer _that_ broke, they
             | can afford it and it makes financial sense.
             | 
             | Again, not a "nice" thing to say but it's the cold truth
             | here. I feel like I've made some variation on this post
             | about every 6 months for like 5 years now but it's still
             | relevant. If you as a vendor don't care about writing good
             | code for key features/libraries for your product, nobody
             | else is either, and you'll never get uptake. It's the same
             | thing with AMD/ATI not embedding developers with studios to
             | get those optimizations for their architectures. Console
             | lock-in will only get you so far. If you don't care about
             | the product as a vendor, nobody else will either.
             | 
             | It's remarkable how much flak Jensen got for "NVIDIA is a
             | software company now" back in 2009, and how people _still
             | don 't get it_, AMD is _not_ a software company and that 's
             | why they keep failing. Writing the framework that turns
             | into StableDiffusion and sells a billion dollars of GPUs is
             | the NVIDIA business model, AMD keeps trying to jump
             | straight to the "sell a billion dollars of GPUs" part and
             | keeps failing.
        
           | paulmd wrote:
           | AMD has "tensor cores" (called something different but
           | they're very similar matrix accelerator units) in CDNA. RNDA3
           | is supposed to have "something", it has the same instruction
           | as CDNA, it's supposed to be less than a full unit but
           | presumably there wouldn't be an instruction without some
           | level of hardware acceleration either.
           | 
           | The bigger problem is that AMD doesn't want to pay to keep up
           | on the software side... at the end of the day when you're
           | coming from behind you just have to pay someone to port key
           | pieces of software to your platform. AMD has really coasted
           | for a long time on letting the open-source community do their
           | work for them, but that's not going to fly with things like
           | PyTorch or other key pieces of software... if AMD wants the
           | sales and the adoption of their hardware, it's just going to
           | have to pay someone to write the software, so that people who
           | want to do research and not PyTorch maintenance can justify
           | buying the hardware.
           | 
           | I am not particularly interested in the perceived historical
           | justifications for the current situation, it doesn't matter
           | to the businesses who might be AMD's customers. And actually
           | in many ways they've gotten even shakier recently, what with
           | dropping RDNA support from their NN/ML package. As a cold
           | statement of reality, this is table stakes going forward and
           | if AMD doesn't want to do it they won't get the sales.
           | 
           | It's not even just PyTorch either, it's... everything. AMD is
           | just coming from a million miles behind on the software, and
           | "welp just write it yourself if you want to use our hardware"
           | is not an attitude that is conductive to selling hardware.
        
             | rektide wrote:
             | Which libraries is AMD dropping RDNA support in?
             | 
             | That seems very inadvised. Nvidia's libraries being usable
             | by a broad range of developers on a wide range of hardware
             | is critical to their wide adoption. AMD cannot expect to
             | have real adoption if only their fancy enterprise grade
             | unobtanium cards support ML systems. AMD needs a wide &
             | engaged community trying to yse their stuff to figure out
             | what software/drivers they simply have to build.
        
               | paulmd wrote:
               | Talking specifically about ROCm here, their ML package.
               | 
               | God this is such a tough paragraph to write accurately.
               | AMD themselves have conflicting information all over
               | their docs and repos and half of it is not even marked as
               | "outdated"...
               | 
               | https://docs.amd.com/bundle/ROCm-Getting-Started-
               | Guide-v5.2....
               | 
               | The official supported platforms at this point are RDNA2
               | (pro), GFX9 (pro), and CDNA. Consumer versions of these
               | (RDNA2, Radeon VII, and Vega 56/64) _probably_ work,
               | although Vega 56 /64 are an older version with much less
               | hardware support as well. RDNA2 support is also "partial"
               | and ymmv, things are often broken even on supported
               | cards.
               | 
               | If RDNA2 works, then RDNA1 _may_ work, but again, mega
               | ymmv, things may not even be great with RDNA2 yet.
               | 
               | The "hardware guide" link talks about supporting Vega 10
               | (that's V56/64) and says GFX8 (Polaris) and GFX7 (Hawaii)
               | are supported... but that doc is tagged 5.0 and 5.1 was
               | the release that dropped the other stuff. So I'd _say_
               | Vega 64 /56 chips are _probably_ broken at this point, on
               | the current builds.
               | 
               | Up until earlier this year though, it was unsupported on
               | any consumer card except Radeon VII. They dropped
               | Hawaii/Polaris/Vega support about 6 months before they
               | started adding partial RDNA2 support back.
               | 
               | And in contrast... NVIDIA's shit runs on _everything_ ,
               | going back 10 years or more. At least to Kepler, if not
               | Fermi or Tesla uarch (GTX 8800 series). It may not run
               | _great_ , but CUDA is CUDA, feature support has been a
               | _nearly_ forward ratchet, and they 've had PTX to provide
               | a rosetta stone in between the various architectural
               | changes.
               | 
               | I mean... at the end of the day the hardware hasn't
               | changed _that_ much, surely you can provide a shader
               | fallback (which is required anyway since RDNA2 doesn 't
               | have tensor acceleration). I don't get what the deal is
               | tbh.
               | 
               | https://www.techpowerup.com/288864/amd-rocm-4-5-drops-
               | polari...
               | 
               | https://www.phoronix.com/news/Radeon-ROCm-5.1
        
               | rektide wrote:
               | Yeah there was unbelievably terrible lag getting
               | Polaris/RDNA2 generation going. It was jaw droppingly
               | slow to happen! But it did finally happen.
               | 
               | I was afraid all RDNA[n] were going fully unsupported,
               | which felt like a sure invitation of death.
               | 
               | Sounds like there's still a lot of uncertainty, but it
               | also doesnt sound as bad as I'd first feared; it seems
               | like RDNA2+ could probably hopefully possibly work decent
               | well. As opposed to, you have to buy unobtanium hard to
               | find stupidly expensive cards. Seems like it's still
               | playing out, & we dont know what RDNA2+ is good for yet,
               | but it doesnt sound like the walk towards certain death
               | this originally sounded like.
               | 
               | Thanks for the intense hard to develop reply. A lot of
               | in-flight status updates to gather. Appreciated. Really
               | should be better established, what AMD is shooting for &
               | what we can expect.
        
         | ActionHank wrote:
         | AMD already has dedicated raytracing hardware on their GPUs,
         | but are behind Nvidia.
         | 
         | PC games that opted for Nvidia raytracing earlier on run poorly
         | with an AMD GPU with raytracing turned on. Cyberpunk 2077 is an
         | example of this, runs beautifully on Nvidia gpus with rt on,
         | framerate falls through the floor on an AMD card.
         | 
         | Raytracing on the "next gen" consoles PS5 and XBox Series X is
         | done using AMD hardware and runs really well.
        
         | Maciek416 wrote:
         | IMO, a major ingredient of raytracing running well enough in a
         | game that _also_ looks as good as any other current AAA title
         | in raster-only mode has been DLSS. Raytracing is (or at least,
         | at time of 2xxx series GPUs, was) still quite expensive to run
         | at full res at resolutions (1440p, 4K) and framerates
         | (60,120,144) that PC gamers demand. However, rendering
         | raytraced games at a much lower resolution is just within reach
         | for dedicated CUDA hardware. So DLSS makes up the difference
         | with very sophisticated upscaling. Without DLSS, I think the
         | raytracing in titles like Cyberpunk 2077 might not be
         | performant enough.
         | 
         | In light of this, you might go back and see that AI and RT for
         | NVidia have gone hand in hand, because one enables the other to
         | be performant enough for AAA titles. Opinions may vary greatly
         | on this, but personally, I don't think AMD's FSR upscaler is
         | capable of matching what DLSS can do in this regard. (Intel's
         | upscaling does seem to be capable of doing it, but very high
         | performance parts are still some ways away from release).
        
         | ckozlowski wrote:
         | I'm theorizing here, but I suspect it's because AMD feels like
         | building out a dedicated or proprietary capability could really
         | hurt them if it didn't take off. By that logic, NVidia's risk
         | here could hurt them as well (and that they are taking a risk
         | is the point the article it trying to make.)
         | 
         | AMD has for a long time favored an inclusive, compatible
         | approach. First (that I can recall) with x86-64, more recently
         | with AdaptiveSync over G-Sync, and now with their Ray Tracing
         | approach. Each time they chose a move efficient path that was
         | open to the industry as a whole.
         | 
         | This seems to have had some pros and cons. On the one hand,
         | they've been able to keep up with the market with a solution
         | that is the best value for them. They've never been a large
         | company against the likes of Intel and NVidia, so I suspect
         | there's less appetite for risk.
         | 
         | On the other hand, by always going that route, they cede the
         | leadership role to others, or if they do have leadership, it's
         | not in a way they can really leverage. It becomes commoditized.
         | Note how when the industry was moving to 64bit, AMD ended up
         | setting the direction over IA-64 with their more inclusive
         | approach. But it didn't turn into any significant leverage for
         | them. They set a standard, but it was one that everyone else
         | then adopted, including Intel.
         | 
         | So I feel like while AMD's approach keeps them alive and always
         | in the running, it's an approach that will never put them on
         | top. Whether or not this is a bad thing really depends on what
         | the goals of the company are, and if the the goal is to remain
         | steadily in the race, then they're doing great.
         | 
         | But arguably, NVidia pulls the industry in directions by way of
         | its choices. They're risky and sometimes irritable. It's also
         | put them in front.
         | 
         | So in my opinion, AMD hasn't ceded the market, but they have
         | ceded leadership in many instances by their safe approach. It's
         | still profitable and safe for them. But they'll always remain
         | second place as a result.
        
           | viscanti wrote:
           | The argument is made in the article. AMD is cutting all the
           | things that are expensive but with limited markets (no
           | special hardware on chip for raytracing or AI as well as
           | using a slightly older fab). They'll focus on being much
           | cheaper with nearly the same performance and less energy
           | requirements than NV.
        
           | bee_rider wrote:
           | It makes sense -- as the underdog they need to erode
           | entrenched advantages, starting up a standards-based
           | compatible approach is a cheap way of doing so (on top of
           | being totally laudable and good for the community).
           | 
           | I wonder if we could ever see Radeon Rays on Intel's GPUs, or
           | even their iGPUs. Raytracing in every low-resource-
           | requirement MOBA, I say!
        
         | gamdevthrowaway wrote:
         | > as to why AMD can't create a similar dedicated ray-tracing
         | functionality in their chips?
         | 
         | They do and it works well.
         | 
         | > out how Meta and Google are doing this themselves.
         | 
         | Meta and Google develop products everywhere all the time.
         | 
         | The author doesn't play or develop games, so it's okay that he
         | doesn't really know anything or can meaningfully comment on it.
         | He just took what Huang said and Ctrl+C Ctrl+V'd it.
         | 
         | WHY DO PEOPLE CARE ABOUT RAYTRACING?
         | 
         | Photoreal gets financing. For your game, for your crypto thing,
         | whatever. Raytracing makes photoreal demos at pre-financing
         | costs.
         | 
         | Also essential reason why Unreal Engine is appealing. Unity is
         | for people who make games, Unreal is for people who _finance_
         | games.
         | 
         | WHY DO PEOPLE CARE ABOUT AI-GENERATED CONTENT?
         | 
         | The Darwinian force here is the same: people believe you can
         | make a game (or whatever) out of DALL-E or whatever dogshite.
         | They don't believe you when you say you would hire an artist in
         | Indonesia for the same cost (or whatever). So you'll get
         | financed by saying one and not the other.
         | 
         | The reasons why don't really matter. It's idiosyncratic. You're
         | going to spend the money however it is. Just like every
         | startup.
         | 
         | Also the AI generation thing attracts people who think they're
         | the smartest, hottest shit on Earth. That attitude gets
         | financing. Doesn't mean it reflects reality.
         | 
         | DO THESE TECHNOLOGIES MATTER?
         | 
         | I don't know. What does Ben Thompson know about making fun
         | video games? It's so complicated. I doubt Big Corporate or VC
         | capital is going to have financed the innovative video game
         | that uses raytracing or AI generated content. It's going to be
         | some indie game developer.
        
           | TillE wrote:
           | > I doubt Big Corporate or VC capital is going to have
           | financed the innovative video game that uses raytracing
           | 
           | Sorry, you _don 't_ believe that a AAA game developer is
           | going to take full advantage of the latest high-end GPU
           | capabilities? That's the one thing they _have_ reliably done
           | for the past 25+ years.
        
             | Jasper_ wrote:
             | Uh, AAA game developers not using NVIDIA-developed tech?
             | Yes. Tessellation was a huge fail that didn't deliver on
             | performance. Geometry shaders were another huge fail. Mesh
             | shaders are shaping up to be pretty unimpressive. Pretty
             | much all of NVIDIA's middleware (Ansel, Flow, HairWorks,
             | WaveWorks, VXGI) haven't really caught on fire; usually
             | they're placed into the game through NVIDIA paying
             | developers large sums of money, and usually ripped out upon
             | the next game release.
             | 
             | What instead happens is that game developers are developing
             | large bunches of tech in-house that exploit new features
             | like compute shaders, something NVIDIA has struggled to
             | keep up with (lagging behind AMD in async compute,
             | switching compute/graphics pipelines has a non-zero cost,
             | lack of FB compression from compute write).
             | 
             | I say all of this as a graphics developer working inside
             | the games industry.
             | 
             | Ben makes a bunch of historical errors, and some pretty
             | critical technical errors, that I consider it to be almost
             | a puff piece for NVIDIA and Jensen.
             | 
             | (NVIDIA definitely did not invent shaders, programmability
             | is a blurry line progressing from increasingly-flexible
             | texture combiners, pioneered by ArtX/ATI, NVIDIA looooved
             | fixed-function back then. And raytracing really does not
             | function like however Ben thinks it does...)
        
               | gamdevthrowaway wrote:
               | > something NVIDIA has struggled to keep up with (lagging
               | behind AMD in async compute, switching compute/graphics
               | pipelines has a non-zero cost)
               | 
               | Someone who knows what he is talking about.
               | 
               | Like Ben Thompson could write an e-mail to any of the
               | 10,000 people who know something about this and just ask,
               | right?
        
             | gamdevthrowaway wrote:
             | The keyword there was innovative, I'm sorry. Like imagine
             | Portal, an innovative rendering game - now, is there
             | something innovative that is going to be done with a
             | raytracing feature like accelerated screen space
             | reflections? No, probably not, people could render mirrors
             | for ages. It's going to be like, one guy who creates cool
             | gameplay with the hardware raytracing APIs, just like it
             | was a small team that created Portal.
        
           | 411111111111111 wrote:
           | > _Also essential reason why Unreal Engine is appealing._
           | 
           | Is that why unreal engine deprecated ray tracing?
           | 
           | You're honestly quiet clueless considering your phrasing. I
           | bet you haven't worked in a large studio working with unreal
           | engine either.
        
             | gamdevthrowaway wrote:
             | > Is that why unreal engine deprecated ray tracing?
             | 
             | Unreal Engine didn't deprecate ray tracing, they are using
             | the same DirectX DXR APIs in a different place. But I think
             | you already know that.
             | 
             | > I bet you haven't worked in a large studio working with
             | unreal engine either.
             | 
             | Have you?
        
               | 411111111111111 wrote:
               | > _Unreal Engine didn 't deprecate ray tracing,_
               | 
               | They literally did. They found that Lumen looked better
               | and used less performance then hardware ray tracing. At
               | least thats what they said in a developer interview a few
               | month ago.
               | 
               | > _Have you?_
               | 
               | No, I haven't. My day job isn't even in the game
               | industry. I've only dabbled a little after I bought my
               | first VR headset.
               | 
               | The either was because you were dissing the article
               | author by saying they weren't a game dev, with the
               | argument that UE is only used because it looks shiny to
               | management, with is pure nonsense.
        
               | cypress66 wrote:
               | That's incorrect. Lumen supports both software and
               | hardware raytracing.
               | 
               | Software rt lumen is lower quality, often faster,
               | supports any hardware and has some pretty big
               | limitations.
               | 
               | Hardware rt lumen is higher quality, often slower, needs
               | hardware support and has a lot less limitations.
        
               | 411111111111111 wrote:
               | I cannot find the interview I remembered so you're both
               | almost certainly right wrt hardware ray tracing. I still
               | think that their original premise about unreal engine is
               | nonsense however.
        
       | stephc_int13 wrote:
       | I am not worried about Nvidia.
       | 
       | They will clearly suffer for a while, a bad year or maybe two,
       | not enough to sink the company or to kill the golden goose.
       | 
       | They are spreading themselves a bit thin with all their projects,
       | but I can see them succeeding with enough of them to be in a
       | better relative position 2-3 years from now.
       | 
       | GPUs are becoming a general-purpose data-plane computing
       | platform, this not simply related to gaming, crypto or AI
       | training, but everything.
        
       | paulpan wrote:
       | Personally I'm very eager to see AMD's upcoming RX 7000-series
       | from a pricing as well as performance standpoint compared to
       | Nvidia's new 4000 series.
       | 
       | As noted in the article, "Nvidia likely has a worse cost
       | structure in traditional rasterization gaming performance for the
       | first time in nearly a decade" due to the much lower die costs
       | that AMD is targeting. I predict Nvidia's share price will
       | continue its downward trajectory - both on the weakness of 4000
       | series (given high price and surplus 3000 series inventory) and
       | on AMD pricing pressure.
       | 
       | What's more interesting is Nvidia's long term focus on ray
       | tracing and continuing to dedicate more and more precious die
       | space to it. AMD, on the other hand, seems to be more focused on
       | the short term: rasterization performance. It's a bit reminiscent
       | of Apple vs. Samsung, where AMD is Apple as it waits on
       | technologies to become mainstream before full adoption and Nvidia
       | is Samsung as it bets on new technologies (e.g. foldables).
        
       | bullen wrote:
       | How far off can you be: electricity is never going to become
       | cheaper.
       | 
       | Right now playing 5 hours of 1060 PC games costs $1 in EU.
       | 
       | To even think of selling something that draws more than 30W is
       | ridicoulus.
       | 
       | The metaverse has to scale DOWN to 5W Raspberry 4.
       | 
       | VR/120Hz/RTX are completely out of the question.
        
       | vimota wrote:
       | >Similarly, if AI applications become democratized and accessible
       | to all enterprises, not just the hyperscalers, then it is Nvidia
       | who will be positioned to pick up the entirety of the long tail.
       | 
       | How does this square with Cloud providers themselves being
       | hyperscalers ? What's to stop Google/AWS/Microsoft's hardware
       | divisions from outpacing Nvidia since those businesses themselves
       | may need the hardware and can then provide it to end users of
       | their cloud platforms.
        
       | m3kw9 wrote:
       | With gfx slowly trending towards cloud and edge computing, I
       | wonder how that benefits them, or not?
        
         | dotnet00 wrote:
         | I've been hearing about this for a while but has any mainstream
         | game actually demonstrated meaningful realtime graphics
         | streaming? I suspect that kind of like NFTs or Metaverse, most
         | developers see that it's a pretty bad idea in a practical
         | sense, but talk about it because it's easy PR.
         | 
         | People just barely tolerate games which require an internet
         | connection to run for DRM, I can't imagine them appreciating
         | the requirement for the connection to also be fast and stable
         | enough for streaming, especially with handheld PCs starting to
         | catch on. Especially since they'll also want to take down the
         | expensive servers quickly after sales fall off, most likely
         | making the games unplayable.
        
           | squeaky-clean wrote:
           | There's several games released on Nintendo Switch that run
           | only via cloud streaming. The Kingdom Hearts switch "ports"
           | are all streamed.
           | 
           | It's also the only way to play Fortnite on iOS after Apple
           | removed Fortnite from the app store.
           | 
           | https://www.epicgames.com/fortnite/en-US/mobile
           | 
           | > People just barely tolerate games which require an internet
           | connection to run for DRM
           | 
           | While some people are generally very loud about this on
           | Reddit and gaming forums, the average person doesn't seem to
           | care. Even people who complain loudly about it online will
           | quietly buy it if it's the only choice. It reminds me of that
           | old screenshot of a Steam group dedicated to "Boycotting
           | Modern Warfare 2" in 2009. The day after release? Most of the
           | group was playing it.
           | 
           | https://i.kinja-img.com/gawker-media/image/upload/s--
           | W8ZYHe9...
        
             | izacus wrote:
             | There's a difference between being angry at DRM in general
             | and not being able to play a game because your cheapskate
             | ISP gave you a shitty overbooked connection.
             | 
             | Stadia hit that issue quite a bit. And I haven't really
             | seen cloud titles on Switch (a famously offline console
             | bought by people who really want to play away from stable
             | networks) have significant sales success.
        
           | SketchySeaBeast wrote:
           | I played with Microsoft Flight Simulator 2020 gamepass
           | streaming. It was fine, games like that don't require
           | fractions of a second response time, it looked like I was
           | watching an interactive high compressed youtube video of the
           | game set to medium, but the whole time I was going "Why don't
           | I just play the prettier version on my PC?"
        
         | stephc_int13 wrote:
         | Not sure about this trend.
         | 
         | The idea of a dumb terminal had been around forever, and so
         | far, it has been cyclically going up and down.
         | 
         | In my opinion, we'll always have both.
        
       | Melatonic wrote:
       | This guy really loves his leather jackets eh?
        
       | russli1993 wrote:
       | if chiplet is the silver bullet to die costs, I don't get why
       | Nvidia won't go for it. After all, Nvidia has smart ppl too.
       | There must be some design tradeoffs that made Nvidia engineers to
       | believe their approach is still the best in terms of perf and
       | area. For 4090 and 4080 price, I think Nvidia strategy is: they
       | are higher tier, with 3000s as the lower tier more affordable
       | option. They feel they can succeed with this strategy in the
       | current market (1 year or so) given the competitive landscape.
       | Then they can use 4090 and 4080s to drive up the ASP and margins.
       | Their biggest motivation is sustain the stock price.
        
         | zackmorris wrote:
         | I hadn't heard the term chiplet before, thank you! When I got
         | out of school in the late 90s, system-on-chip (SoC) was a pipe
         | dream. It has issues because memory might have different
         | requirements than CPU, so it's hard to put all of the
         | components on a single die.
         | 
         | I always thought the situation was frankly pretty silly,
         | because I didn't want one big die anyway. I wanted a scaleable
         | grid of chips like the 3D processor from the Terminator 2
         | movie.
         | 
         | It took over 20 years, but maybe we'll have a chance to get
         | something like that with an open source RISC-V. Maybe an array
         | of matchhead-size CPUs at 1 GHz costing $1 each that could be
         | scaled to hundreds or thousands of cores with content-
         | addressable memory for automatic caching and data locality.
         | Then it could be treated as a single processor with access to a
         | one large shared memory space and we could drop the (contrived)
         | distinction between CPU/GPU/TPU/etc and get back to simple
         | desktop computing rather than dealing with the friction of
         | vertex buffers.
         | 
         | My guess is that Nvidia doesn't want any of that to happen, so
         | works to maintain its dominance in single-die solutions.
        
           | paulmd wrote:
           | "match-head sized chiplets" sort of falls into the chiplet
           | mythos I think. Chiplets aren't magic, they actually
           | _increase_ power usage vs an equivalent monolithic chip, data
           | movement is expensive and the more data you move the more
           | expensive it is. People just think chiplets are efficient
           | because AMD made a huge node leap (GF 12nm to TSMC 7nm is
           | like, more than a full node, probably at least 1.5 if not 2)
           | at the same time, but chiplets have their own costs.
           | 
           | The smaller you split the chiplets, the more data is moving
           | around. And the more power you'll burn. It's not desirable to
           | go _super_ small, you want some reasonably-sized chiplet to
           | minimize data movement.
           | 
           | Even if you keep the chiplets "medium-sized" and just use a
           | lot of them... there is still some new asymptotic efficiency
           | limit where data movement power starts to overwhelm your
           | savings from clocking the chips lower/etc. And there's
           | copper-copper bonding to try and fix that, but that makes
           | thermal density even worse (and boy is Zen4 hot already...
           | 95C under _any_ load). Like everything else, it 's just
           | kicking the can down the road, it doesn't solve all the
           | problems forever.
        
         | paulmd wrote:
         | Even AMD is only splitting off IO dies this generation. That's
         | an advantage for sure, because you can push 20% or so of your
         | die off to separate chiplets, which means you can go 25% larger
         | than you otherwise could with a purely monolithic design.
         | 
         | (and in particular you can also push them off to N6 or some
         | other super-cheap node and save your super-expensive N5
         | allocation for the GCD... at the cost of some power for
         | infinity fabric to send a TB/s of data around between chiplets)
         | 
         | But so far nobody seems to have progressed on splitting the GCD
         | itself... which has always been the actual goal here. Multiple
         | chiplets computing together like one, transparently without
         | software needing to be catered to like SLI.
         | 
         | AMD's general approach seems to be "big cache-coherent
         | interconnect between GCD chiplets", which is generally
         | unsurprising on multiple levels (it's how infinity fabric
         | already works in their CPUs, and it's how SMP generally works
         | in general today almost everywhere) but there still seems to be
         | some barrier to making it work with _graphics_. There are of
         | course a lot of gotchas - like temporal data or needing access
         | to other parts of the frame that may be in another GPU - but
         | cache-coherent interconnects generally solve this.
         | 
         | But yeah, NVIDIA isn't avoiding some silver bullet here,
         | they've actually been on the forefront of this with NVSwitch
         | since Pascal, that was the first serious "multi-GPU that acts
         | like one" attempt. I don't know why they're not doing it, or at
         | least the "split the IO dies off" thing.
         | 
         | edit: thinking about this a bit more, one reason they may not
         | be doing the MCD/io die idea is because they're doing L2
         | instead of L3. L3 can be pushed off to a separate die... L2 is
         | really part of the SM, it's a lot "closer". Again, they may
         | have engineering reasons to favor L2 over L3, or it may be a
         | patent thing.
        
         | dotnet00 wrote:
         | It isn't the case that NVIDIA won't go for a chiplet design,
         | they're working on it but it isn't ready yet. Current
         | expectations for NVIDIA's chiplets to be ready seem to be for
         | either a refresh of the 50xx series or the 60xx series.
        
       | maerF0x0 wrote:
       | > I don't think we could have seen it. I don't think I would've
       | done anything different,
       | 
       | This is a CEO admitting they essentially learned nothing, will
       | not admit they made a wrong choice in light of data, and
       | typically means they will do it again.
       | 
       | They could have seen the writing on the wall about how people
       | were angry about crypto emissions and that proof of stake was
       | being bandied about as a solution (and that Cardano already had
       | it, so it's not just a theory).
        
         | johnla wrote:
         | I read that more as CYA than anything.
        
       | treis wrote:
       | Lots of people talking about the economics but my take away from
       | the article was that we're on the cusp of something really cool.
       | The combination of ray tracing, physics via ray tracing, and GPT
       | like content generation are all the ingredients needed to make
       | huge immersive worlds. Imagine GTA with every house, office, and
       | sky scraper full of unique realistic looking people doing
       | realistic things in a fully destructible environment.
        
       | wslh wrote:
       | Offtopic: Stratechery has a podcast at [1] including the
       | translated interview with Nvidia CEO [2]. The weird thing from
       | the marketing/conversion perspective is that the interview
       | translation says "To listen to this interview as a podcast, click
       | the link at the top of this email to add Stratechery to your
       | podcast player." instead of having a direct link in the web page.
       | 
       | [1] https://open.spotify.com/show/1jRACH7L8EQCYKc5uW7aPk
       | 
       | [2] https://stratechery.com/2022/an-interview-with-nvidia-ceo-
       | je...
        
         | draw_down wrote:
        
       | wing-_-nuts wrote:
       | NV saw the profits it made from the shortages of the last mining
       | boom, and determined that it was better to make a much higher
       | profit even at the cost of shipping fewer units. They went full
       | bore on the 'costs and tdp don't matter'. They saw that people
       | were willing to pay any price for cards that were 5% faster at a
       | higher power draw.
       | 
       | Only, the mining boom is over, there are a _mountain_ of 3000
       | series cards floating around out there, and NV is sweating. The
       | flood of used mining cards is going to make this the first good
       | time to buy a GPU in _years_ and no amount of shenanigans with
       | 4000 series cards is going to change that math.
        
         | johnla wrote:
         | Where's the best place to pick up these cheap cards? eBay?
        
           | wing-_-nuts wrote:
           | ebay, fb marketplace, r/hardwareswap
        
         | izacus wrote:
         | This sounds like a win for nVidia in any case, since it'll
         | entrench their RTX, CUDA, NVENC and DLSS moats, making them
         | ubiquitous targets for games and driving purchasing decisions
         | in the future.
         | 
         | But they'll need to wade out a generation for that - which
         | probably isn't the first time.
        
         | adolph wrote:
         | 40 years from now someone will make good use of those in a
         | documentary . . .
         | 
         | https://en.wikipedia.org/wiki/Atari_video_game_burial
         | 
         | https://en.wikipedia.org/wiki/The_Mystery_of_Al_Capone's_Vau...
        
       | monkmartinez wrote:
       | The article kind of led with Crypto, but could have easily
       | started and ended with Crypto. Full Stop.
       | 
       | Overlay Ethereum-USD and Nvidia charts for the past 5 years and
       | what does that tell you? My take is both Nvidia's stock price and
       | Ethereum-USD price were products of helicopter money. Sure some
       | gamers bought cards, but not on the scale to which miners
       | acquired them.
       | 
       | Analysis of mining hash rate post-merge bears this out[0]. With
       | some estimates being $5 to $20 Billion(USD) in newer GPU's that
       | are no longer "profitable."[1] The 'chip' shortage meant Nvidia
       | could not satisfy demand for their products. I posit no amount of
       | chips would have satisfied demand. They, along with many other
       | companies, made record profits during the pandemic. There was
       | just soooo much 'extra' money flowing around with stimulus,
       | overly generous unemployment, PPP, and more.
       | 
       | I can't tell you how many people were YOLO'ing every cent of
       | their stimulus into Crypto and WSB stonks. It was an enormous
       | cohort of people based on the explosion of Robinhood and Coinbase
       | use mentioned in WSB subreddit. Mooning and Diamond hands FTW,
       | eh?
       | 
       | Mining crypto with gainz from a roulette spin with Robinhood
       | options? Yeah, while we are at it, lets buy 4 mining rigs because
       | mining Ethereum is very, very profitable during this time. The
       | thinking that one can literally turn electricity to heat while
       | making money (even while one sleeps no less) and trade options
       | while one is awake. It may look crazy right now, but the
       | subreddits were non-stop talk like this.
       | 
       | There are mining rigs all over my Facebook marketplace and local
       | craigslist now. Crypto winter is here (for who knows how long)
       | and Nvidia is not going to make those types of returns for the
       | extreme foreseeable future. Reality check 101.
       | 
       | [0]https://twitter.com/WuBlockchain/status/1552837630278385664
       | [1]https://www.theblock.co/post/168380/whats-left-for-
       | ethereum-...
        
         | jejeyyy77 wrote:
         | After crypto it will be ML and general cloud computing.
        
           | monkmartinez wrote:
           | How many people do you think are 'into' ML enough to spend
           | $1k+ on a SOTA gpu and associated hardware? I am slowly
           | getting there having moved from Colab to my own set up...
           | 
           | However... Ethereum enabled/lured many everyday joes into
           | buying 4+ $1k cards due to the incentive structure of; 'buy
           | mining rig then buy a lambo.' Setting up a GPT-NeoX is orders
           | of magnitude more difficult than pointing GPU compute at
           | Nicehash. I really have a hard time thinking ML will have any
           | meaningful uptake in that regard because the incentive
           | structure isn't the same and its much, much harder.
           | 
           | Big cloud seems to be going there own way in regards to
           | compute. GPU's are great for ML, but that doesn't mean they
           | will always hold the crown. TPU's, NVMe storage, and chiplets
           | may find a better path with newer software.
           | 
           | I just don't see how Nvidia really thrives without
           | drastically reducing price (less margin). I don't think they
           | are dead, but they are in big trouble as are many companies.
        
             | jejeyyy77 wrote:
             | New Macbook Pros and workstations are now coming with
             | powerful GPU's for ML work.
             | 
             | StableDiffusion alone was trained on 256 x Nvidia A100
             | GPUs.
        
               | monkmartinez wrote:
               | Correct, MBP's can run stable diffusion and other ML
               | workloads on non-nvidia hardware. I clearly see this
               | becoming a trend. GPT-J, Neo and NeoX run really well on
               | Colab TPU's, again these are not made by Nvidia.
               | 
               | Training is dominated by Nvidia, I will not question that
               | as most papers I have seen say something similar. I will
               | say that I do not believe training will always be
               | dominated by Nvidia's datacenter options. Two things that
               | will hasten the withdraw from Nvidia; Cuda and hardware
               | advances around the motherboard (ASICs, RAM proximity,
               | PCIe lanes, data transfer planes, etc).
               | 
               | Think about this... what if a company released an ML
               | training/Inference ASIC that used regular DDR4/NVMe,
               | performed like 4 x A100's and cost $8000? Would you be
               | interested? I would! I don't think this is too far off,
               | there has to be someone working on this outside of
               | Google, Apple and Meta.
        
               | [deleted]
        
               | KaoruAoiShiho wrote:
               | We've had several generations of ASICs already, if TPUs
               | etc aren't much superior to GPUs why would future ASICs
               | be any better.
        
         | justapassenger wrote:
         | > Overlay Ethereum-USD and Nvidia charts for the past 5 years
         | and what does that tell you? My take is both Nvidia's stock
         | price and Ethereum-USD price were products of helicopter money.
         | Sure some gamers bought cards, but not on the scale to which
         | miners acquired them.
         | 
         | While I do get your point, and largely agree, I don't think
         | that stock price is a good comparison. You can overlay also
         | Meta or Amazon stock, and get same conclusions. Last few years
         | were across the board tech bubble.
        
           | monkmartinez wrote:
           | Agreed, it has been very bubbly across the board. However, I
           | think Nvidia (AMD as well) and Ethereum are especially
           | correlated as first and second order effects seem
           | proportional to incentive feedback into their respective
           | ecosystems. Ethereum has been hounded by the 'wastefulness'
           | of mining, which is to say 'proof of stake' wasn't born in a
           | vacuum. Likewise, Nvidia GPU's were being bought at many
           | multiples of msrp in BULK! Call it a perfect storm, call it
           | what you want... but Nvidia has a tough road ahead.
        
       | kaycebasques wrote:
       | Is there any kind of precedent where bad news for the
       | semiconductor manufacturers means good news for specific
       | downstream industries? Examples? I.e. the oversupply of chip X
       | allowed industry Y to buy more chips resulting in innovation Z...
        
       | spicyusername wrote:
       | The last paragraph really sums it up                   In other
       | words, Nvidia has earned the right to be hated by taking the
       | exact sort of risks in the past it is embarking on now. Suppose,
       | for example, the expectation for all games in the future is not
       | just ray tracing but full-on simulation of all particles:
       | Nvidia's investment in hardware will mean it dominates the era
       | just as it did the rasterized one. Similarly, if AI applications
       | become democratized and accessible to all enterprises, not just
       | the hyperscalers, then it is Nvidia who will be positioned to
       | pick up the entirety of the long tail. And, if we get to a world
       | of metaverses, then Nvidia's head start on not just
       | infrastructure but on the essential library of objects necessary
       | to make that world real (objects that will be lit by ray-tracing
       | in AI-generated spaces, of course), will make it the most
       | essential infrastructure in the space.              These bets
       | may not all pay off; I do, though, appreciate the audacity of the
       | vision, and won't begrudge the future margins that may result in
       | the Celestial City if Nvidia makes it through the valley.
        
         | TechnicolorByte wrote:
         | It's a very good summary. Nvidia to me gets a lot of flak
         | (sometimes rightfully so) in large part to being the big risk-
         | taker in the GPU/graphics/AI space that pushes forward industry
         | trends. AMD, meanwhile, is largely a follower in the
         | GPU/graphics space with few contributions to rendering
         | research. Even worse, they even use their marketing to angle
         | themselves as consumer-friendly when I reality they're just
         | playing catch-up. For instance, they started off badmouthing
         | ray-tracing as not being ready back when Turing launched (when
         | is any new paradigm ever "ready" when it requires developers to
         | adapt?) before adopting it with much worse performance the
         | following generation.
        
       | ThomPete wrote:
       | There is going to be no end to demand in the future for Nvidia
       | unless it's going to another chip producer or another technology
       | (like quantum chips)
       | 
       | It's really that simple IMO.
       | 
       | AI have just started becoming mainstream and it will most
       | certainly play a big part in whatever gets us out of our current
       | situatins.
       | 
       | Disclaimer: I invested very early in nvdia stocks and is long on
       | them.
        
       | gitfan86 wrote:
       | Once I got my hands on stable diffusion, I bought a 3090 because
       | I was afraid that they would go up in price, but my power supply
       | wasn't big enough and I returned it for a 2060. But now I see
       | that 3090s are dropping like 40% in price since then. This
       | article does a great job explaining these dynamics.
       | 
       | Metaverse and crypto will be a bust, but democratized AI is going
       | to explode, the big risk for NVDA is that models and tools get
       | small and efficient enough that they don't need $1,000+ hardware.
        
         | flyinglizard wrote:
         | $1,000 is a very reasonable enterprise price point. After all,
         | datatcenter CPUs are much more expensive.
        
           | vineyardmike wrote:
           | Democratic means fit on an iPhone if not on $50 throw-away
           | hardware.
        
           | bongobingo1 wrote:
           | I think democratizing AI means more than "a very reasonable
           | enterprise price point" of $1000.
        
             | smoldesu wrote:
             | It doesn't, you can run this on $150 GTX 1060s if you're
             | willing to wait a few moments. The interesting angle here
             | is that there is a consumer segment of hardware (gaming
             | GPUs) that can be used to drive cool artistic experiments.
             | I think for most people, even non-gamers, getting dedicated
             | acceleration hardware is not any more expensive than an
             | entry-level analog synthesizer.
        
               | aidenn0 wrote:
               | With only 4GB, I'm RAM limited for a lot of models. I'm
               | thinking of buying a dedicated GPU just for ML stuff.
        
           | Dennip wrote:
           | Won't enterprise always go for the quadro cards which are
           | more like $4000+
        
         | jeffparsons wrote:
         | > the big risk for NVDA is that models and tools get small and
         | efficient enough that they don't need $1,000+ hardware.
         | 
         | You will always be able to do more and better with the premium
         | hardware. Maybe for video games the marginal differences aren't
         | enough to sway most consumers towards the high end cards, but I
         | expect there will be a lot of people willing to pay top dollar
         | to be able to run the best version of X AI tool.
        
           | bitL wrote:
           | The new 40x0 series has no NVlink even on their pro cards,
           | which is a major letdown and step back for AI folks.
        
           | solardev wrote:
           | With games, the best card might make sense cuz it could be
           | the difference between 60+ fps and 40.
           | 
           | For workloads though, does it really matter if it completes
           | in 20 seconds instead of 15?
           | 
           | Unless there's some feature difference in the higher end
           | cards (like AI-specific subchips), not just more of the same
           | & faster, then a lower card of the same generation shouldn't
           | impact your ability to run something or not.
           | 
           | Maybe RAM size...
        
             | squeaky-clean wrote:
             | How many times do you have to complete that 20/15 second
             | task in a year? Multiply that by 5 seconds to get the
             | "wasted time". Factor in the employees salary by that
             | wasted time, how much are you paying them to wait on that
             | slower hardware? Is it more money than the price difference
             | between the better GPU and the cheaper GPU?
             | 
             | Or maybe it's an automated thing that runs constantly? The
             | difference between 20 and 15 seconds is 480 additional
             | iterations per 8-hours. How much is that worth to you?
             | 
             | There's also the unmeasurable benefits. What if you're
             | losing potential good employees applying because word on
             | the net is your company uses last-gen hardware for
             | everything, and your competitors use the newest.
        
             | [deleted]
        
             | Blammar wrote:
             | RAM size is the issue. I can't run stable diffusion at 1k x
             | 1k on my 2080ti because I run out of vram.
        
             | [deleted]
        
         | mw888 wrote:
         | Democratized AI will not explode, it will thrive as centralized
         | services until the day real time AI is expected by consumers.
        
         | ronsor wrote:
         | I actually don't think Metaverse will be a bust, but I think
         | the current corporate implementation of it will be. Companies
         | are trying to cash in too early and it's not really working out
         | for them, particularly Zuckerberg's Meta.
        
           | metadat wrote:
           | What other implementation will come to fruition? AFAICT the
           | Metaverse is a shoddy VR clone of Second Life.
           | 
           | The Metaverse is already a bust, essentially a stillborn
           | fetus. The market doesn't actually want it.
        
             | mlsu wrote:
             | Zuck's Metaverse (tm) is very forward looking in some ways,
             | and very backward looking in others.
             | 
             | Although his particular vision is very weird and ugly, Zuck
             | is on the right track: in the future, people will spend
             | real time (not async time, a-la tiktok/instagram) with each
             | other in virtual spaces. Facebook is already a text-based
             | version of the metaverse, and it minted a pretty penny.
             | 
             | The metaverse, as Zuck imagines it, is already here though:
             | video games. People who are hardcore about video games wake
             | up, go to work, get home, and then plug in. Their friends
             | and social circle exists on the internet, in the game. This
             | is the metaverse. Video-game communities; Fortnight [1] ,
             | WoW, League, CS-Go, etc; these are the most complete so far
             | existing metaverse(s). One exists for every major flavor of
             | gameplay we can think of. Look on steam for people who have
             | 10,000 hours in a game. These people live as Zuck wants us
             | all to: in the metaverse.
             | 
             | Zuck's vision is to expand these virtual communities beyond
             | the hardcore gamer group. You go to work (in a metaverse)
             | then you come home and spend time with your friends (in a
             | different metaverse), then you play a game (in yet another
             | metaverse), then you sleep and do it all again the next
             | day. It's actually a very forward-looking vision, and one
             | that probably will come to pass over the next few decades.
             | A metric for this is "time spent in front of a screen." It
             | inexorably climbs year after year, generation after
             | generation.
             | 
             | But the key thing is that the experience has to be good.
             | The metaverse of Fortnight or CS-GO is worth spending time
             | in because those games are _fun_ and engaging in their own
             | right; for some, more fun and engaging than reality itself.
             | These games print money, because a small group of people
             | legitimately spend every leisure moment in them. Zuck 's
             | vision is to expand this beyond the market for hardcore
             | gamers and into work, socialization -- life itself.
             | 
             | I suspect that the market doesn't quite fully understand
             | this part of Zuck's metaverse dreams. In part because of
             | Meta's spectacular failure at making the metaverse look fun
             | and compelling.
             | 
             | Personally, my money's on Epic to actually bring the
             | Metaverse to fruition. They already own a Metaverse,
             | Fortnight, and they also own Unreal, which will probably be
             | the software foundation upon which the normie metaverse
             | gets built.
             | 
             | [1] https://medium.com/@ow/the-metaverse-is-already-here-
             | its-cal...
        
             | quantumwannabe wrote:
             | A spiritual successor to Second Life with modern graphics
             | and VR support would likely be very successful. Dumbed-down
             | versions like Fortnite Creative and Roblox have been very
             | popular. Zuckerberg won't be the one to build a successful
             | metaverse game though, not because Meta doesn't employ
             | skilled game developers, but because culturally the
             | leadership is completely out of touch. Second Life thrived
             | because of its uncontrolled player creation environment,
             | which Meta would never allow. It let you 3D model and
             | script _anything_ you could think of using in-game (and
             | external) tools and had a free market economy where the
             | developer only took a cut when converting to /from real
             | world currency. The whole debacle with the "virtual
             | 'groping'" and the personal space bubble has demonstrated
             | that even if Facebook's Metaverse did have good creative
             | tools, all it would take is a single article written by a
             | journalist with an axe to grind for them to cripple it.
             | They've already removed legs from avatars in Horizon Worlds
             | for "safety", which is ridiculous.
        
             | not2b wrote:
             | The market doesn't want Zuckerberg's version. Maybe some
             | day someone will get it right in a way that's good enough
             | to be interesting.
        
               | metadat wrote:
               | Yes, maybe someday, somewhere else.
        
         | WFHRenaissance wrote:
         | >Metaverse and crypto will be a bust
         | 
         | I love when HN netizens just throw stuff like this out without
         | further explanation. It's as if they're monks reciting the
         | dogma. Sigh.
        
           | flumpcakes wrote:
           | It is probably the default safe position. I think there would
           | be a larger need in explaining how they won't be a "bust"
           | compared to saying that they will. The internet needed
           | explaining to many people to see it's real advantages before
           | it took off.
        
       | yuan43 wrote:
       | > I [nvidia CEO] don't think we could have seen it [massive
       | effect of Ethereum merge on bottom line]. I don't think I
       | would've done anything different, but what I did learn from
       | previous examples is that when it finally happens to you, just
       | take the hard medicine and get it behind you...We've had two bad
       | quarters and two bad quarters in the context of a company, it's
       | frustrating for all the investors, it's difficult on all the
       | employees.
       | 
       | This is not confidence inspiring. It was obvious that the Etherum
       | merge would affect the bottom line in a big way. Why this
       | professed ignorance? Does it have to do with the fact that to
       | admit that it was visible a mile away would have been to admit
       | the deep reliance the company had come to have on the short-term
       | Ethereum mining boom?
        
         | SkyMarshal wrote:
         | Many folks _in_ crypto expected those Ethereum-mining GPU farms
         | to just switch to some other GPU-minable cryptocurrencies. It
         | wasn 't a certainty all those farms would just close up and
         | dump their GPUs on the market en masse. But Fed interest rate
         | policy hitting at the same time, driving down the crypto market
         | across the board (and all other risk assets), may have
         | unexpectedly changed the ROI calculation there and resulted in
         | the dump.
        
           | Tenoke wrote:
           | Many poorly educated folk maybe. All other chains were
           | already less profitable and had a small fraction of
           | Ethereum's hashrate. Even 5% of ETH's hashrate moving to them
           | is plenty to make them unprofitable for most. This was never
           | a likely outcome.
        
           | paulmd wrote:
           | > Many folks in crypto expected those Ethereum-mining GPU
           | farms to just switch to some other GPU-minable
           | cryptocurrencies.
           | 
           | This was a pretty common take but if you did the math
           | Ethereum had about 90% of the GPU-mining market (by hashrate)
           | so it was obvious the profitability was going to tank on
           | those other currencies as soon as Ethereum switched.
           | 
           | In the long run yes, there will probably be another big spike
           | in another cryptocurrency that starts another GPU boom. But
           | it's not magic where one instantly springs up to absorb all
           | the ethereum hardware at equivalent profitability.
           | 
           | A GPU crash was inevitable regardless of the interest rate
           | drop hitting at the same time.
        
           | swalsh wrote:
           | I hoped there would be a rise in proof of work like chains,
           | where in the work was something useful like training an AI or
           | brute forcing a hard but useful problem. Like a SETI@Home,
           | but paying crypto for successful solutions as opposed to
           | relying on altruism.
        
             | asciimike wrote:
             | There are a few of these type of things, e.g. RNDR token
             | (https://rendertoken.com) and rent a flop
             | (https://rentaflop.com) in rendering, and golem
             | (https://www.golem.network) and sonm (https://sonm.com/) in
             | the "general purpose computing on the blockchain"
        
             | omegalulw wrote:
             | It's hard to pull this off, if not impossible. A key
             | attribute of proof of work systems is that the difficulty
             | should be dynamically adjustable and that everyone has
             | perfect consensus on what "work" is. Doing meaningful work,
             | while admirable, puts the owners of those projects in
             | control of defining "work" and adjusting difficulty, i.e.,
             | people in the loop. That's not trustworthy from a currency
             | POV, no matter who the people are.
        
         | KaoruAoiShiho wrote:
         | It was sudden and couldn't have been predicted. Ethereum ended
         | PoW just this month but the GPU crash was 7 months ago. In
         | reality the PoW transition had nothing to do with the GPU
         | crash, it was the end of WFH and the crypto decline caused by
         | the russian invasion that resulted in the GPU crash.
        
         | [deleted]
        
         | varelse wrote:
        
         | opportune wrote:
         | I think Huang does not want to draw investors' attention to
         | crypto because he doesn't want people to equate Nvidia's
         | performance as a company with crypto performance. He doesn't
         | want Nvidia to just be a crypto company.
         | 
         | At the same time, he also definitely wants to cash in on any
         | future crypto booms, because they are lucrative.
         | 
         | It is best for him to take a position that mostly ignores
         | crypto. I think he legitimately doesn't want crypto to be the
         | future of Nvidia and doesn't want to build for that use case,
         | nor does he want to be financially reliant on it, but there is
         | also no point in him talking shit or spreading doom about
         | crypto when he can just shut up and still sell gpus.
        
           | swalsh wrote:
           | Nvidia doesn't have a role in any future crypto boom, unless
           | it's being used for analytics or AI. All modern chains use
           | PoS.
        
             | kranke155 wrote:
             | They do now but they didn't just a few days ago.
        
         | AceJohnny2 wrote:
         | As the article says, the timing on the Ethereum dropping proof-
         | of-work was shorter than Nvidia's production pipeline.
        
         | vineyardmike wrote:
         | I think we forget that Silicon manufacturing is planned a lot
         | farther out than Silicon shopping.
         | 
         | They were likely trying to make TSMC purchase orders in the
         | start of the pandemic, before a crypto boom. They also tried to
         | handicap their GPUs wrt crypto. They likely didn't expect the
         | absolute shit show of a chip shortage (because who predicted or
         | understood the pandemic early).
         | 
         | The rest of the market was desperate, and they probably
         | expected it to be more robust than it ended up being. The merge
         | would have been so far away at the time that they wouldn't
         | predict if it would happen at all nevermind when.
        
         | sophrocyne wrote:
         | Couple of comments to this point suggest that it couldn't have
         | been predicted since the official timing of the merge was only
         | announced in 2022, and silicon supply chain requires planning
         | in advance of that.
         | 
         | But that point is ignorant of this truth - Proof-of-stake has
         | been on the roadmap since ~2017 if not earlier.~ Edit: 2016 -
         | Thanks friend! :)
         | 
         | I think the reality is that the impact of Ethereum on Nvidia's
         | business was not fully appreciated, and that 'veil of
         | ignorance' may well have been intentional. They never truly
         | served the crypto market directly (e.g., there wasn't really a
         | "miner" line of cards), and as a result didn't do the due
         | diligence to understand how those customers played into their
         | business performance and strategy. Or they did, and just really
         | underestimated the Ethereum devs on ever making the merge
         | happen. But I lean towards the first.
         | 
         | Either way, I think that with crypto in the rearview, I'm
         | actually more confident in their leadership team. They seem
         | better suited to gaming and AI.
        
           | throw101010 wrote:
           | > They never truly served the crypto market directly (e.g.,
           | there wasn't really a "miner" line of cards)
           | 
           | There were deliberate, and miserably failed, attempts to make
           | lines of cards that could not be used for mining, while in
           | parallel keeping their non-limited lines in production,
           | making them the defacto miners' lines.
           | 
           | So yes they did know about it and tried to address it by
           | catering to both markets, but were unable to do it correctly.
        
             | izacus wrote:
             | What would "correct" approach look like?
        
               | sophrocyne wrote:
               | ASICs probably.
               | 
               | The point you responded to is also fair; the fact they
               | tried to lock miners out was a de facto acknowledgment
               | that crypto had an impact on their sales, and is more
               | critical evidence they handled it poorly.
        
           | [deleted]
        
           | doikor wrote:
           | Nvidia did (does?) have a miner line of cards (CMP HX).
           | Though they were mainly their server cards that failed QA but
           | could still work as a miner.
           | 
           | https://www.nvidia.com/en-us/cmp/
           | 
           | These also had models that you can't find anywhere on Nvidia
           | website like HX170 that is basically a A100 with less memory
           | 
           | A lot of miners preferred consumer cards though as those can
           | be sold to gamers once the bust comes again (and with crypto
           | it always will every few years)
        
             | sophrocyne wrote:
             | Completely unaware of this line. Thanks for sharing.
             | 
             | Frankly, doesn't paint a better picture for an Nvidia...
        
           | paulmd wrote:
           | > But that point is ignorant of this truth - Proof-of-stake
           | has been on the roadmap since 2017 if not earlier.
           | 
           | it's been on the roadmap since 2016. That's actually still a
           | problem though, a perpetually-rolling-deadline is effectively
           | worse than not having a deadline at all.
           | 
           | Was NVIDIA just supposed to cut production for the last 6
           | years in anticipation of something that was continuously
           | pushed back 6 months every 6 months? That's not a reasonable
           | expectation.
        
             | sophrocyne wrote:
             | As an observer, I never got the sense that there were
             | strong commitments being made on timelines until A) beacon
             | chain was live (running in parallel), and B) testnets
             | started getting merged successfully.
             | 
             | The moment of Genesis for the beacon-chain started a clock
             | that Nvidia should have been paying attention to, and I
             | think would have given them plenty of time to foresee the
             | present situation.
        
               | [deleted]
        
               | jonas21 wrote:
               | Regardless of whether they foresaw it or not, what should
               | they have done differently?
        
               | cma wrote:
               | Probably not much, but different investor guidance.
        
               | sophrocyne wrote:
               | Great question, and perhaps at the point where decisions
               | were being made, it'd be hard to argue a different path
               | internally (hindsight being 20/20).
               | 
               | However, it seems clear that the business built both
               | insane prices and the crypto lockup of devices (whether
               | explicitly, or implicitly) into their forecasts for the
               | business. They didn't have a good pulse on the actual
               | demand/usage of their product, and when that usage
               | pattern would shift.
               | 
               | The path they're taking right now, specifically regarding
               | pricing towards & serving higher-end enthusiasts with
               | newer products, makes sense while the used inventory gets
               | cycled around the lower end of the market.
               | 
               | From a product perspective, I don't have any useful
               | opinions to share because I'm not in hardware, and I
               | don't have the information set they're operating from
               | internally. But, they should have hoovered up as much
               | cheap capital as they could while their stock price was
               | high and the going was good to make the next period of
               | heavy investments (to be fair, shares outstanding did
               | grow, just not by a ton, %-wise, and they have a fair bit
               | of cash on the balance sheet)
               | 
               | https://www.wsj.com/market-
               | data/quotes/NVDA/financials/quart... is painful to see,
               | and I don't foresee it getting better in the next year.
        
               | paulmd wrote:
               | people absolutely SCREAMED a year ago when there was a
               | rumor going around that NVIDIA was pulling back on new
               | chip starts, it was going all around that it was a plan
               | to "spike prices during the holidays".
               | 
               | In the end 2021Q4 shipments were actually up according to
               | JPR, of course. But people were _mad_ , and I still see
               | that MLID article brought up as proof that NVIDIA was
               | deliberately trying to "worsen the shortage" and "spike
               | prices during the holidays".
               | 
               | https://www.notebookcheck.net/Nvidia-allegedly-halting-
               | RTX-3...
               | 
               | https://www.jonpeddie.com/press-releases/q421-sees-a-
               | nominal...
               | 
               | Now, what MLID may not really know, is that wafer starts
               | typically take about 6 months, so if he's hearing about
               | reduced starts in October, it's probably more like NVIDIA
               | is pulling back on expected Q1/Q2 production... which
               | indeed did come down a bit.
               | 
               | But as to the public reaction... people were fucking
               | _mad_ about _any sign_ of pulling back on production.
               | People are just unreasonably mad about anything involving
               | NVIDIA in general, _every single little news item_ is
               | instantly spun into its worst possible case and
               | contextualized as a moustache-twirling plan to screw
               | everyone over.
               | 
               | Like, would it have really been a bad thing to pull back
               | on chip starts a year ago? That actually looks pretty
               | sensible to me, and gamers will generally also suffer
               | from the delay of next-gen products while the stockpile
               | burns through anyway.
               | 
               | It's _nowhere near_ the  "sure miners may be annoying,
               | but deal with it for 6 months and then we all get cheap
               | GPUs and everyone holds hands and sings" that LTT and
               | some other techtubers presented it as. Like, yeah, if you
               | want a cheap 30-series card at the end of its
               | generation/lifecycle great, but, you'll be waiting for
               | 4050/4060/4070 for a while. Even AMD pushed back their
               | midrange chips and is launching high-end-only to allow
               | the miner inventory to sell through.
               | 
               | And people hate that now that they've realized the
               | consequence, but they were cheering a year ago and
               | demanding the removal of the miner lock / etc. More cards
               | for the miners! Wait, no, not like that!
               | 
               | It's just so tiresome on _any_ article involving NVIDIA,
               | even here you 've got the "haha linus said FUCK NVIDIA,
               | that makes me Laugh Out Loud right guys!?" and the same
               | tired "turn everything into a conspiracy" bullshit,
               | _constantly_.
        
               | smoldesu wrote:
               | And that's just the _hardware_ drama. The software hate
               | against Nvidia is partially unwarranted too - Nvidia 's
               | Wayland issues mostly boil down to GNOME's refusal to
               | embrace EGLStreams, which got whipped up into a narrative
               | that Nvidia was actively working to sabotage the Linux
               | community. The reality is that desktop Linux isn't a
               | market (I say this as an Nvidia/Linux user), and they
               | have no obligation to cater to the <.5% of the desktop
               | community begging for changes. Honestly, they'd get more
               | respect for adding a kernel-mode driver to modern MacOS.
               | 
               | In the end, Nvidia is still a business. Putting any money
               | towards supporting desktop Linux isn't going to have an
               | adverse effect on their overall sales. We're just lucky
               | that they patch in DLSS/ray tracing support to Linux
               | games and software like Blender.
        
         | colechristensen wrote:
         | I think it's more like there are so many games at play as CEO
         | in that position that anything but vague denial would be far
         | more trouble than it's worth. Anything you say is going to
         | attract a lot of criticism so the only thing you can say is the
         | least damaging one.
         | 
         | In other words, most public statements are mostly nonsense
         | engineered for response and have only a casual association with
         | the truth.
        
         | modeless wrote:
         | They have consistently underestimated the effects of crypto,
         | it's been screwing up their demand forecasts for a long time. I
         | think what happened was they had all these efforts to prevent
         | miners from buying cards so gamers could buy them instead, and
         | they thought they were successful. So they attributed strong
         | demand to gaming, but they were actually failing and miners
         | were still buying all the cards. I don't know why they thought
         | they were successful...
        
       | Night_Thastus wrote:
       | Calling it a "valley" almost seems silly. They're returning to
       | _normal_ after a boom perpetuated on smoke.
       | 
       | Nvidia will be fine. Investors don't like to see it because they
       | somehow couldn't comprehend that the growth they saw was
       | completely artificial (how is beyond me), but the company will be
       | fine.
       | 
       | Their latest decision with the 4000 series was smart though. They
       | realized suckers will pay insane amounts for the cards, even
       | disregarding prior crypto. So, make 4000 series insanely
       | expensive. That will drive sales of 3000 series to empty the
       | over-supply and make their relatively lower prices look like a
       | steal.
       | 
       | In the end, they get people to way over-pay on both the 3000 and
       | 4000 series. Double dipping!
        
         | theandrewbailey wrote:
         | Same thing happened during the last crypto boom. It was
         | impossible to find 1000 series cards, and Nvidia saw how much
         | people were willing to pay, so they priced the 2000 series
         | high, just as (former) crypto miners were selling 1000 series
         | cards.
        
         | bitL wrote:
         | Whether it is smart or not remains to be seen. AMD might step
         | up and obliterate 40x0 in price/performance.
        
           | capableweb wrote:
           | In the price-sensitive consumer space, price/performance
           | matters a lot. But all the other places,
           | libraries/SDKs/interoperability matters as much or more. Most
           | of all the Stable Diffusion stuff that is appearing is
           | heavily powered by nvidia cores, with AMD support being
           | spotty at best. Same goes for many other areas in AI/ML.
        
           | izacus wrote:
           | That's about as likely as Matrox suddenly being resurrected
           | as the leader in professional graphics. These are huge
           | complex chips.
        
             | ckastner wrote:
             | What makes you think AMD's chips are that much less
             | complex? They hold up well in benchmarks.
             | 
             | And add price to the comparison (since we are commenting on
             | price/performance), and AMD already comes out ahead of
             | Nvidia. Here's an article [1] that basically reproduces
             | AMD's on PR on this, but other sites corroborate this.
             | 
             | [1] https://videocardz.com/newz/amd-claims-to-offer-better-
             | perfo...
        
         | omegalulw wrote:
         | > They realized suckers will pay insane amounts for the cards,
         | even disregarding prior crypto.
         | 
         | Press X to doubt. People are PISSED, and the 30 series is
         | already excellent value. 4090 will sell but I have my doubts on
         | the 4080 (esp 12 GB).
        
           | izacus wrote:
           | I'm willing to bet a case of lager that 40xx will sell out
           | like hot cakes when gamer teenage angst passes.
           | 
           | Gamer community is known for teenage rage outbursts which are
           | of dubious practicality or reason.
        
             | filoleg wrote:
             | Yeah, I remember a similar fit thrown around the 20xx
             | series release not that long ago. I had a 1080Ti at the
             | time, so i didn't care to upgrade. Comes around the release
             | time of 30xx, suddenly almost everyone i knew and everyone
             | on reddit was upgrading from their 20xx series cards.
        
             | fragmede wrote:
             | Belittling the community, calling them teenagers, and
             | telling them to "get over it" doesn't actually materialize
             | $1,600 to be able to buy the card though. Especially right
             | in the middle of a period of inflation and economic
             | downturn. And then there's the competition Nvidia is in
             | with themselves given then glut of used 3090 cards on the
             | market, and the 3090 was only selling priced so high
             | because of crypto mining in the first place, and that's
             | gone now.
             | 
             | Who knows, maybe you're right or maybe Nvidia's in for a
             | valley, or Nvidia will end up dropping their prices.
        
           | Night_Thastus wrote:
           | People say they're pissed, but open their wallets all the
           | same. I wouldn't correlate internet outrage to anything real.
           | Same as every time people say they're going to boycott
           | anything online.
        
         | izacus wrote:
         | I wonder if "the investors" are mostly just AI trading
         | algorithms with impulse control and intelligence of a 6 year
         | old.
         | 
         | Nothing else explains the short sightedness of modern markets.
        
           | GekkePrutser wrote:
           | I think the investors are just hopping stocks. Nvidia was
           | generating extreme profits, now that that's over they'll jump
           | to another hypey thing. Energy perhaps. They don't care about
           | the companies they back. Just about money.
           | 
           | It's causing good companies with a long-term vision to suffer
           | (note I'm not considering Nvidia one of these) and promoting
           | the hollow-shell money grabbers and pyramid schemes.
           | 
           | I don't know how it can be solved though.. We've made this
           | situation ourselves by speeding business up to "the speed of
           | light". Perhaps a bigger role for governments in investment
           | but I know that's cursing in the American church of the free
           | market :)
           | 
           | But in the long term we really have to find some stability
           | back in the system IMO.
        
             | lmm wrote:
             | IMO the instability was always "really" there, but reduced
             | information flow hid those fluctuations. Maybe we just need
             | to get used to the instability - business is risky and will
             | have ups and downs, responding quickly ultimately makes the
             | system more efficient in the long run.
        
         | ckastner wrote:
         | > _Investors don 't like to see it because they somehow
         | couldn't comprehend that the growth they saw was completely
         | artificial (how is beyond me)_
         | 
         | This was absolutely fascinating to watch over the past 18
         | months.
         | 
         | Anyone looking for GPUs starting from February 2021 new exactly
         | what was going on. Cards were being gobbled up by miners. They
         | never made it to any shelf for a consumer to buy; webshops were
         | botted, very few humans had a chance to buy.
         | 
         | Regular consumers only got them on eBay or similar. And it was
         | blindingly obvious that consumers weren't paying four figures
         | markup for certain cards. When ETH skyrocketed to almost $5000,
         | a friend of mine was reselling 3060 Tis he bought for EUR419
         | from Nvidia's webshop, using a bot he wrote, and resold them
         | for EUR1250. His regular buyer took all cards he could get his
         | hands on (dozens per month), and resold them to central
         | European countries (popular among miners) for even more markup.
         | 
         | Again, this was blindingly obvious. Availability of cards
         | followed the ETH price; when ETH dipped in summer 2021, cards
         | became available again. When ETH went up towards the end of the
         | year, my friend was selling used 3080s for EUR1800 again. Then
         | ETH started to crash again, and suddenly Nvidia was facing a
         | massive oversupply.
         | 
         | The fact that Nvidia to this day refuse to acknowledge the role
         | that miners played in artificially inflating growth is weasily,
         | to say the least.
        
         | tomatotomato37 wrote:
         | The pricing of their cards are probably fine once you consider
         | how stupid expensive fab costs are going to be, but their PR
         | both communicating that and their weirdass 4080 naming nonsense
         | is still hot garbage.
         | 
         | I do agree though that their long-term fundamentals are fine.
         | They're still reactive enough to be competitive with AMD
         | (another company with strong fundamentals), avoided amateur
         | mistakes like going full cryptobro, and they just generally
         | positioned themselves well in the global market.
        
           | flumpcakes wrote:
           | Fab expenses haven't raised that much, meanwhile the margin
           | for AIB has gone from 30% to 5%. Nvidia is making huge
           | margins on their chips, probably the most of any company
           | selling to consumers.
        
       | Jsharm wrote:
       | Is a GPU previously used for mining worth anything in the second
       | hand market? I was under the impression it thoroughly knackered
       | them out.
        
         | causi wrote:
         | There are a number of seriously conflicting stories. Some of
         | them say "the cards were individually tuned and undervolted to
         | run at maximum efficiency to make the most money so they'll be
         | fine" and some say "these cards were overclocked and left to
         | run in a boiling hot shipping container then they washed them
         | off with a water hose".
        
         | neogodless wrote:
         | Results may vary.
         | 
         | But I used my Radeon RX 6700 XT for mining nearly 24/7 for
         | about 10 months (between purchase and when it paid itself off),
         | while using it for gaming in between (I'd obviously stop
         | mining). It ran around 65degC during that time. Very low core
         | clocks, but memory was run at close to the maximum recommended
         | speed by AMD's Adrenalin software. At least so far no signs of
         | any problems.
        
         | bentcorner wrote:
         | LTT goes over it pretty well (IMO):
         | https://www.youtube.com/watch?v=hKqVvXTanzI
         | 
         | tldr: Cards (like any piece of other electronics) do have a
         | lifespan, but mining doesn't affect that. Cards that are kept
         | clean and in better working conditions will run faster.
        
       ___________________________________________________________________
       (page generated 2022-09-26 23:01 UTC)