[HN Gopher] Nvidia Announces Financial Results for Second Quarte... ___________________________________________________________________ Nvidia Announces Financial Results for Second Quarter Fiscal 2024 Author : electriclove Score : 115 points Date : 2023-08-23 20:24 UTC (2 hours ago) (HTM) web link (nvidianews.nvidia.com) (TXT) w3m dump (nvidianews.nvidia.com) | xnx wrote: | The good new is that Nvidia's high GPU prices motivate everyone | (Intel, AMD, ARM, Google, etc.) to try and tackle the problem by | making new chips, making more efficient use of current chips, | etc. For all the distributed computing efforts that have existed | (prime factorization, SETI@Home, Bitcoin, etc.), I'm surprised | there isn't some way for gamers to rent out use of their GPU's | when idle. It wouldn't be efficient, but at these prices it could | still make sense. | NavinF wrote: | You can do that for inference, but most gamers have a single | GPU with <24GB VRAM which kinda sucks for training. 3090 or | 4090 is the minimum to use reasonable batch sizes | Uehreka wrote: | They're all pretty motivated, they've been motivated for years, | and almost nothing is happening. This situation isn't exactly a | poster child for the Efficient Markets Hypothesis. | | Every year just sounds like "Nvidia's new consumer GPUs are | adding new features, breaking previous performance ceilings, | running games at huge resolutions and framerates. Their | datacenter cards are completely sold out because they can spin | straw into gold, and Nvidia continues to develop new AI and | graphics techniques built on their proprietary CUDA framework | (that no one else can implement). Meanwhile AMD has finally | sorted out raytracing, and their consumer GPUs are... well not | as good as Nvidia's but they're a better value if you're | looking for a competitor to one of Nvidia's 60 or 70 line | GPUs!" | ericmay wrote: | > This situation isn't exactly a poster child for the | Efficient Markets Hypothesis. | | I'm unsure why you're criticizing the Efficient Markets | Hypothesis or even using it here, but you need to also | analyze this with some time horizon because the market and | marketplaces are not static. | willis936 wrote: | Their description could be used to describe the situation | in 2023, 2022, 2021, 2020, 2019, 2018, 2017, and 2016. | lotsofpulp wrote: | Efficient market hypothesis is unrelated to Nvidia's | competitors being unable to offer a competing product so far. | | https://www.investopedia.com/terms/e/efficientmarkethypothes. | .. | | > The efficient market hypothesis (EMH), alternatively known | as the efficient market theory, is a hypothesis that states | that share prices reflect all information and consistent | alpha generation is impossible. | IshKebab wrote: | There have been various attempts but you need a workload that's | basically public and also runs on a single GPU (because you | don't have NVLink or similar). | tric wrote: | > I'm surprised there isn't some way for gamers to rent out use | of their GPU's when idle. | | https://rendernetwork.com/ | | "The Render Network(r) Provides Near Unlimited Decentralized | GPU Computing Power For Next Generation 3D Content Creation." | | "Render Network's system can be broken down into 2 main roles: | Creators and Node Operators. Here's a handy guide to figure out | where you might fit in on the Render Network: | | Maybe you're a hardware enthusiast with GPUs to spare, or maybe | you're a cryptocurrency guru with a passing interest in VFX. If | you've got GPUs that are sitting idle at any time, you're a | potential Node Operator who can use that GPU downtime to earn | RNDR." | tzhenghao wrote: | > motivate everyone (Intel, AMD, ARM, Google, etc.) to try and | tackle the problem by making new chips | | Yes, there has been repeated efforts to chip at Nvidia's market | share, but there's also a graveyard full of AI accelerator | companies that fail to find product market fit due to lack of | software toolchain support - and that applies even for older | Nvidia GPUs and their compatible toolchains, let alone other | players like AMD. This isn't a hit on Nvidia, I'm just saying | things move so quickly in the space that even the only-game-in- | town is trying to catch up. | | Nvidia is also leading by being one or two hardware cycles | ahead of their competition. I'm pretty confident AI workloads | in enterprise is their next major focus [1]. I think this more | than anything else will accelerate AI adoption in enterprise if | well executed. | | To your point, I think the industry needs to focus more on the | toolchains that sit right between the deep learning frameworks | (PyTorch, Tensorflow etc.) and hardware vendors (Nvidia, AMD, | Intel, ARM, Google TPU etc.) Deep learning compilers will | dictate if we allow all AI workloads run on just Nvidia or | several other chips. | | [1] - https://www.nvidia.com/en-us/data- | center/solutions/confident... | Conscat wrote: | I am certain that several years ago, I was given an ad for | exactly such a service and even tried it out, but I cannot for | the life of me remember its name. It had some cute salad motif, | and its users are named "chefs". | | EDIT: It was just named Salad. https://salad.com/ | https://salad.com/download | WanderPanda wrote: | With interconnect being the biggest limitation these days I | don't think this would work. | xnx wrote: | I'm not familiar with all the varied uses of GPUs but it | seems like image generation could feasibly be distributed: | large upfront download of models, then small inputs of text | and settings, and small output of resulting images. | WanderPanda wrote: | For inference I agree! But training requires centralized | gradient steps | vajrabum wrote: | If you're in a data center and running large training | jobs then RDMA over Nvidia Mellanox Infiniband cards over | high speed ethernet (like 100GB) are used to ship | coefficients around without having that transfer | bottleneck in the CPU. | xxpor wrote: | 100 gig, that's considered cute nowadays. | | https://aws.amazon.com/blogs/aws/new-amazon- | ec2-p5-instances... | | 3.2 terabits. | Goronmon wrote: | _The good new is that Nvidia 's high GPU prices motivate | everyone (Intel, AMD, ARM, Google, etc.) to try and tackle the | problem by making new chips..._ | | Or their dominance leads to competition throwing in the towel | and investing resources in a market with less stiff | competition. | | I wouldn't be surprised to see AMD start to pair back | ivnestment on high-end GPUs if things continue down this path. | I would say Intel likely keeps pushing, but I'm less convinced | they can actually make much headway in the near future. | xnx wrote: | As was mentioned in another thread on a slightly different | topic, it wouldn't be surprising to see all non-Nvidia | parties unit around some non-CUDA open standard. | josemanuel wrote: | Do you mean something like OpenCL? | xnx wrote: | Exactly. More resources might get applied to improving | it. | haldujai wrote: | > I would say Intel likely keeps pushing, but I'm less | convinced they can actually make much headway in the near | future. | | It seems that Intel is making great headway on their fabs and | may somehow pull off 5 nodes in 4 years. Intel 3 is entering | high volume production soon and according to Gelsinger 20A is | 6 months ahead of schedule and planned for H2 2024. | | If they do pull this off and regain leadership that would | change outlook. | myth_drannon wrote: | vast.ai allows you to rent out gpu | TheAlchemist wrote: | What's also pretty interesting that they actually didn't sell | more chips this quarter - they ... just pretty much doubled the | prices (hence the huge margin). | | This is what having a monopoly looks like ! | | This is also why companies that manufacture their cards didn't | report any uptick in profits. I'm wondering how this play out in | some months ? Do they have any pricing power with respect to | NVidia ? Or NVidia could just switch to another manufacturer ? | thfuran wrote: | There probably isn't another manufacturer they can switch high | end stuff to. They recently tried moving at least some of their | cards to Samsung but switched back last generation due to yield | issues. | wmf wrote: | You have to distinguish between fabs and AIBs. | thfuran wrote: | If they treat their AIBs for their enterprise stuff | anything like they do in the consumer space, they don't | really have anything to worry about there (aside from the | rest of them giving up on dealing with Nvidia's BS, I | guess). | pb7 wrote: | [flagged] | parthdesai wrote: | > Raising prices means you are a monopoly? | | Not sure if you're intentionally choosing to ignore their | point, but what they meant is Nvidia can unilaterally choose | to raise the prices and customers can't do anything since | they're a monopoly. You can't just say well, i'll go to the | next shop and buy something for cheaper. | issafram wrote: | Not that it's much better, but wouldn't it be a duopoly | considering that AMD is also a big player? | | Hopefully Intel continues to improve it's GPU offerings | tric wrote: | > wouldn't it be a duopoly considering that AMD is also a big | player? | | I don't think GPUs are commoditized. You can't swap a Nvida | GPU with a AMD GPU, and get the same performance/results. | midhir wrote: | AMD seem to be catching up quickly lately. I'm running | Stable Diffusion, Llama-2, and Pytorch on a 7900XTX right | now. Getting it up and running even on an unsupported Linux | distro is relatively straightforward. Details for Arch are | here: https://gitlab.com/-/snippets/2584462 | | The HIP interface even has almost exact interoperability | with CUDA, so you don't have to rewrite your code. | capableweb wrote: | > Not that it's much better, but wouldn't it be a duopoly | considering that AMD is also a big player? | | Not sure AMD would be considered a big player, what would be | the percentage threshold for that? | | According to the Steam Hardware (& Software) Survey | (https://store.steampowered.com/hwsurvey/Steam-Hardware- | Softw...), ~75% of computers with Steam running has a NVIDIA | GPU, while ~15% has a AMD GPU. | | AMD is the closest to a competitor NVIDIA has, but they are | also very far away from even being close to their market- | share. | | I'm sure in AI/ML spaces, NVIDIA holds a even higher market- | share due to CUDA and the rest of the ecosystem. | steno132 wrote: | Nvidia's undervalued. | | Once enterprise adoption of AI picks up, demand for chips will | increase 2-3 times further. | | I'm told Nvidia's building their own fab in Southeast Asia over | the next few years. This will massively boost their output. | kccqzy wrote: | It remains debatable whether mass enterprise adoption of AI | would happen first, or Nvidia's competitors coming up with | equivalent chips would happen first. | danielmarkbruce wrote: | On the surface, it's not debatable. Enterprises are going | full steam ahead on AI. Building out an ecosystem to | challenge Nvidia seems like a decade long battle, if it's | even possible. | haldujai wrote: | What is full steam ahead for enterprises? It's not like | they're throwing autoregressive LLMs into production any | time soon. | | In any case Nvidia is expecting to ship ~550k H100s in | 2023, hardly enough to satisfy every user. | | Tesla decided to in-house. TPUv4 and Gaudi2 exceeded A100 | performance, they just never hit scale or the market and | then Hopper added optimization for transformers rendering | these chips relatively obsolete. | | Nvidia's lead is not unassailable and it seems incredibly | unlikely that they would not face serious competition | within the next 2-3 years given the $ being thrown around. | danielmarkbruce wrote: | Large enterprises are already putting them into | production. I have direct experience with it. | | It's not unassailable. But it's going to take a lot to | make _any_ difference to Nvidia 's volume or pricing, let | alone a meaningful difference. They already face serious | competitors in google and aws with TPU and inferentia, | but those competitors are at a pretty big disadvantage | for now (and others too). The cuda ecosystem is a big | advantage. Nvidia has a lot of leverage with semi | manufacturers because of volume. They spend way more on | chip R&D than their competitors in the space. They have | brand recognition. You can buy and own Nvidia chips v tpu | and inferentia. It's... a tough road ahead for | competitors. | haldujai wrote: | It's hard to imagine Nvidia will maintain what is right now | effectively 100% market share for training forever, | especially given the $ being thrown around. | steno132 wrote: | There's no competitor to Nvidia for the next 10 years. | | They've got a monopoly. And with AI's coming explosion, I'd | wager 50/50 odds Jensen becomes the world's first | trillionaire. | johnvanommen wrote: | > Once enterprise adoption of AI picks up, demand for chips | will increase 2-3 times further. | | Possibly their greatest asset, as an investment, is their crazy | high margins. Nvidia in 2023 is where Intel was in 2007, where | they could basically charge almost any price because they were | so dominant in the market. I remember when E5s were selling for | $2000 a pop and data centers were using thousands of them. | qwytw wrote: | > will increase 2-3 times further. | | That and possibly way more than that is already priced in. | Nvidia's stock is extremely expensive not because of they are | making now (which is not a lot relative to valuation, they just | barely surpassed Intel this quarter in revenue) but because | investors expect pretty much exponential growth over the next | few years.. | seydor wrote: | It s come to the point that people are begging competitors to do | something in the space. Who knows, maybe some cheap Chinese asic | that can do matrix multiplication ends up eating their lunch. | | You d think that, at the level of capitalization of tech | companies, competition would be cutthroat | zapdrive wrote: | There are a bunch of startups trying to develop AI GPUs. | Someone linked them in a comment a few days ago. | smoldesu wrote: | You're kinda underselling what exactly Nvidia is doing right | now. If any Chinese company could compete with something like | the DGX GH200, they would be building GPUs for the PRC, not | exporting them. | | There's also the problem of industry hostility, anyways. Even | _if_ Nvidia was dethroned in the hardware-space, it 's unlikely | their successor would improve the lock-in situation. It will | take an intersectional effort to change things. | TheAlchemist wrote: | Capitalization in itself is meaningless. If you have 50% of | NVidia outstanding shares, and you try to sell 10% of that, the | capitalization would crater. | | What really counts is the profit. It is pretty huge now, but | not 'that' huge (at least yet). | pier25 wrote: | So gaming is now less than 20% of their business? Holy shit. | gigatexal wrote: | A blockbuster quarter for sure with eps up 854%. | TechnicolorByte wrote: | Incredible company. It's absolutely insane how far ahead they are | with the investments they made over a decade ago. | | So nice to see a "hard" engineering (from silicon to software) | SV-founded company getting all this recognition. Especially after | what has felt like a decade of SV hype software companies | dominating the mainstream financial markets pre-pandemic with a | spate of overpriced IPOs or large ad-revenue generating mega | corporations. | kccqzy wrote: | The moniker of "hard" engineering is neither precise nor | useful. What makes engineering hard? Is solving problems with | distributed systems, even if these systems are for ads, hard? | Or do you mean hardware? In that case even Nvidia is not hard | enough since they don't fabricate their own chips. Or do you | mean designing hardware? Then what makes writing system verilog | at a desk hard but writing Python not hard? | TechnicolorByte wrote: | I admit that was a glib comment and unnecessary. | | I'm really speaking about Nvidia's ability to perform well in | both hardware and software, at chip-scale and datacenter- | scale. Also speaking of their product/business direction that | revolutionizes multiple industries (leaders in graphics with | ray tracing and AI frame/resolution sacking; leaders in AI | infra and datacenter systems, etc.) all resulting in big | impacts to their respective industries. | | You're right that many of those software-only companies do | very real engineering with distributed systems and such. I | should've been more precise and was really complaining about | the SV hype of the 2010s focusing on regulating-breaking | companies like Airbnb, Uber, wework, etc. and on companies | like Meta and Google who focus on pushing ads for their | revenue. | omniglottal wrote: | I suppose the difference is engineering something | deterministic (i.e., physics, electronics, logic) versus | something soft and indistinct (SEO, ad impressions, customer | conversion rate). | epolanski wrote: | Are they so far ahead? | | AMD GPUs get comparable results as of late on Stable Diffusion. | | Software and hardware from competitors will catch up, crunching | 4/8/16 bit width numbers is no rocket science. | johnvanommen wrote: | > Software and hardware from competitors will catch up, | crunching 4/8/16 bit width numbers is no rocket science. | | I made the mistake of buying an A770 from Intel, based on the | spec sheet. Hardware is comparable to what Nvidia is selling, | for 70% of the price. | | It's basically a useless paperweight. The AI software crashes | constantly, and when it's not crashing, it performs at half | the level of Nvidia's cards. | | Turns out that drivers and software compatibility are a big | deal, and Intel is way way behind in that arena. | david-gpu wrote: | > Software and hardware from competitors will catch up, | crunching 4/8/16 bit width numbers is no rocket science. | | I used to think like that, until I got a job there and... Oh, | boy! I left five years later still amazed at all the ever | more mind bending ways you can multiply two damn matrices. It | was the most tedious yet also most intellectually challenging | work I've ever done. My coworkers there were also the | brightest group of engineers I've ever met. | smoldesu wrote: | Nvidia has a small lead on the industry in a few places, | adding up to _super_ attractive backend hardware options. | They aren 't invincible, but they profit off the hostility | between their competitors. Until those companies gang up to | fund an open alternative, it's open season for Nvidia and HPC | customers. | | The recent Stable Diffusion results are great news, but also | don't include comparisons to an Nvidia card using the same | optimizations. Nvidia claims that Microsoft Olive doubles | performance on their cards too, so it might be a bit of a | wash: https://blogs.nvidia.com/blog/2023/05/23/microsoft- | build-nvi... | | Plus, none of those optimizations were any more open than | CUDA (since it used DirectML). | | > crunching 4/8/16 bit width numbers is no rocket science. | | Of course not. That's why everyone did it: | https://onnxruntime.ai/docs/execution-providers | | The problem with that "15 competing standards" XKCD is that | normally one big proprietary standard wins. Nvidia has the | history, the stability, the multi-OS and multi-arch support. | The industry can definitely overturn it, but they have to | work together to obsolete it. | NickC25 wrote: | Absolutely monster numbers. The aftermarket trading is up over 8% | as of right now, roughly $41 USD to approximately $513 a share. | Insane. | | Anyone who is a lot more versed in company valuation methodology | see this as being near peak value, or does Nvidia have a lot more | room to run? | epolanski wrote: | This incredible growth was already priced in at 250$. | | Now it's just crazy. | squeaky-clean wrote: | It's pretty overpriced already if you're looking at the | fundamentals, and has been for a while. But fundamentals | haven't really mattered in tech stocks for a long time. | | If you want the responsible advice, it's overpriced. If you | want my personal advice, well I bought more yesterday | afternoon. | reilly3000 wrote: | It's basically a meme stock now. I don't think anyone should be | surprised by wide swings and irrational pricing going forward | into the next few months. | vsareto wrote: | I don't think the market leader for graphics cards -- a | technically complex product compared to a bunch of brick | stores selling video games -- is what you can consider a meme | stock | pb7 wrote: | What makes it a meme stock? It's printing money from an | industry that is only starting. This isn't crypto nonsense. | kelvie wrote: | (Not sure if it's true), but a meme stock is one whose | price is propped up by retail traders, and spreads through | social media / word-of-mouth, as memes do. | | How we prove it's one is probably another matter. | pb7 wrote: | Retail investors make up a low single digit percent of | individual stock ownership. /r/wallstreetbets is not | putting even a dent in a $1T company's stock price. | fnordpiglet wrote: | Yeah, every company of any note is planning how to use AI, | and a lot of the use cases are already proved out. This | isn't speculative nonsense. The question is how big does it | get, not will it be big. | | Crypto and blockchain never had an actual proved out use | case. There was an interesting idea but no one ever could | figure out a way it was useful. The costs associated were | much higher than the risks of not using it. | | People who think this is a meme aren't paying attention, | and they're certainly not in the rooms of power where AI | planning is happening at megacorps. I've been in them, and | it's serious and material and we are just now beginning to | scratch the surface. | danielmarkbruce wrote: | Top line growing 100% a year, faster recently..... Doesn't | take long for $50 bill pa to turn into 1 trillion pa at that | rate... | mikeweiss wrote: | In my opinion it's likely mostly pull forward demand. Companies | are racing to buy as many chips as possible and hoard them. | | I already saw a few posts here on HN from companies that threw | down insane amounts of $$ on H100s and are now looking to rent | out their excess capacity. I'm guessing we'll be seeing a lot | more posts like that soon. | mholm wrote: | Nvidia is the pickaxe seller in a gold rush. Their valuation is | very much tied to how big AI grows in the next several years, | and how quickly competitors can arise. I could easily see them | continuing to go up from here, especially if AI keeps on | expanding utility instead of leveling off as some fear. | tmn wrote: | Valuation fundamentals don't justify current prices. That said | it could easily go higher (much higher). Passive investing has | created a constant bid that has significantly distorted price | discovery compared to pre passive era. | rvz wrote: | > The aftermarket trading is up over 8% as of right now, | roughly $41 USD to approximately $513 a share. Insane. | | 8% is close to nothing in stocks. Biotech stocks go up and down | more than that without earnings announcements. | | > Anyone who is a lot more versed in company valuation | methodology see this as being near peak value, or does Nvidia | have a lot more room to run? | | As long as fine-tuning, training or even using these models are | inefficient and no other efficient alternatives to that without | these GPUs, then Nvidia will remain unchallenged unless that | changes. | | EDIT: It is true like it or not AI bros. There are too many to | list. For example, just yesterday: | | Fulcrum Therapeutics, Inc. (FULC) 38% up. | | China SXT Pharmaceuticals (CM:SXTC) down 25%. | | Regencell Bioscience Holdings (RGC) 28% up. | | NanoViricides (NNVC) up 20%. | | Armata Pharmaceuticals (ARMP) down 23%. | | [0] https://simplywall.st/stocks/us/pharmaceuticals-biotech | pb7 wrote: | Biotechs are lottery tickets, not stocks. You're just | gambling on binary results. | rvz wrote: | > Biotechs are lottery tickets, not stocks. | | Please. | | Stocks are lottery tickets and Biotech stocks are stocks. | | > You're just gambling on binary results. | | The risks are no better than most of the AI bros buying | Nvidia and overpriced stocks at the very top or all time | highs or extremely risky 0DTE strategy trades on earnings | announcements. | | Do AI bros who jumped in late really have to be married to | their stocks that are already overpriced to make 8% on | earnings when the very early folks start selling to take | their profits? | gorenb wrote: | Stocks are lottery tickets... | rvz wrote: | Exactly. | | Nvidia is just one of many lottery tickets and 8% in one | day is hardly volatile in stocks. | pb7 wrote: | Not if you understand what stocks are and how betting on | biotech stocks is not a wise investment. | rvz wrote: | My point is, 8% on earnings is hardly volatile. | | > Not if you understand what stocks are and how betting | on biotech stocks is not a wise investment. | | So you're giving investment advice for putting money in | NVDA stock at the top or all time highs, right now on | earnings as a 'wise investment' to make 8% (after hours) | when others are clearly taking their money out of the | market. | | Unless you already invested in NVDA stock last year, that | move is gone and you're just telling retail late comers | to throw money at NVDA at the top for others to take | their profits. | mikestew wrote: | Whelp, I guess those September NVDA call options I sold are going | to get exercised. Who woulda guessed after the crypto fallout | that "AI" would come along and bump the price back up. | | Record revenues, and a dividend of $0.04 on a $450 stock? That's | not even worth the paperwork. For example, if you bought 100 | shares, that's $45K. From that, around September $4 will show up | in your account, which you have to pay taxes on. So $3 or so net | on a $45,000 investment. Sure, there were stock buybacks, but why | keep the token dividend around? | kinghajj wrote: | Should have sold a call credit spread instead! | | For large shareholders, the dividend would still be worthwhile. | From what I could find, Jensen has 1.3 million shares, so he'd | receive over $200k in dividends this year. You might think | that's chump change, but another source lists his salary at | just under $1m; another 20% bump in liquid income is nothing to | sneeze at. | mikestew wrote: | _Should have sold a call credit spread instead!_ | | I'll get right on that...after I go look up what that means. | :-) I'm but a simple options trader who sells calls to unload | stock I didn't want anymore anyway, and the premium is the | icing on that cake. Left some money on the table this time, | but I otherwise would have just sold the shares outright, and | I did make some bank regardless. | | Gonna be missing that sweet, sweet $0.04 dividend, though. | kinghajj wrote: | A call credit spread simply means buying an even more out- | of-the-money call along with the one you sold. It would | have reduced the premium collected, but the long call would | appreciate on sudden moves like today's. | kikokikokiko wrote: | Theta gang ftw. But I would advise you to stay away from | NVDA, as soon as the first quarter with flat or decreasing | revenues comes (and it WILL come), the fall would be one to | tell your grandchildren about. | loeg wrote: | This benefit is basically only to large shareholders who | can't sell stock. Which might be insiders like Jensen and... | anyone else? Everyone else can just sell, like, 0.0001% of | their stock or whatever. | catchnear4321 wrote: | many times what a lot of people make in a year is nothing to | sneeze at. | | especially when it is awarded for merely having a stack of | papers. | Vvector wrote: | The stock is up 9% or $45/share after hours. Jensen just made | $58 million. $200k doesn't pay his dry cleaning bill. | haldujai wrote: | > Should have sold a call credit spread instead! | | Why? | | > From what I could find, Jensen has 1.3 million shares, so | he'd receive over $200k in dividends this year. You might | think that's chump change, but another source lists his | salary at just under $1m; another 20% bump in liquid income | is nothing to sneeze at. | | Jensen Huang is worth $42 billion and has been a billionaire | for probably a decade or so now? Any CEO with that net worth | would use stock-secured loans/LOCs for liquidity. 200k is | very much chump change. | thomas8787 wrote: | Jensen is one of the largest shareholders. With over 80 million | shares that's an over 3 million dollar dividend for him. | epolanski wrote: | Wait 80M shares? He's worth 4B $ then. Not bad. | tyre wrote: | just wait until it's $4bn and another $3m! | pyrrhotech wrote: | He's worth way more than that. | https://www.bloomberg.com/billionaires/profiles/jenhsun- | huan... | _zoltan_ wrote: | I sold 600C for this Friday an hour or so before earnings. Free | money with 168% IV. | HDThoreaun wrote: | Up more than 10% after hours compared to close yesterday. I | really thought NVDA had hit its ceiling at $1+ trillion, | apparently not. Really does feel like a huge opportunity for | Intel to me. They have the fab capacity to pump out at least | reasonably competitive GPUs if they can figure out the software | side of things. | | P/E still above 50 even after the AI craze 9x'd eps this quarter. | Still hard for me to see that valuation ever makes sense but what | do i know. | UncleOxidant wrote: | Intel doesn't seem to be able to execute. It's not just pumping | out GPUs - for AI you need drivers, and the equivilent of CUDA | and all the various libraries built on CUDA like cuDNN. They do | have OneAPI but it hasn't caught on like CUDA in that space. | It's kind of too bad since OneAPI is open and CUDA is not. | highwaylights wrote: | I can really see Intel figuring this out. A lot of people on | HN talking about Intel as an also-ran just like they spoke | about AMD before Zen. | | Raptor Lake is at 7nm and incredibly competitive there (~2700 | single core on geekbench, taken with a pinch of salt). | They're still planning on being on 1.8nm/18A within 2 years, | while at the same time ramping up their GPU efforts (albeit | using TSMC for 4nm). Nvidia is very much in the lead, but | this is just the beginning. | | tldr; I ain't hear no bell. | andromeduck wrote: | The problem with Intel is: | | 1. They don't pay - Nvidia/Google/Apple easily pays 1.5-2x | Intel before appreciattion. | | 2. They're cheap/beaurcratic. The office sucks, your laptop | sucks. | | 3. They suck at software. | https://pharr.org/matt/blog/2018/04/18/ispc-origins | | 4. They can't develop/retain talent. Half the ML-HW/FW | teams at AMD/Google/Nvidia/Apple are ex-Intel. | HDThoreaun wrote: | Right but the market is saying that a dominant GPU business | is worth more than a trillion dollars. Just hard for me to | believe that they can't get the business off the ground with | that kind money on the table. Can't they just hire all of | nvidia's developers and pay them 5x as much? | UncleOxidant wrote: | > Can't they just hire all of nvidia's developers and pay | them 5x as much? | | Lol... Intel is famously stingy when it comes to salaries. | i_have_an_idea wrote: | You have no idea. There are a lot of senior engineers at | NVDA making 7 figures annual total comp. How many are | there at Intel? | HDThoreaun wrote: | for a trillion dollars though... eventually you have to | believe Pat gets fired and replaced by someone who is | 100% all in on GPUs if he can't figure this out | orzig wrote: | Maybe everything changes at $1 trillion, but I definitely | see smaller (but public) companies leaving money on the | table because it would require cultural change. | UncleOxidant wrote: | I'd argue that Intel being stingy with salaries is a big | part of why they're so behind here. They just don't seem | to be very serious about this. Intel has made several | runs at the GPU market over the years and they just keep | ending up where they are. And now NVDA has such a huge | advantage (software and hardware) that it just gets | harder and harder (and more expensive) to overcome. | | Probably Intel's best bet now would be to try to be the | fab for NVDA. | andrepd wrote: | The market is also saying that Tesla is worth more than | BMW, VW, Audi, Mercedes, Toyota, Hyundai, Fiat, Ford, and | dozens of others _combined_. Mehh, I don 't know. | UncleOxidant wrote: | Exactly, the market isn't always rational. There's a lot | of work on quantization in neural nets, for example, that | can allow them to work sufficiently well on less capable | hardware. It could be that some breakthrough there would | obviate the need for NVDA hardware (or at least reduce | how many are needed). | lotsofpulp wrote: | It is rational if you interpret market capitalization and | share price movements as "the market is saying Tesla WILL | be worth more than x,y z combined between time now and | time whenever you want to sell it." | | For different people, the timespan between now and when | they may want or need to sell it is different, and thus | different people will arrive at different conclusions. | | And note that "worth more" above simply means growth in | market capitalization, so as long as someone is willing | to buy the shares at a price supporting that increased | market cap, then it does not matter if Tesla is still | selling fewer cars than the others combined. | scrlk wrote: | "In the short run, the market is a voting machine but in | the long run, it is a weighing machine." | TheAlchemist wrote: | Exactly this ! | | And as somebody with a significant short position, I | would add - "The market can stay irrational longer than | you can stay solvent" ! | | Their numbers for this and next Q are absolutely amazing. | It's also quite "refreshing" - a company with great | product, almost without competition (so far - it will | come real quick). And fun part being their main advantage | is probably CUDA and not even the chips itself (which by | the way they don't manufacture - they "only" do the | design). | | But still - even with those numbers, and even with this | pace of growth (both being absolutely not sustainable, | and will probably reverse hard next year) - the valuation | doesn't make any sense, especially given the current | interest rates. | peanuty1 wrote: | And Rivian has a greater market cap than Nissan. | nemothekid wrote: | > _Can 't they just hire all of nvidia's developers and pay | them 5x as much?_ | | As time goes on I don't see how you break the CUDA moat | even if you had all of nvidia Al's engineers. | | CUDA means you need everyone in AI to target your new | (hopefully open) platform and that platform is faster than | CUDA is. Given how most frameworks of the last 10 years | have been optimized for CUDA you would need to turn around | a global sized cruise ship. | | If Intel's GPUs are only 3% faster, will that be enough to | rewrite my entire software stack for something not CUDA? If | intel opts for a translation layer, could they ever match | nvidia's performance? | HDThoreaun wrote: | Well I'm not super experienced with GPU development but | aren't most people using packages built on top of CUDA | like pytroch etc? Would it be impossible to throw tons of | resources at those packages so they handle whatever intel | comes up with as well as they handle CUDA? | | If Intel is 10% slower but 50% cheaper and the open | source stack you use has been heavily updated to work | well with Intel drivers would that not be an enticing | product? | UncleOxidant wrote: | Intel's been trying this for several years now (OneAPI | and OpenVINO), but so far they haven't gotten the | traction. CUDA is just really entrenched at this point. | michaelt wrote: | _> aren 't most people using packages built on top of | CUDA like pytorch etc?_ | | Yes, and in fact both AMD and Intel have libraries. You | can run Stable Diffusion and suchlike on AMD GPUs today, | apparently. And you can export models from most ML | frameworks to run in the browser, on phones and suchlike. | | _> If Intel is 10% slower but 50% cheaper [...] would | that not be an enticing product?_ | | Sometimes, yes. Some of the largest models apparently | cost $600,000 in compute time to train [1], so halving | that would be pretty appealing. | | However, part of the reason for nvidia's dominance is | that if you're hiring an ML engineer for $160,000/year | spending $1,600 to give them an RTX 4090 is chump change. | | [1] | https://twitter.com/emostaque/status/1563870674111832066 | tgma wrote: | - Can't they just hire all of nvidia's developers and pay | them 5x as much? | | No. | Mountain_Skies wrote: | Over the past decade Intel seems to have become more | interested in social causes than in technology, maybe with a | side of government backrubbing to keep some income flowing. | UncleOxidant wrote: | Nah, the biggest problem is that Intel became very risk | averse. Yeah, they'll talk a good game on taking risks, but | when it comes down to it people who took risks that failed | tend to not be at Intel and other employees see that and | think that maybe they need to play it safe. | johnvanommen wrote: | > Yeah, they'll talk a good game on taking risks, but | when it comes down to it people who took risks that | failed tend to not be at Intel and other employees see | that and think that maybe they need to play it safe. | | I worked at Sears corporate when Amazon was getting big, | about 25 years ago. | | Always made me chuckle when armchair quarterbacks on TV | would wonder why Sears couldn't do what Amazon did. | | Bezos took _tremendous_ risks in the late 90s and early | 00s, while Sears was trying to figure out how to wring a | few more pennies out of their stores. Sears Corporate was | 110% focused on taking the existing business and | maximizing profits, not on innovation of any kind | whatsoever. | marricks wrote: | Why, and what does it mean, for Nvidia to announce fiscal results | a year ahead of time. | | Is it just promise to sell chips in advance, so that's how far | it's booked, do they own a Time Machine...? | danielmarkbruce wrote: | They announced Q2 results, ending July 31. Their fiscal year is | a little unusual, it ends at end of Jan. So their 2024 year | ends Jan 31 2024. | scrlk wrote: | Financial years are named by the calendar year that they end | in, so FY24 is the financial year ending in 2024. | lotsofpulp wrote: | I have never seen it referred to as financial year until now, | but I guess it makes sense too. Fiscal year is the typically | used term. | scrlk wrote: | Looks like it depends on where you are in the world. | "Financial year" appears to be the preferred phrase over in | the UK. | epolanski wrote: | Every company I know of estimates future revenue. | | It's not black magic, they have contracts in place and know | both how many GPUs will be produced and sold give or take few | %s. | rightbyte wrote: | Seems like the shovel seller is on top of this AI thing? | [deleted] | solardev wrote: | I miss the small graphics company that used to care about gamers | :( | fnordpiglet wrote: | Well, they still make gamer cards. As a company with more than | one employee they are able to multitask, and the knock on | benefits of all the investment will improving their gaming | products as well. I think there are a fair amount of dual use | cards being sold - I know I've got a 4090 that I use for local | AI stuff, and it renders RTX Witcher 3 like a beast. | unpopularopp wrote: | I actually have current gen GPUs from all 3 manufacturers | through my job and I'm glad there are choices now but I'd still | recommend Nvidia over AMD or Intel to anyone. Of course it | depends on the budget, the games you play etc. but DLSS alone | is such a difference that AMD still couldn't catch up with. I | really hope Starfield will deliver because that will be the | first game with FSR3.0 and introducing the technology, yet | DLSS3.5 was just revealed yesterday. It's a huge gamble for | sure going all in on Starfield but tbh that's one of the hypest | game of the year so worth it. And Intel is nowhere near that | (apart from the price and getting 16GB for cheap) | FirmwareBurner wrote: | Intel has entered the chat. If you wanna game on abudget with | lots of VRAM go for A750 or A770. | bozhark wrote: | Intel has left the chat. | FirmwareBurner wrote: | Intel is very much in the game. Every of their recent big | driver update ads double digit performance boosts on AAA | titles. | helf wrote: | [dead] | beebeepka wrote: | When was that? Surely it must have been at least a decade | before the GTX 970 "4GB" but maybe after all the driver | cheating in the late 90s and early 2000s. | | I no longer buy nvidia hardware but I do enjoy stock price | getting higher. I just wish I had the sense to buy more, a lot | more, stock when it was much cheaper. How does a chicken shit | like me make big money :( | grouchomarx wrote: | It's a tough game. Gotta have the guts to get in and stay in | wmf wrote: | The 970 was amazing for gaming; the 3.5GB problem was just | for CUDA. | TechnicolorByte wrote: | Nvidia is dragging the entire gaming industry forward with | ray/path tracing and AI-based resolution and frame scaling. | Everyone else (I.e., AMD) is following Nvidia's lead. | | In what way has Nvidia "forgotten" gamers with the rise of | their datacenter business? | wudangmonk wrote: | Raytracing isn't a thing no matter how much nvidia wants to | push it. The performance penalty is too big for what amounts | to something that takes a trained eye to notice. AI- | resolution scaling is nice to have on lower end devices but | the max resolution people actually use is 4k and I can only | think of VR where having more than 4k would be nice to have. | | My main gripe is that at 4k resolution, top of the line GPUs | shouldn't be using AI frame scaling to get decent fps unless | you are taking the raytracing penalty for funsies. | TechnicolorByte wrote: | Feels like this comment is stuck in 2019 or something. Have | you seen DLSS3.5 announced yesterday with ray | reconstruction? Have you seen path tracing in CP2077? | | Seems like you're really dismissing the massive speed ups | these past few years. Agreed that ray tracing in games is | only at the beginning. A lot of that is gated by the | consoles/AMD but that's generally how it goes. Would love | to see Nvidia in one of the powerful consoles to accelerate | adoption of these technologies. | tracerbulletx wrote: | The card prices have gone up pretty significantly and | availability has been bad for the last few years, they also | have been segmenting their product line in ways where some of | the lower tier cards are not very compelling vs previous | release cycles. I don't know if that's attributable to them | "forgetting about gamers" but it's what people are upset | about. | cma wrote: | Compare price performance and it isn't so bad, assuming you | add in an adjustment for AMD's lack of features. | rvz wrote: | So even without crypto, the prices of GPUs are still | expensive regardless. | | The hoarding isn't going to stop unless there are either | efficient alternatives that are competitive on performance | and price. | | Perhaps that is why I keep seeing gamers crying over GPU | prices and unable to find cheap Nvidia cards due to the AI | bros hoarding them for their 'deep learning' pet projects. | | So they settle with AMD instead. | epolanski wrote: | Raw performance isn't increasing much, price/performance | under the 700$ has barely increased both now and in 2000 | series. | cma wrote: | It's a combination of algorithms and hardware, but | raytracing has gone from path traced quake 1 to path traced | Cyberpunk 2077 in just a few years. The raytracing side of | things hardware wise has doubled in perf for the same tier | card each generation. | solardev wrote: | Gamers don't have datacenter budgets | ancientworldnow wrote: | Gamers like to ignore inflation and increasing fab costs | and pretend cards should cost the same forever with double | performance gains every 1.5 years. | Mountain_Skies wrote: | For much of tech hardware world, declining costs and | increasing performance have been the general trends for | as long as most of us have been alive. | sgarman wrote: | Off the top of my head the only thing I can think of that | did that was TVs. | [deleted] | samspenc wrote: | I think both this comment and GP comment are true in their | own ways. Nvidia is still pushing the gaming / 3D industry | faster than its competitors and I would still recommend an | Nvidia card for reliability and performance over others. | | BUT that comes at a price - Nvidia consumer chips are also | notoriously expensive, but if you want best-of-breed for | gaming, it does come at a price. | | I am hoping that AMD and Intel will be able to compete with | Nvidia someday but I'm not holding my breath. | anjel wrote: | I've seen a regular stream of reports on HN about people "sort | of" getting AI done on laptops and non and lowly GPU machines. Is | it unreasonable or far-fetched to imagine that someone figures | out how to efficiently get it all done without GPUs and pull the | rug out from under Nvidia? | bob1029 wrote: | I have an options strategy that is riding on this possibility | right now. | | All you have to do is take 5 seconds in a typical code base to | determine that the way we write software today isn't exactly... | ideal. Given another 6-12 months, I cannot comprehend another | ~OOM not being extracted somewhere simply by making the | software better. | SpacePortKnight wrote: | Just like existence of MariaDB does not prevent Snowflake from | being worth $50B, just being good enough on laptop is not | enough to replace the need for the cutting edge. | lsh123 wrote: | Training is very expensive and requires GPUs. What you read | about is running trained model on consumer devices (even | phones!). | FreshStart wrote: | Let's assume for a moment you could sort of trade parallel | computation for vast space, fast search and retrieval. So in | this hypotheticals computational theory, you could build a | lookup machine from CPUs and ssds.. squeezing the parallel | cores into one CPU, by squeezing the shaders running into a | million hashes.. And before you know it your simulating a micro | verse trying desperately to find out how to avoid climate | change. What if God hates recursion? | nabla9 wrote: | Their revenues are seriously supply restricted. ~2x revenue if | chip manufacturing could keep up with demand. Packaging seems to | be the bottleneck just now. | [deleted] | nathias wrote: | selling shovels for a few different gold rushes seems to be | profitable | parhamn wrote: | A tangent to this I've been thinking about quite a bit is how big | a moat drivers are to the software/hardware ecosystem. | | They're a major moat/hurdle (depending on your perspective) for | operating systems, new hardware platforms, graphics cards, custom | chips, and more. | | It's interesting to think that we're not _that_ far from being | able to generate decent drivers for things on the fly with the | latest code gen advancements. Relevant to this, that could reduce | the monopolies here, but perhaps as interesting is we can have | more new complete OSes with more resources allocated to the user | experience vs hardware compatibility. ___________________________________________________________________ (page generated 2023-08-23 23:00 UTC)