NVIDIA GPU market domination hits almost 100%, AMD dwindles, Intel non-existent
-
Who the fuck buys a consumer GPU for AI?
Plenty. Consumer GPU + CPU offloading is a pretty common way to run MoEs these days, and not everyone will drop $40K just to run Deepseek in CUDA instead of hitting an API or something.
I can (just barely) run GLM-4.5 on a single 3090 desktop.
… Yeah, for yourself.
I’m referring to anyone running an LLM for commercial purposes.
Y’know, 80% of Nvidia’s business?
-
At the end of the day I think it is this simple. CUDA works and developers use it so users get a tangible benefit.
AMD comes up with a better version of CUDA and you have the disruption needed to compete.
I’m not sure that would even help that much, since tools out there already support CUDA, and even if AMD had a better version it would still require everyone to update apps to support it.
-
This post did not contain any content.
AMD needs to fix their software.
I had an AMD GPU last year for a couple weeks, but their software barely works. The overlay didn’t scale properly on a 4k screen and cut off half the info, and wouldn’t even show up at all most of the time, ‘ReLive’ with instant replay enabled caused a performance hit with stuttering in high FPS games…
Maybe they have it now, but I also couldn’t find a way to enable HDR on older games like Nvidia has.
-
A persons actual experience with a product isnt real world data? Fan boys for huge companies are so weird.
Please read my entire comment, I also said your experience as one person is statistically insignificant. As in, you cannot rely on 1 bad experience considering the volume of GPUs sold. Anybody can be unlucky with a purchase and get a defective product, no matter how good the manufacturer is.
Also, please point out where I did any fanboyism. I did not take any side in my comments. Bad faith arguments are so weird.
-
Please read my entire comment, I also said your experience as one person is statistically insignificant. As in, you cannot rely on 1 bad experience considering the volume of GPUs sold. Anybody can be unlucky with a purchase and get a defective product, no matter how good the manufacturer is.
Also, please point out where I did any fanboyism. I did not take any side in my comments. Bad faith arguments are so weird.
Sure buddy, we’re all idiots for not liking the product you simp for. Got it.
-
… Yeah, for yourself.
I’m referring to anyone running an LLM for commercial purposes.
Y’know, 80% of Nvidia’s business?
I’ve kinda lost this thread, but what does that have to do with consumer GPU market share? The servers are a totally separate category.
I guess my original point was agreement: the 5000 series is not great for ‘AI’, not like everyone makes it out to be, to the point where folks who can’t drop $10K for a GPU are picking up older cards instead. But if you look at download stats for these models, there is interest in running stuff locally instead of ChatGPT, just like people are interested in internet free games, or Lemmy instead of Reddit.
-
Sure buddy, we’re all idiots for not liking the product you simp for. Got it.
Nice. Did not answer anything, did not point out where I’m simping, or being a fanboy. I’m not pro Nvidia, nor AMD, nor anything (rather than that I’m pretty anticonsumerism actually, not that you care).
You’re being extremely transparent in your bad faith.
-
What’s the VRAM cartel story? Think I missed that.
Basically, consumer VRAM is dirt cheap, not too far from DDR5 in $/gigabyte. And high VRAM (especially 48GB+) cards are in high demand.
But Nvidia charges through the nose for the privilege of adding more VRAM to cards. See this, which is almost the same silicon as the 5090: https://www.amazon.com/Blackwell-Professional-Workstation-Simulation-Engineering/dp/B0F7Y644FQ
When the bill of materials is really only like $100-$200 more, at most. Nvidia can get away with this because everyone is clamoring for their top end cards
AMD, meanwhile, is kind of a laughing stock in the prosumer GPU space. No one’s buying them for CAD. No one’s buying them for compute, for sure… And yet they do the same thing as Nvidia: https://www.amazon.com/AMD-Professional-Workstation-Rendering-DisplaPortTM/dp/B0C5DK4R3G/
In other words, with a phone call to their OEMs like Asus and such, Lisa Su could lift the VRAM restrictions from their cards and say 'you’re allowed to sell as much VRAM on a 7900 or 9000 series as you can make fit." They could pull the rug out from under Nvidia and charge a $100-$200 markup instead of a $3000-$7000 one.
…Yet they don’t.
It makes no sense. They’re maintaining an anticompetitive VRAM ‘cartel’ with Nvidia instead of trying to compete.
Intel has more of an excuse here, as they literally don’t manufacture a GPU that can take more than 24GB VRAM, but AMD literally has none I can think of.
-
I know it’s not indicative of the industry as a whole, but the Steam hardware survey has Nvidia at 75%. So while they’re still selling strong, as others have indicated, I’m not confident they’re getting used for gaming.
Everyone and their manager wants to play with LLMs and and and Intel still don’t have a real alternative to CUDA and so are much less popular for compute applications.
-
If Intel/AMD provides an LLM capable card, by the time I get around to making a big boy server, then that’s what I’ll get. Ideally Intel for the sweet AV1 encoding and Quick sync. Then again, if the card runs LLMs, then transcodes won’t be a problem. Cuda looks nice, but fuck Nvidia, I’ll go without.
Don’t look at me, I’m an Android gamer these days, and my home lab runs on integrated Intel graphics.
Maybe one day I’ll build a gaming rig, but mine’s so old now (gtx970) that I’m pretty much starting from scratch. I can’t justify the expense to make a rig from nothing when my Retroid pocket is >£300 and I can play most everything PS2/Gamecube and before.
AMDs are already LLM capable.
-
I’ve kinda lost this thread, but what does that have to do with consumer GPU market share? The servers are a totally separate category.
I guess my original point was agreement: the 5000 series is not great for ‘AI’, not like everyone makes it out to be, to the point where folks who can’t drop $10K for a GPU are picking up older cards instead. But if you look at download stats for these models, there is interest in running stuff locally instead of ChatGPT, just like people are interested in internet free games, or Lemmy instead of Reddit.
The original post is about Nvidia’s domination of discrete GPUs, not consumer GPUs.
So I’m not limiting myself to people running an LLM on their personal desktop.
That’s what I was trying to get across.
And it’s right on point for the original material.
-
The original post is about Nvidia’s domination of discrete GPUs, not consumer GPUs.
So I’m not limiting myself to people running an LLM on their personal desktop.
That’s what I was trying to get across.
And it’s right on point for the original material.
I’m not sure the bulk of datacenter cards count as ‘discrete GPUs’ anymore, and they aren’t counted in that survey. They’re generally sold socketed into 8P servers with crazy interconnects, hyper specialized to what they do. Nvidia does sell some repurposed gaming silicon as a ‘low end’ PCIe server card, but these don’t get a ton of use compared to the big silicon sales.
-
I’m not sure the bulk of datacenter cards count as ‘discrete GPUs’ anymore, and they aren’t counted in that survey. They’re generally sold socketed into 8P servers with crazy interconnects, hyper specialized to what they do. Nvidia does sell some repurposed gaming silicon as a ‘low end’ PCIe server card, but these don’t get a ton of use compared to the big silicon sales.
I wouldn’t be surprised in the slightest if they are included in the list. I dunno, I’m not the statistician who crunched the numbers here. I didn’t collect the data, and that source material is not available for me to examine.
What I can say is that the article defines “discrete” GPUs instead of just “GPUs” to eliminate all the iGPUs. Because Intel dominates that space with AMD, but it’s hard to make an iGPU when you don’t make CPUs, and the two largest CPU manufacturers make their own iGPUs.
The overall landscape of the GPU market is very different than what this data implies.
-
I wouldn’t be surprised in the slightest if they are included in the list. I dunno, I’m not the statistician who crunched the numbers here. I didn’t collect the data, and that source material is not available for me to examine.
What I can say is that the article defines “discrete” GPUs instead of just “GPUs” to eliminate all the iGPUs. Because Intel dominates that space with AMD, but it’s hard to make an iGPU when you don’t make CPUs, and the two largest CPU manufacturers make their own iGPUs.
The overall landscape of the GPU market is very different than what this data implies.
Well, it’s no mystery:
Q2’25 PC graphics add-in board shipments increased 27.0% from last quarter
Data center GPU boards were up an average of 4.7% from last quarter.
Jon Peddie Research (www.jonpeddie.com)
It’s specifically desktop addin boards:
AMD’s RX 9070 XT and RX 9070 represent AMD’s new RDNA 4 architecture, competing with Nvidia’s midrange offerings. Nvidia introduced two new Blackwell-series AIBs: the GeForce RTX 5080 Super and the RTX 5070. The company also announced the RTX 500 workstation AIB. Rumors have persisted about two new AIBs from Intel, including a dual-GPU model.
It is including workstation cards like the Blackwell Pro. But this is clearly not including server silicon like the B200, H200, MI325X and so on, otherwise they would have mentioned updates. They are not AIBs.
I hate to obsess over such a distinction, but it’s important: server sales are not skewing this data, and workstation sales volumes are pretty low. It’s probably a accurate chart for gaming GPUs.
-
Well, it’s no mystery:
Q2’25 PC graphics add-in board shipments increased 27.0% from last quarter
Data center GPU boards were up an average of 4.7% from last quarter.
Jon Peddie Research (www.jonpeddie.com)
It’s specifically desktop addin boards:
AMD’s RX 9070 XT and RX 9070 represent AMD’s new RDNA 4 architecture, competing with Nvidia’s midrange offerings. Nvidia introduced two new Blackwell-series AIBs: the GeForce RTX 5080 Super and the RTX 5070. The company also announced the RTX 500 workstation AIB. Rumors have persisted about two new AIBs from Intel, including a dual-GPU model.
It is including workstation cards like the Blackwell Pro. But this is clearly not including server silicon like the B200, H200, MI325X and so on, otherwise they would have mentioned updates. They are not AIBs.
I hate to obsess over such a distinction, but it’s important: server sales are not skewing this data, and workstation sales volumes are pretty low. It’s probably a accurate chart for gaming GPUs.
The fact that they’re not mentioned, does not constitute proof that they were not included in the statistics.