I don’t think, that Nvidia will keep its dominance in that branch for much longer. Especially when specialized chips can be run much cheaper and with less electricity cost.
Well, the reason to change is, that specifically made AI chips perform better or use less power for the same performance.
But this is a pretty new field, and because they didn’t exist (besides some intel AI sticks I have somewhere laying around), we used GPUs as the next best thing, because their architecture is well layed out for parallel computations.
So, when now new specific AI chips enter the market, Nvidia will get quite some competition, as their product is still just a GPU, and not actually designed for AI inference.
Although there have been some GPUs sold as AI specific, as far as I remember they were still the same chip and only didn’t sport the outputs - but this is just from the back of my head…
Yeah, they will need to proof that, when customers use them, but it seems a specific chip design is much more suited for the task - who could have guessed …
Nvidia, and every GPU manufacturer, got quite some competition in regards to AI, with specialized chips.
E.g. Cerebras - https://www.forbes.com/sites/craigsmith/2024/08/27/cerebras-speeds-ai-by-putting-entire-foundation-model-on-its-giant-chip/
I don’t think, that Nvidia will keep its dominance in that branch for much longer. Especially when specialized chips can be run much cheaper and with less electricity cost.
Edit: more up to date article with actual working model on their chip:
https://cerebras.ai/press-release/cerebras-launches-worlds-fastest-deepseek-r1-distill-llama-70b-inference
Never head of Cerebras. shrug
The world buys Nvidia. 98% of datacenter GPUs are made by Nvidia.
Source: https://www.hpcwire.com/2024/06/10/nvidia-shipped-3-76-million-data-center-gpus-in-2023-according-to-study/
I see no reason for that to change. Everybody is using them, knows how to use them and pretty much every AI software in existence runs on them.
Well, the reason to change is, that specifically made AI chips perform better or use less power for the same performance.
But this is a pretty new field, and because they didn’t exist (besides some intel AI sticks I have somewhere laying around), we used GPUs as the next best thing, because their architecture is well layed out for parallel computations.
So, when now new specific AI chips enter the market, Nvidia will get quite some competition, as their product is still just a GPU, and not actually designed for AI inference.
Although there have been some GPUs sold as AI specific, as far as I remember they were still the same chip and only didn’t sport the outputs - but this is just from the back of my head…
Edit: couldn’t find the article I was looking for, but here are at least some graphs showing the difference:
https://www.msn.com/en-us/news/technology/deepseek-on-steroids-cerebras-embraces-controversial-chinese-chatgpt-rival-and-promises-57x-faster-inference-speeds/ar-AA1y8g3N
Which shows much lower latency and higher tokens per second output
Yeah, they will need to proof that, when customers use them, but it seems a specific chip design is much more suited for the task - who could have guessed …