
Nvidia (NASDAQ:NVDA | NVDA Price Prediction) may be about to remind the AI world who still runs the show. Reports circulating this week suggest the company is preparing a new chip specifically targeting the inference market, and Jim Cramer took notice on his program.
Hereโs what Cramer said last night:
โNvidia is about to announce a new chip that can compete with pesky offerings from competitors. Several companies, including long standing Nvidia customers, have been bragging about how they make a particular kind of AI semiconductor ones for inference. And Nvidia, well, it looks like they may have something against that that could be better.โ
Nvidia shares rallied nearly 3% on the news. Theyโve given up part of those gains today, but thatโs amid a broad market sell-off that saw the Dow down more than 1,000 points in early trading. Letโs look at what NVIDIAโs new inference chip is and how it could change the battle between the company and rivals likeย Broadcomย (Nasdaq: AVGO) and Alphabetย (Nasdaq: GOOG)(Nasdaq: GOOLG).
Why Inference Is the New Battleground
Training is the expensive, power-hungry process of building an AI model. Inference is what happens every time someone uses that model: every query, every response, every decision. As AI moves from labs into products, inference volume explodes.
Nvidia has long dominated AI training hardware, but inference favors efficiency over raw throughput, which is exactly why competitors have found an opening. Companies like Broadcom have argued that NVIDIAโs GPUs arenโt specialized enough for inference and will soon prove to be too expensive.
Recently, NVIDIA purchased the IP and most the employees from startup Groq.
The specifics of Groqโs technology are fairly technical, but thereโs the key idea. Groq has been working on an entirely different architecture than NVIDIAโs GPUs.
In short, it utilizes a compiler that pre-plans operations. So instead of needing to coordinate high-bandwidth memory, Groqโs chips execute a schedule using on-chip SRAM.
The downside to this architecture is that pre-compiling isย difficult.ย You need chips to be perfectly synchonized, which is an incredibly difficult engineering challenge. NVIDIA has presented innovations at recent conferences that strongly hint at the company discovering a solution to synchonize Groqโs chips.
So, the market is putting two and two together. NVIDIA shelled out $20 billion for Groq, and plans to utilize its โclock-forwarded die-to-die linksโ to commercialize Groq. Itโs expected NVIDIA could announce this new chip built for inferencing workloads at GTC, which takes place later his month.
What this Means for NVIDIAโs Competitors
I know the above portion was fairly technical, but hereโs the key takeaways.
- This isnโt for training:ย Groqโs architecture isnโt built for training models but rather inferencing. It gives NVIDIA a potential strategic โmasterstrokeโ because it would allow the company to bypass markets like high-bandwidth memory that have severe shortages. In short, if NVIDIA releases a new inference system based on Groqโs technology, it could allow the company to further surpass Wall Street targets in the upcoming years as NVIDIA would avoid supply constrains.
- Blunts competitive threats:ย As stated earlier, the argument from companies like Broadcom and Marvell has been that NVIDIA chips have too many โjack of all tradesโ features and arenโt built to be competitive in inferencing over time. They argue custom chips will have better economics. This allows NVIDIA to attack those complaints head on.
Weโll have to see whatย exactlyย NVIDIA announces at GTC, but it does seem as though the company may have pulled off another brilliant move purchasing Groq. Right now Jim Cramer is talking about chips from the company, but after GTC expect talk of NVIDIAโs new chips to be all over the financial world.



