AI Infra Demand Shift - Inference Market Booming For AMD?
As demand for AI services increases, there is a growing need for infrastructure supporting inference workloads.
As we already know that $NVIDIA Corp(NVDA)$ GPUs is dominating the AI training space, rivals are focusing on inference-focused hardware. $Advanced Micro Devices(AMD)$ have launched new products aimed at the booming inference market.
As we can see that Nvidia has better revenue opportunity in AI training than AI inferencing, the recent stock price decline could signal a shift. The AI infrastructure market is witnessing a significant shift in focus from the training of massive AI models to optimizing their inference—i.e., the process of applying trained models in real-world scenarios.
As the demand for AI-powered services continues to rise, there is a growing need for infrastructure that can support the rapid and efficient processing of data in real-time. This transition is driving substantial investments in specialized hardware and data center architectures designed to handle inference workloads at scale.
The Inference Boom
While AI training has historically been the most resource-intensive phase of AI development, requiring vast computational power and data, the emphasis is now shifting toward inference. Inference involves using pre-trained AI models to make predictions or decisions based on new data, and it is crucial for real-time applications like recommendation engines, autonomous vehicles, and digital assistants.
In general, GPUs are the preferred hardware for AI training due to their parallel processing capabilities. However, for smaller or less complex inference tasks, energy-intensive GPUs may be overkill, leading to unnecessary power consumption and higher costs.
AMD Well Positioned In Inferencing
Investment bank TD Cowen said Nvidia is not the only semiconductor company that will benefit from generative AI.
"AMD is well positioned to grow meaningfully with the AI total addressable market, especially initially on the inference side," TD Cowen analysts said in a report. Other potential beneficiaries for increased inference workloads include chipmakers Broadcom and Advanced Micro Devices.
Oppenheimer has a similar take on the AI inference market. "We are now entering an era in which inferencing infrastructure for generative AI has to be scaled to support every workload across centralized cloud, edge compute, and IoT (Internet of Things) devices," Oppenheimer analysts said in a report.
AMD Ryzen AI process inference models have come at the right time to take advantage of the demand for smaller workload where inference can be started and scale.
AI Stocks: Inferencing Takes Over
The "inferencing" market should boom as Fortune 500 companies now testing generative AI move into commercial deployment. They'll deploy products using cloud computing from AWS, $Microsoft(MSFT)$ , Google, Oracle (ORCL), IBM (IBM) and others.
"AI inferencing presents a larger monetization opportunity than AI training, we believe, as AI models are put into action, making predictions based on new inputs," Oppenheimer said. "This happens repeatedly across myriad devices and applications, including from voice assistants on smartphones to autonomous vehicles, and health care systems."
Meanwhile, data center operators stand to benefit from AI investments by tech companies and large corporations.
Will AMD Come Out Of Its Current Downtrend?
If we looked at how AMD have also participate in the decline brought on by Nvidia, there is a need for AMD to differentiate itself from the semiconductor pack.
With $Broadcom(AVGO)$ dropping more than 5% after sluggish non-AI sales hurt forecast. This show that AI still is in demand. As the AI market evolves, data center operators and chip makers are increasingly turning their attention to specialized hardware optimized for the high-throughput, low-latency demands of inference.
Having fallen behind rival Nvidia in the AI hardware market, AMD is accelerating its AI chip development in a bid to close the gap.
While AMD is certainly vying for a slice of the data center training segment, considering Nvidia’s dominance in the space, it is increasingly positioning its range of AI accelerators as a powerful inference solution.
The upcoming MI350 range, which is expected to be available in 2025, will reportedly be 35 times better in inference than AMD’s current generation of AI chips.
On-device AI could further accelerate the shift to novel hardware designs that prioritize cost- and energy-efficiency.
Summary
I am confident that AMD could take advantage of its accelerators that can increasingly be found in laptops and home computers as AMD seeks to gain an advantage in the nascent market for AI-optimized PCs.
Appreciate if you could share your thoughts in the comment section whether you think AMD could recover with the demand for AI chips on end user computing devices.
@TigerStars @Daily_Discussion @Tiger_Earnings @TigerWire appreciate if you could feature this article so that fellow tiger would benefit from my investing and trading thoughts.
Disclaimer: The analysis and result presented does not recommend or suggest any investing in the said stock. This is purely for Analysis.
Disclaimer: Investing carries risk. This is not financial advice. The above content should not be regarded as an offer, recommendation, or solicitation on acquiring or disposing of any financial products, any associated discussions, comments, or posts by author or other users should not be considered as such either. It is solely for general information purpose only, which does not consider your own investment objectives, financial situations or needs. TTM assumes no responsibility or warranty for the accuracy and completeness of the information, investors should do their own research and may seek professional advice before investing.
- Taurus Pink·09-06[龇牙] [龇牙] [龇牙] [龇牙]LikeReport