An Evaluation of the D.A. Davidson Downgrade
In late September, D.A. Davidson analyst Gil Luria cast a rare shadow over the usually sunny skies of Microsoft. He made the bold move of downgrading the technology behemoth from buy to neutral, a stark shift in the generally favorable outlook from Wall Street. Despite maintaining a price target of $475, reflecting an approximately 8% upside from the current stock price of $410, Luria’s concerns centered on Microsoft’s reliance on Nvidia’s premium-priced chips within the AI sphere.
Challenges to Microsoft’s AI Prowess
Luria’s rationale for the downgrade hinges on two pivotal points. Firstly, he posits that Amazon and Alphabet have surged forward in their AI capabilities, catching up to Microsoft’s previous lead. While debates may ensue regarding OpenAI’s continued dominance in the generative AI realm, Amazon’s investments in Anthropic and Google’s advancements with its Gemini model have, according to Luria, brought them on par with Microsoft. Noteworthy performances by Claude LLMs from Anthropic and the competitive edge claimed by Google’s Gemini model further support this contention.
Furthermore, Luria points out that Amazon and Google hold a cost advantage due to their early forays into developing proprietary AI accelerators – Google with its Tensor Processing Units and Amazon with its Trainium AI chips. In contrast, Microsoft’s lag in this arena places it at a cost disadvantage, heavily reliant on Nvidia’s costly GPUs.
Countering the Narrative: Microsoft’s Resilience and Prospects
It is crucial to acknowledge that fierce competition in the AI landscape was always on the horizon for investors in Microsoft. The recent release of OpenAI o1 by ChatGPT underscores the potential for Microsoft to maintain a slight edge amid the evolving landscape. Moreover, Microsoft’s ongoing efforts in chip design, exemplified by the Maia project, signify a bid to catch up with Amazon and Google in this domain. While currently trailing in this race, Microsoft’s ability to develop effective internal accelerators could close the existing gap.
As the AI race unfolds, all major cloud players are poised to possess competitive AI frameworks, supported by cost-efficient internal accelerators, marking a potential shift away from Nvidia chips for certain workloads. Given their robust financial standing and technical prowess, Microsoft, Amazon, and Google are well-positioned to capitalize on the AI revolution, cementing their status as solid investment options.
