Cloudflare’s NET AI inference strategy has been different from hyperscalers, as instead of renting server capacity and aiming to earn multiples on hardware costs that hyperscalers do, Cloudflare maximizes system efficiency and utilization per capital expenditure dollar spent on developing the infrastructure, hence optimizing the cost structure of AI inference.
While hyperscalers are dealing with the GPU utilization paradox, Cloudflare is maximizing GPU utilization and minimizing overhead cost. Cloudflare has been using a custom large language model (LLM) inference engine written in Rust, namely Infire, built explicitly for its hardware and edge network.
The system has been able to use fewer CPUs and GPUs for more throughput, run models closer to users and improve on startup speed and efficiency, while the hyperscalers still battle with high latency, underutilized GPUs due to CPU limitations and networking bottlenecks. Here’s how Cloudflare is solving these bottlenecks with Infire.
Cloudflare’s Infire works through an OpenAI-compatible HTTP server, a batcher, and an LLM inference engine on which the models run. Infire downloads the model weights from R2 storage when models are scheduled to run. Furthermore, Infire allows these weights to be cached locally on the edge node for faster loadups in the future and start inference quickly.
Cloudflare’s supply chain is also highly optimized as it uses off-the-shelf hardware, especially in tier-1 cities, which allows it to quickly set up and start generating revenues before fully paying for the hardware, lending the company flexibility and a fast response time when capacity needs to be added.
How Competitors Fare Against Cloudflare
When it comes to AI inference and edge deployment, Cloudflare’s strategy is very different from hyperscalers, traditional cloud and inference providers like Amazon AMZN and Microsoft MSFT. Hyperscale data centers built for Amazon Web Services and Microsoft Azure use large-scale facilities for high-volume data processing, data storage, and massive workloads. However, this method suffers from higher power consumption and latency.
Amazon Web Services is solving this with the introduction of Lambda@Edge as a feature of Amazon CloudFront that lets users run code closer to their application, which massively improves performance and reduces latency. Microsoft has taken a different strategy by using a hybrid cloud strategy, where Microsoft allows its customers to run AI workloads on-premises at the edge.
NET Price Performance, Valuation and Estimates
Shares of Cloudflare have risen 9.9% in the past six months against the Zacks Internet – Software industry’s decline of 3.1%.
Cloudflare 6-Month Price Performance Chart
Image Source: Zacks Investment ResearchFrom a valuation standpoint, Cloudflare trades at a forward price-to-sales ratio of 26.19X, much higher than the industry’s average of 4.86X.
NET Forward 12-Month (P/S) Valuation Chart
Image Source: Zacks Investment ResearchThe Zacks Consensus Estimate for Cloudflare’s 2025 earnings implies year-over-year growth of 21.3%. The estimate for 2025 has been revised upward in the past 30 days.
Image Source: Zacks Investment ResearchCloudflare currently carries a Zacks Rank #2 (Buy). You can see the complete list of today’s Zacks #1 Rank (Strong Buy) stocks here.
Want the latest recommendations from Zacks Investment Research? Today, you can download 7 Best Stocks for the Next 30 Days. Click to get this free report
Amazon.com, Inc. (AMZN): Free Stock Analysis Report Microsoft Corporation (MSFT): Free Stock Analysis Report Cloudflare, Inc. (NET): Free Stock Analysis ReportThis article originally published on Zacks Investment Research (zacks.com).
Zacks Investment Research