- Nexan Insights
- Posts
- AI Chips, NVIDIA, and a Game of Inference
AI Chips, NVIDIA, and a Game of Inference
ASIC vs GPU in a Clown Car

The image provides a comparative analysis of GPUs vs. ASICs in AI workloads, visualizing their strengths and trade-offs within the evolving landscape of AI infrastructure. Inspired by the metaphor of a clown car balancing speed, efficiency, and scalability, this breakdown highlights the ongoing rivalry between NVIDIA’s CUDA-powered GPUs and specialized ASICs optimized for inference.

Imagine you're driving a clown car. You're trying to balance speed, efficiency, and the sheer weight of clowns (they keep piling in, unannounced). That's the reality AI companies face when choosing between GPUs, ASICs, and those other smart-sounding acronyms. Let's dive into this clown car journey—one filled with silicon, liquid cooling, and a market that may, or may not, stay under NVIDIA's thumb.
1. GPUs vs. ASICs - The Smackdown in the Data Center
Let's start with the most immediate rivalry: GPUs versus ASICs. It's like David versus Goliath, except both of them are actually Goliaths, and they take turns pretending to be David depending on the use case. GPUs, like NVIDIA's trusty CUDA-fueled workhorses, dominate training workloads in AI—they're generalists with incredible flexibility, like a chef that can make a mean risotto, but also bake cakes, roast chickens, and even run a farm in their spare time.
ASICs, on the other hand, are like a singularly obsessive risotto chef. If you want to do just one thing (inference), and do it really well, ASICs shine. They're optimized to drive a specific task—in this case, efficiently using trained models to make predictions. The downsides? They lack the mature software stack and flexibility of GPUs, and using them often feels like trying to tune an engine with a toolkit meant for a different model.
In recent years, cloud providers like AWS, Google Cloud, and Microsoft Azure have heavily invested in GPUs (Graphics Processing Units) to fuel AI workloads. But as the industry scales, the competition has transformed into an arms race for GPUs, with each provider vying for technological dominance, operational efficiency, and cost-effectiveness.
For investors, the question isn’t only about who has the most GPUs but who can maximize efficiency and scale profitably. For example, AWS has an advantage in global infrastructure, but how does this translate into revenue growth and profit margins when GPU demands surge? Understanding how GPU capacity directly impacts scalability, cost structure, and AWS's positioning against other giants can reveal where the future investment potential lies.
Comparative Analysis of GPU Allocation Across Leading Cloud Providers

GPUs vs. ASICs: The AI Smackdown

2. Why Compile When You Can CUDA?
When it comes to training models, NVIDIA’s CUDA platform has been king for over two decades. CUDA is like a well-stocked kitchen—any chef, even an inexperienced one, can whip up something delicious. They’ve got libraries, reusable components, and a lot of shortcuts for ease of development.
Contrast this to developing software for ASICs. Imagine moving into an empty house, the kind you build from scratch, starting with the foundation. ASIC compiler software is nowhere near CUDA in maturity. Engineers end up writing parts of the stack from scratch, sometimes using languages like C++, which were invented when dinosaurs roamed the earth (or at least, in computer science years).
AWS’s approach to pricing reflects its balance between maintaining a competitive edge and ensuring profitability. Unlike traditional CPU workloads, GPUs are resource-intensive and costly, which necessitates innovative pricing strategies. AWS’s Spot Instances, for instance, offer cost savings of up to 90% off on-demand pricing by allowing customers to bid on unused capacity—a significant consideration for investors looking at cost efficiency.
However, investors should examine the specific cost metrics, like the cost per training run, uptime, and resource utilization rates. Tracking these metrics reveals AWS's ability to scale and meet demand while maintaining low operational costs, impacting its gross profit margins.
Table 1: Cost Efficiency and Savings on Spot Instances for GPU Training Across Cloud Providers

CUDA Kitchen vs. ASIC Construction Site: The Battle of AI Development

3. Simple Inference vs Complex Inference - Know Your Clowns
All inference is not created equal. We’ve got simple inference, which is like asking a trained chatbot to respond to customer questions, and then we’ve got complex inference, which is like trying to predict shipping routes based on the dynamic movements of 50,000 cargo ships around the world.
Simple inference is what ASICs are made for—the clown car that's fully trained and just has to keep making the same rounds efficiently. Simple questions come in, answers go out. It doesn’t get much better for an ASIC; GPUs, while more capable, are overkill for this kind of workload, like bringing a chainsaw to slice bread.
But complex inference? Here, the clown car starts breaking down. If the number of tokens and parameters (think clowns) needed to navigate the model exceeds a certain threshold, it all becomes too much. In comes a GPU, built to wrangle thousands of tokens in parallel with efficiency, like an air-conditioned bus that takes on all those clowns without breaking a sweat.
AWS has invested in proprietary GPU and ML models that offer fine-tuned performance for AWS services. In comparison, Google’s TPUs (Tensor Processing Units) represent a more specialized, proprietary solution tailored for their ecosystem. Proprietary models may have short-term cost advantages but risk obsolescence as open-source solutions advance. AWS’s strategy combines proprietary and open-source compatibility, which could be a safer bet for future scalability.
Investors should look at the market share of proprietary vs. open-source models in cloud computing, as it provides insight into AWS's ability to adapt to market trends. Proprietary dominance now could mean AWS’s larger market share, but a shift to open-source models may require re-evaluation of growth potential.
Proprietary Models Dominate the AI Market, with Open-Source Gaining Momentum (2024 Projections)

Simple Inference vs. Complex Inference: The Clown Car Dilemma

4. The Networking Hurdle
There’s another element to this analogy—networking. You can have the best GPU or ASIC, but if they aren’t networked properly, they’re as useless as a clown car without a horn. Enter InfiniBand, 800 gig, 1.6G hype—terms that essentially mean connecting everything together in the most sophisticated way possible.
Imagine NVIDIA wants to be the entire circus. They want to provide the tent, the clowns, the clown car, and even the food stands. They’re building out ASICs and InfiniBand networking chips, competing with traditional network hardware companies, so their product controls everything end-to-end. That's why NVIDIA isn't just about GPUs anymore—they're moving into network territories too.
For AWS, scaling its GPU capacity involves navigating high capital expenditures and operating costs. Still, by increasing Spot Instances and reducing downtime through efficient utilization, AWS has optimized its infrastructure to maximize returns. Tracking metrics like operating margin, revenue per GPU instance, and utilization rate gives investors clarity on AWS's financial leverage.
Scalability hinges on balancing these elements effectively. With AI workloads growing exponentially, AWS’s ability to meet demand at low operating costs positions it strongly in the cloud market. Investors focused on long-term returns should consider how AWS’s infrastructure investments today will shape its financial landscape over the next decade.
Projected Financial Performance of AWS GPU Instances in 2024

The NVIDIA Circus: Pulling the Strings of AI and Compute Power

5. Market Fragmentation and the Future of AI Chips
The AI chip landscape is headed towards inevitable fragmentation. Right now, NVIDIA holds the cards, but companies like Amazon are pushing forward with custom silicon—Google has its TPUs, Amazon has Graviton, and there’s a general push for alternatives like RISC-V. ASICs will play an increasing role as the demand for power efficiency and specialized inference rises, but GPUs aren’t going anywhere anytime soon.
Imagine trying to supply a whole carnival with power. Initially, you have one massive power line—the GPU. But now, with diverse workloads and energy concerns, you're hooking up solar panels (ASICs), battery packs (TPUs), and whatever else is needed to meet the circus's needs. In this evolving world, the infrastructure adapts to specialize, and NVIDIA may not always be the only electricity supplier.
AWS also leverages compliance with regional regulations on data sovereignty to expand globally. By adhering to stringent data residency laws, AWS unlocks markets previously inaccessible to non-compliant providers. AWS’s compliance strategy could be a strong predictor of customer acquisition in emerging markets.
Investors interested in AWS's expansion into regulated markets should look at data sovereignty as a key competitive advantage, affecting customer retention rates and regulatory costs across different geographies.
Comparative Analysis of Compliance Scores and Market Penetration by Region

The Future of Fragmented Power Supply: A New Era of AI Compute

In our circus of silicon, the clown car is morphing—sometimes it's a nimble, single-purpose ASIC handling one specific ride, and sometimes it's a multi-purpose, adaptable GPU managing the whole show. NVIDIA's dominance might hold for now, but as companies learn to play the roles of architect, builder, and performer, a new era is emerging where the market is fragmented, with specialized chips taking on specific workloads.
The next act is all about finding efficiencies—shrinking models, upgrading infrastructure, and making sure those clowns all fit just right, with the least effort and energy. And if one thing's for sure, the tech circus is only getting bigger.
Clown Cars, Buses, and the Future Big Top: The AI Compute Circus

