What NVIDIA’s bet on OpenClaw means for the future of AI and your token budget

Azeem Azhar's Exponential View Podcast Recap

Published:

What Happened

Nvidia's latest GTC Jamboree showcased AI inference as a pivotal area for future development. The company's valuation stands at $4 trillion, affirming its dominance in the AI accelerator chip market. Nvidia has $1 trillion in committed AI chip orders, a significant leap from $500 billion the previous year.

OpenClaw, an open-source platform enabling AI agents reminiscent of 1980s science fiction, is gaining rapid popularity. Within 45 days, it surged from 5,000 GitHub stars to becoming a major focal point. Exponential View has increased its compute capacity by 50% to accommodate these agents, highlighting the growing demand for AI inference.

Inference is now the primary driver of AI compute demand, shifting away from training. Nvidia's acquisition of Groq for $20 billion aims to bolster their inference capabilities. The planned release of new chips with Groq technology promises a 35-fold improvement in throughput per megawatt.

Exponential View's AI token usage has skyrocketed, jumping from 100 million to 870 million tokens daily in a matter of weeks. This increase is attributed to the deployment of various AI models for simulations, coding, documentation, and security audits. A model registry ensures these AI systems remain up-to-date.

AI agents like R. Mini Arnold and R. Veblen currently trigger inference workloads at Exponential View. Although these processes require manual approval, there is a noticeable trend towards allowing agents to autonomously initiate large token workloads. Governance over these agents and their compute usage is deemed necessary to maintain efficiency.

Token budgets have become a critical consideration, especially in smaller organizations. Jensen suggests allocating half of an engineer's salary to their token budget, underscoring its importance. In many companies, IT departments manage token budgets, but this may not be the ideal approach.

The inference economy is experiencing a significant shift, having grown a million-fold over the past two years. Nvidia and its partners anticipate continued growth in this sector. The transition from a training economy to an inference economy marks a new era in AI development.

Key Insights

View all Azeem Azhar's Exponential View recaps