Back to The Twenty Minute VC (20VC)

20VC: OpenAI and Anthropic Will Build Their Own Chips | NVIDIA...

The Twenty Minute VC (20VC)

Full Title

20VC: OpenAI and Anthropic Will Build Their Own Chips | NVIDIA Will Be Worth $10TRN | How to Solve the Energy Required for AI... Nuclear | Why China is Behind the US in the Race for AGI with Jonathan Ross, Groq Founder

Summary

The discussion focuses on the immense demand for AI compute, the challenges and opportunities in chip manufacturing, and the critical role of energy in the AI revolution. The episode highlights how companies are innovating in chip design and supply chains to meet this demand, while also examining the geopolitical implications of AI leadership.

Key Points

  • The demand for AI compute is insatiable, with major AI companies like OpenAI and Anthropic able to double their revenue if given more inference compute, indicating a direct correlation between compute availability and business growth.
  • NVIDIA is projected to maintain significant market value due to brand strength and customer reliance, even as the demand for specialized chips increases and companies like OpenAI and Anthropic explore building their own silicon for greater control and efficiency.
  • Building custom AI chips is incredibly difficult and capital-intensive, requiring deep expertise in design, software, and manufacturing, which is why only a few companies have succeeded in creating competitive alternatives to NVIDIA.
  • The energy required for AI compute is a major bottleneck, and solving this will necessitate a combination of efficient nuclear power, scaled renewables, and strategic placement of data centers in areas with cheap energy.
  • China is currently ahead in terms of AI action and energy production for AI, but the US and its allies have a significant opportunity to lead in the "away game" of AI by developing more efficient chips and securing energy resources, potentially creating a new global AI landscape.
  • The US leads in AI due to a massive compute advantage, which allows for more extensive model training and optimization, leading to lower inference costs and a competitive edge globally.
  • Europe's slow adoption of AI, often hindered by regulatory approaches, puts it at risk of becoming an economic outlier, and it needs to act decisively to secure energy and compute resources to remain competitive.
  • The "hardware lottery" is significant, where models are often designed for existing hardware, creating a feedback loop that benefits incumbents like NVIDIA, making it challenging for new chip designs to gain traction quickly.
  • The shift towards inference as models mature is crucial, and while GPUs currently dominate, the development of specialized inference chips like Grok's LPUs could offer a competitive advantage by addressing specific bottlenecks and increasing efficiency.
  • The labor market is expected to experience massive shifts due to AI, potentially leading to deflationary pressures, reduced need for human labor in certain sectors, and the creation of entirely new industries and job roles.
  • The debate around AI chip development centers on whether companies like OpenAI and Anthropic will successfully build their own chips or continue to rely on incumbents like NVIDIA, with the key differentiating factor being the ability to control destiny rather than just the chip itself.
  • The market is currently underestimating the sheer amount of compute needed for AI, as adding more compute directly improves product quality, expands the total addressable market (TAM), and drives economic growth in a way not seen before.
  • NVIDIA's CUDA ecosystem, while a moat for training, is less of a barrier for inference, where the focus is shifting towards more specialized hardware and software solutions, opening doors for companies like Grok.
  • The market for AI is not a zero-sum game; companies can and will coexist and innovate, with the leaders in AI compute and model development likely to expand the overall market rather than just capture existing market share.

Conclusion

The demand for AI compute is insatiable, driving significant investment and innovation in chip design and energy solutions, which will redefine industries and economies globally.

Companies that can provide scalable, efficient, and cost-effective compute, alongside robust energy infrastructure, will be the leaders in the AI revolution, shaping the future of technology and global power dynamics.

The AI era is ushering in unprecedented changes, from economic structures and labor markets to the very nature of human intelligence and our understanding of the universe.

Discussion Topics

  • How can nations ensure they have sufficient energy and compute resources to remain competitive in the AI race?
  • What are the ethical and societal implications of AI-driven labor market shifts, and how can we prepare for them?
  • As AI capabilities advance, what new business models and competitive advantages will emerge in the chip and AI software industries?

Key Terms

Compute
The processing power required to perform calculations, essential for running AI models and applications.
Inference
The process by which a trained AI model makes predictions or generates outputs based on new data.
GPUs (Graphics Processing Units)
Specialized processors originally designed for graphics rendering, now widely used for AI training and inference due to their parallel processing capabilities.
HBM (High Bandwidth Memory)
A type of DRAM memory that operates at higher speeds than standard DRAM, crucial for AI workloads that require fast data access.
Monopsony
A market situation where there is only one buyer for a particular good or service, giving that buyer significant power over sellers.
Capex (Capital Expenditure)
Funds used by a company to acquire, upgrade, and maintain physical assets, such as property, buildings, technology, or equipment.
Opex (Operational Expenditure)
Ongoing costs incurred to run and maintain a business or system, such as electricity, salaries, and maintenance.
TAM (Total Addressable Market)
The total market demand for a product or service.
AGI (Artificial General Intelligence)
A hypothetical type of AI that possesses the ability to understand or learn any intellectual task that a human being can.
Vibe Coding
A colloquial term for using AI tools to generate code without traditional programming expertise, relying on natural language prompts and AI assistance.
SRAM (Static Random-Access Memory)
A type of semiconductor memory that uses bistable latching circuitry to store each bit of data, generally faster but more expensive than DRAM.
DRAM (Dynamic Random-Access Memory)
A type of volatile semiconductor memory that stores each data bit in a separate capacitor within an integrated circuit.
LPU (Language Processing Unit)
A specialized chip designed to accelerate language-related AI tasks, potentially offering advantages over general-purpose GPUs for inference.
Moat
In a business context, a sustainable competitive advantage that protects a company's market share and profitability from competitors.
CUDA
A parallel computing platform and application programming interface model created by NVIDIA, enabling software to use NVIDIA graphics processing units for general-purpose processing.

Timeline

00:00:03

The countries that control compute will control AI, and you cannot have compute without energy, highlighting the foundational importance of these resources for AI dominance.

00:00:17

Jonathan Ross predicts NVIDIA could be worth $10 trillion in five years due to insatiable demand for compute, emphasizing the scale of the current AI infrastructure race.

00:00:19

The demand for compute is insatiable, with OpenAI and Anthropic stating they could double revenue with just double the inference compute, demonstrating the direct link between compute and AI service scalability.

00:05:23

Ross analyzes the current AI market not as a bubble, but as a high-investment phase akin to early oil drilling, where smart players are making significant returns despite the lumpiness of successes.

00:06:05

The AI market spend is characterized as "lumpy," with a few companies driving most of the revenue, mirroring the early days of oil exploration where instinct and a few big finds dictated success.

00:07:40

Ross questions investors about AI's future impact, finding no one certain it won't displace their jobs, which he uses to illustrate why hyperscalers are investing heavily to maintain leadership.

00:09:05

Ross shares an anecdote about using AI to deliver a custom feature in four hours with no human code writing, showcasing the value proposition of AI in rapid development and problem-solving.

00:10:11

The discussion explores whether major tech companies will move into chip manufacturing, with Ross emphasizing the immense difficulty and risk involved, citing failed efforts like Dojo.

00:10:58

The conversation touches on the concept of "Mag 7" companies needing to invest in chips to stay relevant, but Ross argues that successfully replicating NVIDIA's chip performance is a near-impossible task.

00:11:09

Ross clarifies that the investment in chips by AI companies is not an "infinite money loop" but a necessary spend on productive infrastructure that injects capital into the ecosystem.

00:12:01

The value increase in NVIDIA's stock is attributed to the belief that its AI compute will continue to drive revenue, fueled by the insatiable demand for more computational power.

00:12:13

The demand for compute is insatiable, with Ross suggesting that doubling compute for OpenAI or Anthropic could double their revenue within a month due to current rate limits and customer demand.

00:12:33

The importance of speed in AI applications is highlighted, drawing parallels to CPG margins and conversion rates on web pages, arguing that quick responses are crucial for user engagement and brand affinity.

00:13:59

Ross disputes the idea that AI latency is acceptable, stating that speed directly correlates to user engagement and competitive advantage, a lesson learned from building early internet companies.

00:14:40

The possibility of OpenAI or Anthropic verticalizing into chip manufacturing is discussed, with Ross acknowledging their ability but emphasizing the complexity of software and keeping pace with evolving technology.

00:16:09

NVIDIA's near-monopoly on HBM supply creates a bottleneck, incentivizing hyperscalers to explore building their own chips to secure necessary capacity and control their destiny.

00:17:17

The dominance of NVIDIA's GPUs is explained by how small performance edges translate into massive system value, making the chip cost negligible compared to the overall system deployment cost.

00:18:03

The HBM market presents a challenge due to NVIDIA's buying power, but a diversified supply chain for HBM manufacturers could create opportunities for competitors.

00:18:37

The capital cost and conservative nature of memory suppliers create a supply constraint for GPUs, even for a company with NVIDIA's financial power.

00:19:23

The cost of building data centers versus buying systems is debated, with data centers representing a larger, amortized cost over a longer period, but still a significant investment.

00:19:59

A faster payback period for chip investments is advocated for, suggesting that a three to five-year amortization cycle for chips might be too long given rapid technological advancements.

00:20:24

The value of a chip has two phases: initial deployment and ongoing operation, with the latter being more critical for profitability as long as operational costs are met.

00:21:38

The potential for excess compute supply due to rapid chip obsolescence is a concern, but long-term contracts and the cost of breaking them complicate this scenario.

00:22:00

The emphasis is on minimizing payback periods and operational costs to manage the lifecycle of chips effectively and avoid the risk of having underperforming, costly hardware.

00:22:44

Grok's value proposition is providing more compute capacity with a faster supply chain, a critical advantage when hyperscalers cannot meet demand.

00:23:05

The ability to provide more compute capacity is the primary value proposition, as it directly enables companies to serve more customers and therefore increase their revenue and lock-in.

00:23:55

Grok's supply chain advantage, with a six-month delivery timeframe for LPUs compared to NVIDIA's two-year GPU commitment, is a key differentiator that attracts hyperscalers.

00:24:24

The "hardware lottery" concept suggests that models are designed for existing hardware, creating a loop that benefits incumbents and makes it hard for new entrants to compete without a faster iteration cycle.

00:25:01

For new chip market entrants, a faster development and deployment cycle is essential because customers will not design their models for chips that won't be available for two years.

00:25:20

The current market prediction is that the US has a two-to-three-year advantage in the "away game" of AI, emphasizing the importance of speed and allied collaboration.

00:25:41

Despite past underestimations of compute needs, companies are now trying to overbuild capacity, but still often fall short due to the rapid growth of AI demand.

00:26:00

AI's impact on product quality differs from SaaS; AI allows for continuous improvement on each query by spending more compute, meaning more compute directly leads to better products.

00:27:02

The assumption that efficiency is the sole focus for future AI models like GPT-5 is questioned, as cost-sensitive markets require different optimization strategies than performance-driven ones.

00:27:16

The current focus on efficiency for models like GPT-5 is driven by the need to penetrate cost-sensitive markets like India, where low-cost AI access is paramount.

00:27:44

The misconception that Chinese AI models are cheaper to run is debunked; US-based OSS models are significantly more cost-effective for inference despite potentially higher training costs.

00:28:39

The cost to train an AI model is amortized over its inference usage, making cheaper inference a key competitive advantage, especially when scaling to a global audience.

00:29:29

The US holds a substantial compute advantage, enabling more intensive model training that lowers inference costs, giving it a strategic edge in the AI race.

00:29:44

The subsidy potential of the Chinese government for AI compute is acknowledged, but the US and its allies can compete by having more energy-efficient chips, crucial for countries with limited power capacity.

00:30:02

The "away game" for AI leadership is won by having better chips, which provide an advantage in regions where building massive energy infrastructure is not feasible, solidifying the US's position for the next few years.

00:30:42

The idea of open-sourcing models is discussed, with the conclusion that while brand strength can initially drive adoption, prompt compatibility and cost efficiency are key for wider uptake.

00:31:40

Prompt reusability is a significant factor in adopting open-source models, allowing users to transition to premium models without losing their existing work.

00:32:10

The intense energy requirements for AI compute are addressed, with nuclear power being efficient but renewables also playing a crucial role, and the key strategy being to locate compute where energy is cheap.

00:32:44

The US's fear of "mistakes of omission" drives its aggressive AI investment, contrasting with Europe's caution and legislative focus, which could lead to being left behind.

00:33:42

Europe could unlock significant energy potential through renewables like wind in Norway and nuclear power, but fear and a lack of decisive action are hindering progress.

00:34:39

Japan's rapid progress in building a two-nanometer fab and allocating substantial funds to AI shows a determination to compete, serving as a model for Europe.

00:35:18

The speed of building renewable energy infrastructure like wind turbines is a challenge, suggesting that governments or hyperscalers should fund these projects to meet AI's energy demands.

00:35:39

Collaboration with regions like Saudi Arabia, which are investing heavily in data centers and energy, could offer a solution for Europe's energy needs for AI.

00:36:36

The high cost of permitting in the US and Europe for infrastructure like nuclear power plants is a significant hurdle, potentially slowing down AI energy deployment.

00:36:39

The core takeaway is that controlling compute means controlling AI, and compute requires energy, making energy infrastructure a critical factor in global AI leadership.

00:36:48

Europe's current lag in the AI race is significant, but acting now could allow it to catch up, especially by leveraging allied strengths and focusing on energy solutions.

00:38:04

Model sovereignty alone is insufficient to win in AI; abundant compute is the ultimate determinant of AI performance, as even a superior model will underperform if it lacks sufficient computational power.

00:38:13

The question of whether European models like Mistral can compete hinges on the availability of sufficient compute, not just on their origin or ownership.

00:38:59

Grok, despite being a compute provider, aims to avoid competing with its customers by not developing its own AI models, creating a safer ecosystem for developers.

00:39:07

While companies like CoreWeave offer on-demand GPUs, their finite allocation limits their ability to fully address the massive demand for compute.

00:39:20

As AI models mature, inference becomes more critical, and specialized hardware like Grok's LPUs may offer advantages over GPUs for this specific task.

00:39:33

NVIDIA is expected to continue selling GPUs due to their established brand and the virtuous cycle between training and inference, where more inference requires more training.

00:40:01

The inference market is evolving rapidly, with the unexpected rise of language-based AI making it accessible to a wider audience and driving faster adoption than initially predicted.

00:40:33

The massive adoption of AI, with 10% of the world's population being weekly active users of services like ChatGPT, is astounding, but compute remains a bottleneck for quality and broader language support.

00:41:11

The AI ecosystem relies on data, algorithms, and compute; improving any one of these can enhance AI, but compute is the most accessible and consistently improving factor.

00:41:47

The demand for compute is underestimated because adding more compute directly improves AI quality and user experience, leading to increased demand and economic growth without traditional bottlenecks.

00:42:49

The economy's strength is predicated on a shift from labor spend to AI, with AI expected to create labor shortages rather than widespread unemployment due to increased productivity and new industries.

00:43:13

AI is predicted to cause massive deflationary pressure, leading to reduced costs for goods and services, and consequently, people working less and retiring earlier.

00:44:09

AI will create new jobs and industries that are currently unimaginable, similar to how agriculture's automation freed up labor for new economic pursuits.

00:45:09

The US can benefit from AI advancement due to policies that facilitate energy production and infrastructure development, which is crucial for scaling AI compute.

00:45:48

The concept of "vibe coding" is examined as an enduring skill, analogous to reading and writing becoming universal, enabling non-programmers to create functional tools.

00:47:07

Margins in the compute business are important for stability and weathering market volatility, but the primary focus should be on customer value and solving unmet needs.

00:48:10

The strategy for margins is to keep them as low as possible for business stability while retaining the ability to increase prices if needed, leveraging volume to drive profit.

00:49:33

Jevon's paradox applies to compute: increased availability leads to increased sales as costs decrease, creating a virtuous cycle of innovation and adoption.

00:50:17

Businesses should focus on customer needs and differentiation rather than solely on the bottom line, as solving unique problems leads to customer loyalty and revenue growth.

00:51:11

AI's impact is expanding the Total Addressable Market (TAM) by making products easier to use and enabling new functionalities, leading to overall revenue increases even if per-unit costs decrease.

00:51:48

The value of AI lies in its ability to deliver tangible benefits, as evidenced by private equity firms seeking cheap AI compute to improve their businesses' bottom lines.

00:53:04

The potential for AI to add more labor to the economy by enhancing compute and AI capabilities is unprecedented, suggesting a significant shift in economic paradigms.

00:53:21

A short-term speed bump in AI development could have a significant multiplier effect on the economy due to the concentration of value in major AI companies.

00:54:03

Economic downturns can be challenging but also create opportunities for robust businesses to emerge, highlighting the cyclical nature of markets.

00:54:44

The economy's unpredictability stems from human behavior and the impact of predictions on those predictions, making forecasting difficult.

00:54:54

A major problem in AI is the dispersion of talent, where engineers can raise vast sums to start new companies, hindering the growth of existing startups by fragmenting expertise.

00:56:18

The war for tech talent is more aggressive than ever, but unlike sports, there are no salary caps or limitations on the number of startups, allowing for unlimited team creation.

00:56:57

Google is praised for its turnaround and AI DNA, with Gemini showing success in adoption, though consumer product integration needs further refinement.

00:58:15

OpenAI has established a significant distribution advantage, making it challenging for incumbents like Google to catch up, despite their efforts to iterate on AI products.

00:58:34

The enduring value of OpenAI and Google in the AI landscape is likely, with Anthropic potentially carving out a niche in areas like coding.

00:59:03

Engineers are shifting between AI tools like Sourcegraph, Anthropic, and Codex based on what is currently the "best tool," indicating low switching costs and a dynamic competitive environment.

00:59:29

The rapid switching of AI tools by cutting-edge engineers raises questions about the enduring value of specific platforms if switching costs are low.

00:59:59

Both OpenAI and Anthropic are considered undervalued, as their R&D efforts are expanding the market itself, not just competing within a finite space.

01:00:13

The bull case for AI labs is that they will achieve similar market value to current tech leaders and potentially expand the "Mag 7" into "Mag 9" or more.

01:00:53

Successful tech companies naturally tend to move up the value stack and subsume parts of their customers' businesses, a tendency that AI labs may also follow.

01:01:18

Grok's decision not to build its own AI models is a strategic choice to avoid competing with its customers, ensuring a safe infrastructure for developers, though this could be a potential risk.

01:01:50

Grok's recent $750 million raise at a nearly $7 billion valuation was driven by investor confidence in their ability to execute and the positive margins on their hardware sales.

01:02:15

Hardware companies like Grok have an advantage with positive margins on their products, unlike software companies whose margins can vary depending on the model and operational costs.

01:03:29

Grok aims for low margins for business stability while retaining the ability to increase prices if market demand necessitates it, leveraging volume for profit.

01:04:11

In five years, NVIDIA is predicted to retain significant revenue share but a smaller portion of chip sales as custom chip development increases and customers gain more power to choose alternatives.

01:04:37

NVIDIA's brand value allows them to charge premium prices, but this can make them less competitive as customers prioritize business success over brand loyalty when making purchasing decisions.

01:05:26

Customers with significant spending power, like those in the top 35-36 clients in the AI market, will increasingly make decisions based on business success rather than brand name.

01:05:45

The prediction is that NVIDIA will be worth $10 trillion in five years, driven by its established position and the ongoing demand for AI compute.

01:05:53

Grok's potential to reach $10 trillion valuation is linked to its ability to produce near-unlimited compute without the supply chain constraints faced by others.

01:06:11

Grok's competitive advantage lies in its ability to produce compute at scale and cost-effectively, addressing the market's primary need beyond just brand reputation.

01:06:26

Grok's innovation in chip design, particularly with SRAM integration, allows for more efficient multi-user capabilities on a single hardware unit, optimizing system-level performance.

01:07:11

The cost of SRAM per bit is inherently higher than DRAM due to its architecture, but Grok's system-level approach and massive chip deployment reduce overall cost per token.

01:08:04

Grok optimizes compute at a world-scale distribution level, load balancing across multiple data centers globally to provide tailored solutions based on geographical needs and model performance.

01:08:46

Grok has the capacity to significantly increase its supply chain build-out rate, having raised substantial funds and being oversubscribed, indicating a readiness to meet extreme demand.

01:09:59

Grok's cost per token, especially with its speed advantage, is highly advantageous, allowing them to charge less than the market while still building a stable business through volume.

01:10:33

The focus for Grok is purely on execution and satisfying compute demand, not on immediate profitability or public offerings, reflecting a commitment to core business growth.

01:10:48

The biggest misconception about NVIDIA is that its CUDA software ecosystem is an insurmountable moat for inference, as specialized hardware and developer adoption of alternative platforms are challenging this.

01:11:09

If founding Grok today, Ross would not pursue chip development due to the long lead times, advocating instead for faster-moving areas of the AI ecosystem.

01:11:18

The temporal moat of chip development, requiring years from design to production, is a significant barrier to entry, making it difficult for new chip providers to compete with established players.

01:13:04

Larry Ellison's meteoric rise with Oracle is attributed to brilliant business decisions and a willingness to move aggressively and embrace risk, a strategy Ross advocates for in the AI space.

01:13:13

When everyone else is fearful about AI, it's an opportunity for smart investors to be greedy, and vice versa, emphasizing contrarian investment strategies.

01:13:43

Investors should be greedy where moats exist, and in the early stages of AI investment, focus on identifying companies that have the potential to build strong, defensible moats.

01:14:14

Ross's approach has shifted from preserving optionality to intense focus, believing that saying "yes" to fewer, more strategic opportunities leads to better business outcomes.

01:14:41

Elon Musk's Grok and XAI may succeed by differentiating their offerings, as AI markets are expected to diverge, with different companies excelling in specific niches.

01:15:01

AI markets will eventually diverge, with companies like Anthropic focusing on coding and others potentially specializing in different AI applications.

01:15:43

Companies that do not differentiate themselves risk dying, as seen in the tech landscape where clear business focuses are essential for long-term survival.

01:15:53

Microsoft's AI strategy, leveraging its OpenAI partnership and ability to deploy compute, positions it well, while Amazon's lack of AI DNA is a concern despite its compute capacity.

01:16:47

Amazon's lack of AI DNA is identified as a weakness, contrasting with Google and Meta, which have historically strong AI foundations, and Microsoft's strategic acquisition of AI capabilities through OpenAI.

01:16:52

Ross is excited about the future of AI, viewing LLMs as a "telescope of the mind" that will expand human intelligence and reveal new truths about the universe.

01:17:45

The biggest problem in AI is the difficulty in attracting and retaining top engineering talent due to excessive funding for new startups, which fragments expertise.

01:17:59

The rapid adoption of AI by 10% of the world's population is astonishing, but compute limitations and language support are holding back even wider use.

01:21:03

AngelList is recommended as a modern platform for venture funds, offering an all-in-one software solution and a dedicated service team to manage fund operations efficiently.

01:21:20

.tech domains are suggested for tech founders to secure the ideal startup name without compromise, signaling innovation and a focus on technology.

Episode Details

Podcast
The Twenty Minute VC (20VC)
Episode
20VC: OpenAI and Anthropic Will Build Their Own Chips | NVIDIA Will Be Worth $10TRN | How to Solve the Energy Required for AI... Nuclear | Why China is Behind the US in the Race for AGI with Jonathan Ross, Groq Founder
Published
September 29, 2025