20VC: Cerebras CEO on Why Raise $1BN and Delay the IPO | NVIDIA...
The Twenty Minute VC (20VC)Full Title
20VC: Cerebras CEO on Why Raise $1BN and Delay the IPO | NVIDIA Showing Signs They Are Worried About Growth | Concentration of Value in Mag7: Will the AI Train Come to a Halt | Can the US Supply the Energy for AI with Andrew Feldman
Summary
Cerebras CEO Andrew Feldman discusses the company's $1.1 billion Series G funding round, the rapid evolution of the AI market, and the challenges and opportunities in scaling AI infrastructure.
The conversation delves into NVIDIA's market position, the complexities of chip depreciation, the energy demands of AI, and the concentration of value in large tech companies.
Key Points
- Cerebras secured a $1.1 billion Series G at an $8.1 billion valuation, underscoring investor confidence and providing capital to scale manufacturing, expand data centers, and pursue further innovation in AI hardware.
- The current AI market is characterized by immense demand and uncertainty about future needs, leading companies to make large "options on the future" with significant capital investments.
- Traditional planning cadences are insufficient for the rapid pace of AI development; instead, frequent planning and option-based strategies are necessary to navigate the uncertainty.
- NVIDIA is showing signs of strategic maneuvering, possibly indicating concern about future growth, by using its balance sheet and engaging in predatory pre-announcements to maintain market dominance.
- Chip depreciation is not solely about the age of the hardware but rather how much faster future generations are, making older, even fully depreciated, chips obsolete due to superior performance and efficiency.
- The critical bottleneck for AI performance is not just raw compute power but the entire system, including memory bandwidth and data transfer, which can become limiting factors.
- Cerebras' wafer-scale approach addresses the limitation of SRAM's low capacity on traditional chips by utilizing a much larger silicon area to store vast amounts of fast SRAM, overcoming previous impossibilities in chip manufacturing.
- While NVIDIA's GPUs are strong for training, Cerebras claims superiority in both training and inference, with inference being easier to demonstrate due to simpler software integration compared to the complex porting required for training models.
- The exponential growth of AI inference is driven by increasing user adoption, higher frequency of use, and more complex AI tasks, all multiplying to create mind-numbing demand.
- Historical parallels with the adoption of electricity and computers suggest that AI will lead to significant productivity gains only when the entire economy and workflows are reorganized around it, not just used as a replacement for existing tools.
- The US has sufficient energy resources, but they are geographically dispersed, and the primary challenge is infrastructure and distribution to meet AI's localized demand.
- The concentration of value in a few large tech companies (Mag7) presents market risk, as a slowdown in AI could lead to significant market downturns due to a lack of true diversification.
- Bottlenecks in the AI ecosystem include a shortage of AI expertise and talent, the inability of foundries like TSMC to build fabs quickly enough, and a lack of sufficient data center capacity.
- The high compensation for top AI talent is justified by the immense value they can generate for companies, far exceeding their salaries.
- Building AI infrastructure, particularly data centers, is complex and costly, with many potential pitfalls for those who underestimate the requirements for power access, permitting, construction, and tenant management.
- The most successful AI companies to date (OpenAI, Anthropic) are not vertically integrated, suggesting that a fully horizontal model from chip to software is not the only path to success.
- Building chips is an extremely difficult and capital-intensive endeavor, often requiring deep expertise and long-term vision, which challenges many software-focused companies.
- The AI market is unlikely to become a monopoly, with value likely to accrue to multiple specialized companies, similar to the historical distribution of market share in the broader silicon industry.
- Margins are becoming increasingly important for companies, especially as they approach public markets, and high gross margins like NVIDIA's can create long-term incentives for competitors like cloud providers to develop their own solutions.
- Sovereignty can be a compelling driver for AI adoption in regions like Europe, especially when combined with advanced hardware that offers superior performance.
- The US-China AI race is framed as potentially detrimental, drawing parallels to the costly arms race of the past, and peaceful engagement is suggested as a more beneficial path.
- The decentralized nature of US governance, with local regulations interfering with large projects, presents a challenge for strategic infrastructure development compared to China's centralized planning.
- Attracting and retaining global talent is crucial for the US AI effort, and historical pathways like J1 and H1 visas, along with green cards, have been vital.
- AI's impact on education and the nature of entry-level jobs will be profound, shifting from rote tasks to higher-level problem-solving and innovation as AI handles more routine work.
Conclusion
The AI revolution is accelerating at an unprecedented pace, requiring companies to be agile and strategic in their investments and planning.
Significant bottlenecks exist across the AI ecosystem, from talent and manufacturing to data center capacity and energy infrastructure, demanding focused solutions.
The future of AI development and its impact on society and the economy will depend on how effectively we reorganize around this technology and ensure its development yields substantial, tangible benefits.
Discussion Topics
- How can companies navigate the immense and unpredictable demand in the AI market while mitigating the risks of over- or under-investment?
- What are the most significant overlooked bottlenecks in the AI ecosystem, and what innovative solutions are needed to address them?
- Given the rapid advancements in AI, how should educational systems and early-career development adapt to prepare the future workforce for evolving job markets?
Key Terms
- SRAM
- Static Random-Access Memory, a type of semiconductor memory that uses bistable latching circuitry to store each bit of data. It is faster and more expensive than DRAM and is used for cache memory and in high-performance computing.
- HBM
- High Bandwidth Memory, a type of DRAM integrated circuit. It is used in systems that require higher memory bandwidth than standard DDR memory can provide, common in GPUs and AI accelerators.
- Wafer Scale
- A manufacturing process where an entire silicon wafer is used to create a single large integrated circuit, rather than cutting the wafer into multiple smaller chips. This allows for significantly larger and more powerful chips.
- FLOPS
- Floating-point Operations Per Second, a measure of computer performance, particularly relevant in scientific and high-performance computing tasks.
- GPU
- Graphics Processing Unit, a specialized electronic circuit designed to rapidly manipulate and alter memory to accelerate the creation of images in a frame buffer intended for output to a display device. GPUs are also widely used for parallel processing in AI and machine learning.
- IPO
- Initial Public Offering, the process by which a private company becomes public by selling its shares to the public for the first time.
- Series G Funding
- A later stage of venture capital funding for a company that has already undergone several earlier rounds (Series A, B, C, etc.), typically indicating a mature company seeking significant capital for expansion or preparing for an IPO.
- Mag7
- Refers to the seven largest technology companies in the S&P 500 index, often cited for their significant market capitalization and influence on the stock market.
- EDA Toolmaker
- Electronic Design Automation toolmaker, companies that provide software used to design and verify integrated circuits (chips).
- PA Semi
- A fabless semiconductor company acquired by Apple in 2008, contributing to Apple's in-house chip design capabilities.
- Anapurna Labs
- A semiconductor design company acquired by Amazon in 2015, instrumental in developing Amazon's custom chips.
- TSMC
- Taiwan Semiconductor Manufacturing Company, the world's largest contract chip manufacturer.
- SOC 2
- System and Organization Controls 2, a framework that specifies how organizations should manage customer data, focusing on security, availability, processing integrity, confidentiality, and privacy.
- ISO 27001
- An international standard for information security management systems.
Timeline
Andrew Feldman discusses the immense, unpredictable demand in the AI market.
Feldman clarifies that the US has ample power, but it's in the wrong locations.
Feldman explains the strategic importance and investor confidence generated by Cerebras' $1.1 billion Series G funding round.
Feldman describes the current AI market dynamics characterized by massive claims and a lack of accountability for future projections.
Feldman addresses the challenge of planning in a rapidly evolving AI environment, emphasizing adaptive strategies.
Feldman discusses the sustainability of AI market growth, acknowledging skepticism but highlighting the potential for economic transformation.
Feldman touches upon NVIDIA's rapid growth and the potential for market shifts.
Feldman suggests NVIDIA may be exhibiting signs of concern about growth by leveraging its balance sheet.
Feldman details strategies of large companies using their strengths, specifically mentioning NVIDIA's pre-announcement tactics.
Feldman discusses the complexities of chip depreciation and how it relates to the pace of technological advancement.
Feldman questions the current stage of performance improvement in AI chips, asking if we are at incremental gains or early stages.
Feldman explains that true performance gains are often less dramatic than marketing suggests, with memory bandwidth being a key constraint.
Feldman discusses the limitations of SRAM and HBM for large-scale AI, highlighting Cerebras' wafer-scale approach.
Feldman explains how Cerebras' wafer-scale chip design overcomes traditional limitations by utilizing extensive fast SRAM.
Feldman elaborates on the difficulty of building large chips, noting past failures and the unique challenges Cerebras overcame.
Feldman clarifies that Cerebras is faster in both training and inference, but inference is easier to market due to simpler adoption.
Feldman discusses how the AI inference market has developed unexpectedly, driven by exponential growth in user numbers, frequency, and compute needs.
Feldman draws parallels between AI adoption and the historical diffusion of electricity and computers, emphasizing the need for systemic reorganization for significant productivity gains.
Feldman addresses the immense energy requirements of AI and the feasibility of meeting them in the US.
Feldman discusses the societal obligation to ensure AI delivers significant value to justify its resource consumption.
Feldman offers an assessment of the Trump administration's impact on the US AI effort.
Feldman debates the necessity of nuclear power for AI energy demands, discussing alternatives and cost-effectiveness.
Feldman expresses concern about the AI community's rapid, perhaps helter-skelter, approach to resource consumption without guaranteed extraordinary outcomes.
Feldman discusses the risk associated with the concentration of value in Mag7 companies and its implications for market stability.
Feldman explains how underestimating risk in financial markets, particularly sector-specific risks within concentrated indices, can lead to surprises.
Feldman comments on NVIDIA's market valuation and its current position.
Feldman identifies expertise and manufacturing capacity as key bottlenecks in the AI industry.
Feldman addresses the "wolf of talent" and the unprecedented compensation for highly skilled AI professionals.
Feldman discusses the limitations in TSMC's ability to build fabs fast enough, impacting chip supply and cost.
Feldman notes a shortage of data center capacity despite significant investment talk.
Feldman explains the difficulties and ways to lose money in building data centers.
Feldman discusses the unclear necessity of full vertical integration in AI companies.
Feldman weighs in on whether OpenAI and Anthropic will build their own chips to reduce reliance on NVIDIA.
Feldman highlights the fundamental differences in mindset required for chip design versus agile software development.
Feldman reflects on how leaders missed the dominant compute market of the early 21st century.
Feldman considers the future market landscape for silicon.
Feldman discusses the importance of margins for Cerebras as it prepares for a public offering.
Feldman comments on NVIDIA's extraordinary margins and how they impact competitors.
Feldman discusses the role of sovereignty as a driver for AI development in Europe.
Feldman shares his perspective on the US-China AI race.
Feldman outlines areas where the US needs to improve its AI strategy, particularly regarding power infrastructure and university resources.
Feldman discusses the importance of attracting and retaining global talent through visa programs.
Feldman reiterates the challenges with US power infrastructure and permitting processes at local levels.
Feldman mentions the need to seriously consider and support university research in AI.
Feldman participates in a quick-fire round of questions.
Feldman expresses belief in future peace in the Middle East, citing visits and economic incentives.
Feldman explains the concentration of Cerebras' revenues in the UAE due to significant demand and consumption.
Feldman admits to making mistakes in resource planning but emphasizes the bold bets that have been successful.
Feldman recounts the biggest bet at Cerebras: the successful development of wafer-scale chips after numerous failures.
Feldman advises against young CEOs entering the silicon industry without prior experience, contrasting it with markets where founder-customer proximity is an advantage.
Feldman highlights under-invested but critical areas like data cleaning and pipelines as crucial for AI project success.
Feldman discusses the curious and evolving data provision market.
Feldman predicts AI's diffusion into the economy will be gradual, not immediately causing mass labor shortages.
Feldman uses AlphaFold as an example of a breakthrough with future but not immediate widespread impact.
Feldman foresees AI significantly changing education and the nature of entry-level jobs.
Feldman states he thrives on competing against overwhelming odds, drawing pride from facing Goliath daily.
Feldman praises the host's writing for capturing the entrepreneurial spirit, comparing it to Ben Horowitz's "The Hard Thing About Hard Things."
Feldman challenges the notion of achieving greatness through part-time work and work-life balance, emphasizing the all-consuming nature of building something new.
Episode Details
- Podcast
- The Twenty Minute VC (20VC)
- Episode
- 20VC: Cerebras CEO on Why Raise $1BN and Delay the IPO | NVIDIA Showing Signs They Are Worried About Growth | Concentration of Value in Mag7: Will the AI Train Come to a Halt | Can the US Supply the Energy for AI with Andrew Feldman
- Official Link
- https://www.thetwentyminutevc.com/
- Published
- October 6, 2025