
The new diplomacy of the AI chip wars
Unlocking AI’s Potential: SambaNova’s Vision for Efficient Inference and Data Sovereignty
(This article was generated with AI and it’s based on a AI-generated transcription of a real talk on stage. While we strive for accuracy, we encourage readers to verify important information.)
SambaNova Systems, led by Co-founder and CEO Mr. Rodrigo Liang, is a pivotal force in the computing revolution. The company builds data center infrastructure for high-performance, low-latency inference, utilizing custom chips and servers. Their unique architecture, developed from Stanford research, optimizes specifically for neural network inference.
Over seven years, SambaNova has achieved six tape-outs, deploying products in 15 countries. Their technology efficiently handles models from 8 billion to multi-trillion parameters within a single 10-kilowatt rack. Mr. Liang emphasized efficient scaling, not just raw performance.
The AI landscape is rapidly shifting from a primary focus on training models to a dominance of inference. High costs of training proprietary models drive many organizations to adapt and fine-tune open-source or existing frontier models. This makes adaptation a more viable and economical strategy.
Inference is now the prevailing function, encompassing agentic AI and synthetic data generation. Mr. Liang noted that not all tasks require large frontier models. Agentic AI, which interconnects specialized models for tasks like voice, video, or code generation, can often utilize smaller, fine-tuned open-source models.
Enterprises typically start with frontier models, but quickly face escalating token costs. This prompts them to explore cheaper, smaller models. Ultimately, businesses seek differentiation by training unique models with their proprietary data for competitive advantage.
Mr. Liang described the demand for AI infrastructure as “effectively unlimited,” fueled by the build-out of hyperscale, frontier, neo, and sovereign clouds. However, traditional enterprises have yet to fully invest in their own AI infrastructure. Industry constraints include the scarcity of chips and suitable data center space.
SambaNova addresses these constraints by offering an efficient alternative for inference workloads. They systematically optimize popular open-source models to run up to ten times faster and more efficiently on their platforms. This approach frees up Nvidia racks for other tasks, saving on capital and operational expenditures.
Differentiation is crucial; SambaNova’s 5x-10x speed improvements for specific workloads allow businesses to enhance services and improve margins. Data sovereignty is also a critical driver, with countries investing in local infrastructure to protect sensitive data.
Mr. Liang foresees a heterogeneous future for the industry, with diverse infrastructure, model use cases, and data sets. Supply chain issues, particularly for DRAMs and chips, remain a bottleneck, leading to aggressive pre-ordering as AI access becomes a crucial competitive factor.

