SambaRack™
Air-cooled hardware built for cloud-scale agentic AI solutions
Introducing SambaRack SN50
Built from our fifth-generation chips, SambaRack SN50 is purpose-built for agentic inference. It combines 16 SN50 RDU chips in one rack and can run inference workloads scaled out over a dozen racks.
The advantages of SambaRack
Efficiency at the core
At the heart of SambaRack is the reconfigurable dataflow unit (RDU), which leverages a dataflow architecture. This approach minimizes data movement to achieve better performance and energy efficiency.
Learn more -->Cloud-scale token factories
SambaRack SN50 delivers 5X more compute per accelerator and 4X more network bandwidth than the previous generation.
With the ability to scale up to 256 accelerators over a multi‑terabyte‑per‑second interconnect, the time‑to‑first‑token can be reduced, allowing for support of larger batch sizes.
As a result, organizations can deploy models with higher throughput and responsiveness. These inference clusters are managed with our SambaStack full-stack solution.
Learn more -->
Fast inference on the largest models
SambaRack delivers low latency with high throughput, resulting in better tokenomics.
Unlike SRAM-based systems that require dozens of racks to run models, a single SambaRack SN50 can run models of up to 10 trillion parameters and with a context length of up to 10 trillion tokens.
This performance advantage makes SambaRack the ideal choice for complex agentic and coding workloads.
Learn more -->Related resources

SambaNova Launches First Turnkey AI Inference Solution for Data Centers, Deployable in 90 Days
Seamless turnkey deployments
A turnkey data center solution, SambaRack integrates hardware, networking, and software into a single, self-contained system. With racks readily available, systems can be installed in as few as 90 days — getting your AI inference cloud up and running quickly. We have helped customers around the globe set up their inference services quickly.
Learn more -->
Secure, fast, flexible
Industries worldwide tap into the SambaRack performance advantage.
Developers & Enterprise
Powers leading businesses with private, plug-and-play, and fast AI.
Learn more →Government & Public Sector
Gain secure, flexible, and fast AI inference for all nations.
Learn more →FAQs
SambaRack is a turnkey AI rack system designed to deploy and run large AI models efficiently in data centers. It integrates hardware, networking, and software into a single self-contained system built around SambaNova RDU chips to deliver fast inference performance.
SambaRack is designed for AI inference workloads that require high throughput and low latency when running large models in production. It is used to deploy and serve inference on AI models at scale across enterprise, developer, and data-center environments.
SambaRack is a rack-level integrated system built specifically for inference using RDU processors and its Dataflow architecture, whereas GPU clusters typically assemble general-purpose accelerators across multiple servers. This purpose-built design of SambaRack focuses on improving efficiency and performance for inference on AI models. SambaRack can co-exist alongside GPUs in a data center for different workloads.
Yes. SambaRack combines multiple RDU chips in one rack, typically 16 SN50 RDUs, to provide the compute and memory capacity required to run very large AI models efficiently within a single system.
Yes. SambaRack is designed to integrate into existing air-cooled data center infrastructure with minimal modification, enabling organizations to deploy AI inference on-premises as well as in hosted environments.
SambaRack scales by combining multiple RDU chips within each rack and connecting multiple racks together into larger inference clusters. Large deployments can span multiple racks to deliver higher throughput and support large-scale AI inference services.


