AI Chips for Blazing Fast Inference: Meet SambaNova's RDU
RDUs are the only solution that run the largest AI models on a single system with blazing fast perfomance
Delivering fast & energy efficient Al inference
Our fourth-generation RDU, the SN40L, is designed to deliver extraordinary performance. Uniquely capable of powering the most demanding workloads, the SN40L does it all with the highest efficiency rate.

Seamlessly achieve high performance
From chips to racks
The combination of 16 SN40L RDUs creates a single, high-performance rack that can run the largest models, such as DeepSeek R1 671B and Llama 4 Maverick, with the fast inference. These racks can be seamlessly integrated into any existing air-cooled data center.


Dataflow architecture
Our innovative compute and memory chip layout enables seamless data flow between operations when processing Al models. This approach results in high-speed data traffic and significant gains in performance and efficiency.
Three-tier memory for efficiency
The SN40L design enables multiple models to run in memory and switch models in microseconds. This unique layout enables SambaNova to scale to the largest models, like DeepSeek and Llama 4 — all on a single rack.
