Resources

The Only Inference Provider with High Speed Support for the Largest Models

Written by SambaNova Systems | April 1, 2025

With support for DeepSeek R1 671B running at 250 tokens per second per user, SambaNova continues to be the only provider to offer high speed inference on the largest open source models.