Blog 2
Use this space to tell everyone about what you have to offer.
SambaNova Holds Speed Record on Llama 3.1 405B - 4X faster than the rest
In today's fast-paced business landscape, enterprises need more than just the latest AI model to solve their biggest...
Does reduced precision hurt? A bit about losing bits.
SambaNova and Groq recently achieved 1000 tokens per second on their inference system for Meta’s LLaMa 3 8b Instruct...
Introducing Fugaku-LLM in Composition of Experts
The Composition of Experts (CoE) architecture that the Samba-1 model is based upon has many features that make it...
Sovereign AI
Artificial intelligence has become vital to nations, governments, and large corporations. Many of these large...
NAIRR; Govt-funded AI Research Resources
Artificial intelligence (AI) is driving the next generation of technological innovation and scientific discovery....
Tokens Per Second is Not All You Need
In the fast-paced world of LLM inference, there's been a growing buzz around achieving high tokens per second...
Samba-CoE v0.3: The Power of Routing ML Models at Scale
*A twitter user said that CausalLM-34b-beta is suspected to have MMLU contamination. On further investigation we do...
Responsible AI
Generative AI will be the defining technology of this century, fundamentally reshaping how businesses, governments,...
SambaLingo hits 15,000+ downloads, now integrated with Samba-CoE-v0.2
SambaLingo, our cutting-edge multilingual language expert series, surpassed 15k downloads and is now integrated into...
Subscribe
Lorem ipsum dolor sit amet, consectetur adipiscing elit. Nullam ultrices massa sit amet auctor scelerisque. Cras vel quam non lorem tincidunt facilisis.