Blog 2

Use this space to tell everyone about what you have to offer.

May 1, 2024

Tokens Per Second is Not All You Need

In the fast-paced world of LLM inference, there's been a growing buzz around achieving high tokens per second...

April 11, 2024

Samba-CoE v0.3: The Power of Routing ML Models at Scale

*A twitter user said that CausalLM-34b-beta is suspected to have MMLU contamination. On further investigation we do...

April 10, 2024

Responsible AI

Generative AI will be the defining technology of this century, fundamentally reshaping how businesses, governments,...

April 8, 2024

SambaLingo hits 15,000+ downloads, now integrated with Samba-CoE-v0.2

SambaLingo, our cutting-edge multilingual language expert series, surpassed 15k downloads and is now integrated into...

March 27, 2024

AI Power: Accurate Models at Blazing Speeds | SambaNova

In late February, we announced Samba-1, a CoE architecture that is a paradigm-shift, and will ultimately become the...

March 21, 2024

Using Mixed Precision on RDUs

Deep learning frameworks have traditionally relied upon the 32-bit single-precision (FP32) format. However, FP32...

February 28, 2024

Benchmarking Samba-1

Today, we announced Samba-1, a Composition of Experts model with over 1 trillion parameters, built on top of open...

February 28, 2024

Samba-CoE v0.1: Unlocking the Power of Experts

We're thrilled to unveil Samba-CoE-v0.1, a scaled down version of Samba-1, our latest breakthrough model that...

February 26, 2024

High-Accuracy AI Models in 9 Languages

SambaNova is excited to open source a collection of expert models that adapt Llama 2 [12] to a diverse set of 9...

February 13, 2024

SambaCoder-nsql-Llama-2-70B model

Using a Text-to-SQL generative AI solution can have a significant impact on enterprise organizations. This will...

February 7, 2024

BLOOMChat-v2 Long Sequences at 176B

We are proud to release BLOOMChat-v2, a 32K sequence length, 176B multilingual language model trained on top of...

January 25, 2024

SambaNova RDAs: Mastering Fault Management (Part 2)

In Part 1 of this two-part blog, we discussed the SambaNova product components, discussed some fault management...

January 25, 2024

SambaNova RDAs: Mastering Fault Management (Part 1)

SambaNova delivers the first, full stack enterprise grade generative AI platform. Designed specifically for the...

December 22, 2023

ALiBi Deep Dive: Interpolation vs. Extrapolation

Introduction Long sequence capabilities enable a variety of applications like summarizing long form documents, such...

November 21, 2023

Elevating Information Retrieval and Augmenting Large Language Models

We are pleased to announce that SambaStudio now supports text embedding models. This new feature significantly...

September 8, 2023

Enabling Open Source LLMs to Become Effective Tool Manipulators

INTRODUCTION Using tools can extend the capabilities of LLMs’ to access knowledge beyond their training data. As an...

August 7, 2023

Training long sequence size models with SambaNova

At SambaNova, we have been researching and developing methods to train long sequence size (SS) models on our...

May 19, 2023

BLOOMChat: Open-Source Multilingual Chat LLM

[1] The image is created with Midjourney Highlights SambaNova, in collaboration with Together, is excited to present...

February 13, 2023

Achieving GPT 175B Level Accuracy with a 10x More Efficient Model

In this blogpost, we show how one can use the SambaNova platform to develop a GPT 13B parameter model that can...

December 22, 2022

Achieving Best-in-Class Large Language Model Accuracy in Low-Resource Settings

The opportunity: solving a range of language tasks using large language models with zero and few-shot learning

December 16, 2022

Dataflow Architecture Leads to a Performance Breakthrough on GNN Fused Kernels

A Collaboration between SambaNova Systems and Argonne National Laboratory Using the capabilities of the SambaNova...

Subscribe

Lorem ipsum dolor sit amet, consectetur adipiscing elit. Nullam ultrices massa sit amet auctor scelerisque. Cras vel quam non lorem tincidunt facilisis.