SambaNova
Overview
SambaNova provides high-performance inference using custom RDU (Reconfigurable Dataflow Unit) hardware, optimized for large language model serving. Floopy proxies requests to SambaNova’s OpenAI-compatible API.
Supported Models
| Model | Context Window | Notes |
|---|---|---|
Meta-Llama-3.3-70B-Instruct | 128K | Llama 3.3 70B |
DeepSeek-R1 | 128K | DeepSeek R1 reasoning model |
DeepSeek-V3-0324 | 128K | DeepSeek V3 |
Qwen3-32B | 128K | Qwen 3 32B |
Setup
- Go to Settings > Providers in the dashboard.
- Click Add provider and select SambaNova.
- Paste your SambaNova API key and click Save.
Usage
import OpenAI from "openai";
const client = new OpenAI({ baseURL: "https://api.floopy.ai/v1", apiKey: process.env.FLOOPY_API_KEY,});
const response = await client.chat.completions.create({ model: "Meta-Llama-3.3-70B-Instruct", messages: [{ role: "user", content: "Explain quantum computing." }],});from openai import OpenAI
client = OpenAI(base_url="https://api.floopy.ai/v1", api_key=os.environ["FLOOPY_API_KEY"])
response = client.chat.completions.create( model="Meta-Llama-3.3-70B-Instruct", messages=[{"role": "user", "content": "Explain quantum computing."}],)curl https://api.floopy.ai/v1/chat/completions \ -H "Authorization: Bearer $FLOOPY_API_KEY" \ -H "Content-Type: application/json" \ -d '{"model": "Meta-Llama-3.3-70B-Instruct", "messages": [{"role": "user", "content": "Explain quantum computing."}]}'Provider-Specific Features
- RDU acceleration — SambaNova’s custom hardware delivers fast inference for large models.
- High throughput — Optimized for batch processing and high-volume production workloads.
- Model naming — Models use PascalCase names (e.g.,
Meta-Llama-3.3-70B-Instruct).