Skip to content

Nebius

Overview

Nebius AI Studio is an inference provider offering a curated set of open-source models with competitive pricing and optional fast-inference tiers. Floopy proxies requests to Nebius’ OpenAI-compatible API.

Supported Models

ModelContext WindowNotes
deepseek-ai/DeepSeek-R1-0528128KDeepSeek R1 reasoning model
deepseek-ai/DeepSeek-V3-0324128KDeepSeek V3
Qwen/Qwen3-Coder-480B-A35B-Instruct128KQwen 3 Coder flagship
Qwen/Qwen3-235B-A22B-Thinking-2507128KQwen 3 reasoning model
Qwen/Qwen3-32B128KQwen 3 32B
meta-llama/Llama-3.3-70B-Instruct128KLlama 3.3 70B
meta-llama/Llama-3.1-405B-Instruct128KLlama 3.1 405B

Setup

  1. Go to Settings > Providers in the dashboard.
  2. Click Add provider and select Nebius.
  3. Paste your Nebius API key and click Save.

Usage

import OpenAI from "openai";
const client = new OpenAI({
baseURL: "https://api.floopy.ai/v1",
apiKey: process.env.FLOOPY_API_KEY,
});
const response = await client.chat.completions.create({
model: "Qwen/Qwen3-32B",
messages: [{ role: "user", content: "Explain quantum computing." }],
});

Provider-Specific Features

  • Fast inference tiers — Some models offer a “fast” tier at higher cost for lower latency.
  • Batch inference — Batch processing at 50% of real-time pricing.
  • Model naming — Models use the Org/Model-Name format (e.g., Qwen/Qwen3-32B).