Skip to content

DeepInfra

Overview

DeepInfra is an inference provider hosting a large catalog of open-source and commercial models. Floopy proxies requests to DeepInfra’s OpenAI-compatible API.

Supported Models

ModelContext WindowNotes
deepseek-ai/DeepSeek-V3.1128KDeepSeek V3.1
deepseek-ai/DeepSeek-R1-0528-Turbo128KDeepSeek R1 Turbo
Qwen/Qwen3-235B-A22B128KQwen 3 235B MoE
Qwen/Qwen3-32B128KQwen 3 32B
meta-llama/Llama-3.3-70B-Instruct-Turbo128KLlama 3.3 70B Turbo
meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8128KLlama 4 Maverick
microsoft/phi-4-reasoning-plus16KPhi 4 Reasoning Plus
google/gemma-4-26B-A4B-it128KGemma 4 26B

Setup

  1. Go to Settings > Providers in the dashboard.
  2. Click Add provider and select DeepInfra.
  3. Paste your DeepInfra API key and click Save.

Usage

import OpenAI from "openai";
const client = new OpenAI({
baseURL: "https://api.floopy.ai/v1",
apiKey: process.env.FLOOPY_API_KEY,
});
const response = await client.chat.completions.create({
model: "meta-llama/Llama-3.3-70B-Instruct-Turbo",
messages: [{ role: "user", content: "Explain quantum computing." }],
});

Provider-Specific Features

  • Large model catalog — Access to 60+ text-generation models from DeepSeek, Qwen, Llama, Mistral, Microsoft, Google, and others.
  • Turbo variants — Some models offer Turbo variants optimized for faster inference.
  • Model naming — Models use the org/Model-Name format (e.g., meta-llama/Llama-3.3-70B-Instruct-Turbo).