Skip to content

Together

Overview

Together is an inference provider hosting a wide range of open-source models including Llama, Mixtral, Qwen, and others. Floopy proxies requests to Together’s OpenAI-compatible API.

Supported Models

ModelContext WindowNotes
meta-llama/Llama-3.3-70B-Instruct-Turbo128KLlama 3.3 70B Turbo
meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8128KLlama 4 Maverick
Qwen/Qwen3-235B-A22B128KQwen 3 235B MoE
mistralai/Mixtral-8x22B-Instruct-v0.164KMixtral 8x22B
deepseek-ai/DeepSeek-V3128KDeepSeek V3

Setup

  1. Go to Settings > Providers in the dashboard.
  2. Click Add provider and select Together.
  3. Paste your Together API key and click Save.

Usage

import OpenAI from "openai";
const client = new OpenAI({
baseURL: "https://api.floopy.ai/v1",
apiKey: process.env.FLOOPY_API_KEY,
});
const response = await client.chat.completions.create({
model: "meta-llama/Llama-3.3-70B-Instruct-Turbo",
messages: [{ role: "user", content: "Explain quantum computing." }],
});

Provider-Specific Features

  • Large model catalog — Access to Llama, Qwen, Mixtral, DeepSeek, and many other open-source models.
  • Turbo variants — Some models offer Turbo variants optimized for faster inference at lower cost.
  • Model naming — Models use the org/Model-Name format (e.g., meta-llama/Llama-3.3-70B-Instruct-Turbo).