Docs

DOCUMENTATION

Get your first response in under 2 minutes.

Quick Start

Get your first response in 3 steps:

  1. Create a free account — $2 credit, no card required.
  2. Copy your API key from the dashboard.
  3. Make your first API call:
bash
curl https://api.llmdiscount.com/v1/chat/completions \
-H "Authorization: Bearer YOUR_KEY" \
-H "Content-Type: application/json" \
-d '{"model":"qwen3.5-flash","messages":[{"role":"user","content":"Hello!"}]}'

Authentication

QwenBridge uses Bearer token authentication — the same format as OpenAI. Include your API key in every request:

Authorization: Bearer YOUR_QWENBRIDGE_KEY

Get your key at api.qwenbridge.com → Dashboard → API Keys.

Models

Available models and their IDs:

Model IDNameInput/1MOutput/1M
qwen3.6-plusQwen3.6-Plus$0.22$1.20
qwen3-maxQwen3-Max$0.55$2.80
qwen3.5-plusQwen3.5-Plus$0.22$1.30
qwq-plusQwQ-Plus$0.50$1.50
qwen3.5-flashQwen3.5-Flash$0.07$0.35

SDK Examples

The OpenAI SDK works without modification — just change the base URL and API key.

python
from openai import OpenAI
client = OpenAI(
base_url="https://api.llmdiscount.com/v1",
api_key="YOUR_QWENBRIDGE_KEY"
)
response = client.chat.completions.create(
model="qwen3.6-plus",
messages=[
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": "Hello, how are you?"}
]
)
print(response.choices[0].message.content)

Cursor

Drop-in replacement for OpenAI in Cursor's model config.

  1. Open Cursor → Settings → Models
  2. Add model: "qwen3.6-plus" (or any QB model)
  3. Set OpenAI Base URL to: https://api.llmdiscount.com/v1
  4. Paste your QwenBridge API key
  5. Test with Ctrl+K
openaiBaseUrl = "https://api.llmdiscount.com/v1"

Cline

Set QwenBridge as your OpenAI-compatible provider in VS Code.

  1. Install Cline from VS Code marketplace
  2. Set API Provider: OpenAI Compatible
  3. Base URL: https://api.llmdiscount.com/v1
  4. Paste your QwenBridge API key
  5. Select model: qwen3.6-plus
cline.openAiBaseUrl = "https://api.llmdiscount.com/v1"

Continue.dev

Configure QwenBridge as an OpenAI provider in config.json.

  1. Open ~/.continue/config.json
  2. Add model with provider: "openai"
  3. Set baseURL: "https://api.llmdiscount.com/v1"
  4. Set apiKey to your QwenBridge token
  5. Choose model: qwen3.5-plus
baseURL = "https://api.llmdiscount.com/v1"

Windsurf

Use QwenBridge models in Windsurf via OpenAI-compatible config.

  1. Open Windsurf → Preferences → AI
  2. Choose provider: OpenAI Compatible
  3. Set base URL: https://api.llmdiscount.com/v1
  4. Enter QwenBridge API key
  5. Select qwen3.6-plus as default model
openai.baseUrl = "https://api.llmdiscount.com/v1"

Endpoints

Base URL: https://api.llmdiscount.com/v1

MethodPathDescription
POST/chat/completionsCreate a chat completion. OpenAI-compatible.
GET/modelsList all available models.

Error Codes

CodeLabelDescription
400Bad RequestMalformed request body or missing required fields.
401UnauthorizedInvalid or missing API key.
429Too Many RequestsRate limit exceeded. Retry after the specified delay.
500Internal Server ErrorSomething went wrong on our end. Retry with backoff.

Rate Limits

Free accounts: 60 requests/minute, 100K tokens/day

Paid accounts: 600 requests/minute, unlimited tokens

All limits are per API key. If you hit a rate limit, the API returns 429 Too Many Requests with a Retry-After header.

python
# Handle rate limits with exponential backoff
import time, random
for attempt in range(5):
try:
resp = client.chat.completions.create(...)
break
except openai.RateLimitError:
wait = 2 ** attempt + random.random()
time.sleep(wait)