Just change the base URL. No other code changes needed.
Python
from openai import OpenAI
client = OpenAI(
base_url="https://your-domain/api/v1",
api_key="sk-or-..."
)
resp = client.chat.completions.create(
model="gpt-4o",
messages=[{
"role": "user",
"content": "Hello!"
}]
)JavaScript
import OpenAI from "openai";
const client = new OpenAI({
baseURL: "https://your-domain/api/v1",
apiKey: "sk-or-..."
});
const resp = await client.chat
.completions.create({
model: "gpt-4o",
messages: [{
role: "user",
content: "Hello!"
}]
});Deployed on Vercel's edge network. Optional Aliyun reverse proxy for guaranteed connectivity.
Works with any OpenAI SDK. Just change the base URL. Model is fixed server-side.
SSE streaming support for real-time responses. Up to 300 seconds continuous streaming.
Admin creates and distributes API keys. SHA-256 hashed storage.
See exactly how many tokens each key has used. Track costs per key.
Built-in rate limiting and automatic retry on transient failures.