Keep your prompts and message format. Swap your client to the official LLMWise SDK and get multi-model routing, failover, and orchestration on top of one API key.
You only pay credits per request. No monthly subscription. Paid credits never expire.
Replace multiple AI subscriptions with one wallet that includes routing, failover, and optimization.
pip install llmwise # or npm i llmwise
# Python (LLMWise SDK)
# pip install llmwise
import os
from llmwise import LLMWise
client = LLMWise(os.environ["LLMWISE_API_KEY"])
# Messages keep the same shape: role + content
resp = client.chat(
model="auto",
messages=[
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": "Compare Python and Rust for backend development."},
],
max_tokens=512,
)
print(resp["content"])
# TypeScript (LLMWise SDK)
// npm i llmwise
import { LLMWise } from "llmwise";
const tsClient = new LLMWise(process.env.LLMWISE_API_KEY!);
for await (const ev of tsClient.chatStream({
model: "claude-sonnet-4.5",
messages: [{ role: "user", content: "Write a TypeScript utility type for deep partial." }],
})) {
if (ev.delta) process.stdout.write(ev.delta);
if (ev.event === "done") break;
}Everything you need to integrate LLMWise's multi-model API into your Migration project.
Sign up at llmwise.ai and copy your API key from the dashboard. You get 40 free trial credits to start, then continue with non-expiring paid credits. No credit card required.
export LLMWISE_API_KEY="your_api_key_here"
Use the official SDKs for Python and TypeScript. Your prompt/message structure stays the same; you only swap the client call.
pip install llmwise # or npm i llmwise
LLMWise uses OpenAI-style role/content messages. Choose a model or use model="auto" to route by goal.
resp = client.chat(
model="auto",
messages=[{"role": "user", "content": "Hello, world!"}],
)
print(resp["content"])Specify a fallback chain so requests can retry on another model when the primary is rate-limited or failing.
resp = client.chat(
model="gpt-5.2",
routing={"strategy": "rate-limit", "fallback": ["claude-sonnet-4.5", "gemini-3-flash"]},
messages=[{"role": "user", "content": "Summarize this incident report."}],
)
print(resp["content"])Use Compare to benchmark models, Blend to synthesize, and Judge to score responses. These are native modes designed for production eval and routing decisions.
resp = client.compare(
models=["gpt-5.2", "claude-sonnet-4.5", "gemini-3-flash"],
messages=[{"role": "user", "content": "Explain eventual consistency."}],
)
print([r["model"] for r in resp["responses"]])You only pay credits per request. No monthly subscription. Paid credits never expire.
Replace multiple AI subscriptions with one wallet that includes routing, failover, and optimization.