QwQ-32B
Qwen/QwQ-32B
Prompt: 0.65 credits / 1M tokens
•Completion: 0.65 credits / 1M tokens
Overview
QwQ-32B is a 32-billion parameter reasoning model noted for delivering performance comparable to much larger state-of-the-art models on complex reasoning and coding tasks. Developed through advanced reinforcement learning, it excels in mathematical reasoning and complex problem-solving. With a 128K context window and native tool use support, QwQ-32B is designed for handling extensive information and is highlighted for its speed and cost-efficiency on optimized hardware.
Tags
Chat
CServe Optimized
Dedicated
LLM
Serverless
API
curl -X POST "https://api.centml.com/openai/v1/chat/completions" \
-H "Authorization: Bearer *******************************************" \
-H "Content-Type: application/json" \
-d '{
"model": "Qwen/QwQ-32B",
"messages": [{ "role": "system", "content": "You are a helpful assistant." }],
"stream": false
}'
from openai import OpenAI
client = OpenAI(
api_key="*******************************************",
base_url="https://api.centml.com/openai/v1"
)
completion = client.chat.completions.create(
model="Qwen/QwQ-32B",
messages=[{ "role": "system", "content": "You are a helpful assistant." }],
stream=False,
)
print(completion.choices[0].message)
import OpenAI from "openai";
const client = new OpenAI(
api_key="*******************************************",
baseURL="https://api.centml.com/openai/v1"
)
async function main() {
const completion = await client.chat.completions.create({
model: "Qwen/QwQ-32B",
messages: [{ "role": "system", "content": "You are a helpful assistant." }],
stream: false,
});
console.log(completion.choices[0])
}
main()