Qwen2-VL-2B-Instruct
Qwen/Qwen2-VL-2B-Instruct
Overview
Qwen2-VL-2B-Instruct is a vision-language model with 2 billion parameters capable of understanding images of various resolutions and videos. It demonstrates state-of-the-art performance on several visual understanding benchmarks and can operate devices based on visual input and text instructions. Featuring architectural updates like Naive Dynamic Resolution and Multimodal Rotary Position Embedding, this model is designed for efficient multimodal processing and multilingual text understanding within images.
Tags
CentML Optimized
Chat
Dedicated
LLM
VLM
API
curl -X POST "https://api.centml.com/openai/v1/chat/completions" \
-H "Authorization: Bearer *******************************************" \
-H "Content-Type: application/json" \
-d '{
"model": "Qwen/Qwen2-VL-2B-Instruct",
"messages": [{ "role": "system", "content": "You are a helpful assistant." }],
"stream": false
}'
from openai import OpenAI
client = OpenAI(
api_key="*******************************************",
base_url="https://api.centml.com/openai/v1"
)
completion = client.chat.completions.create(
model="Qwen/Qwen2-VL-2B-Instruct",
messages=[{ "role": "system", "content": "You are a helpful assistant." }],
stream=False,
)
print(completion.choices[0].message)
import OpenAI from "openai";
const client = new OpenAI(
api_key="*******************************************",
base_url="https://api.centml.com/openai/v1"
)
async function main() {
const completion = await client.chat.completions.create({
model: "Qwen/Qwen2-VL-2B-Instruct",
messages: [{ "role": "system", "content": "You are a helpful assistant." }],
stream: false,
});
console.log(completion.choices[0])
}
main()