Qwen

Qwen2-VL-2B-Instruct

Qwen/Qwen2-VL-2B-Instruct

Overview

Qwen2-VL-2B-Instruct is a vision-language model with 2 billion parameters capable of understanding images of various resolutions and videos. It demonstrates state-of-the-art performance on several visual understanding benchmarks and can operate devices based on visual input and text instructions. Featuring architectural updates like Naive Dynamic Resolution and Multimodal Rotary Position Embedding, this model is designed for efficient multimodal processing and multilingual text understanding within images.

Tags

CentML Optimized

Chat

Dedicated

LLM

VLM

API

Select a language:
curl -X POST "https://api.centml.com/openai/v1/chat/completions" \
  -H "Authorization: Bearer *******************************************" \
  -H "Content-Type: application/json" \
  -d '{
    "model": "Qwen/Qwen2-VL-2B-Instruct",
    "messages": [{ "role": "system", "content": "You are a helpful assistant." }],
    "stream": false
  }'
from openai import OpenAI

client = OpenAI(
	api_key="*******************************************",
	base_url="https://api.centml.com/openai/v1"
)

completion = client.chat.completions.create(
    model="Qwen/Qwen2-VL-2B-Instruct",
    messages=[{ "role": "system", "content": "You are a helpful assistant." }],
    stream=False,
)

print(completion.choices[0].message)
import OpenAI from "openai";

const client = new OpenAI(
	api_key="*******************************************",
	base_url="https://api.centml.com/openai/v1"
)

async function main() {
  const completion = await client.chat.completions.create({
    model: "Qwen/Qwen2-VL-2B-Instruct",
    messages: [{ "role": "system", "content": "You are a helpful assistant." }],
    stream: false,
  });

  console.log(completion.choices[0])
}

main()