Llama 3.2 11B Vision Instruct
meta-llama/Llama-3.2-11B-Vision-Instruct
Overview
This model is part of Meta's Llama 3.2 Vision family, a multimodal large language model (LMM) integrating image and text reasoning. With 11 billion parameters, it is designed for a wide array of vision-language tasks including visual recognition, image reasoning, and captioning. Llama 3.2 11B Vision Instruct offers strong performance across various image-text benchmarks and is instruction-tuned to follow directions accurately. Its capabilities make it suitable for applications like Visual Question Answering (VQA) and generating detailed image descriptions.
Tags
CentML Optimized
Chat
Dedicated
LLM
VLM
API
curl -X POST "https://api.centml.com/openai/v1/chat/completions" \
-H "Authorization: Bearer *******************************************" \
-H "Content-Type: application/json" \
-d '{
"model": "meta-llama/Llama-3.2-11B-Vision-Instruct",
"messages": [{ "role": "system", "content": "You are a helpful assistant." }],
"stream": false
}'
from openai import OpenAI
client = OpenAI(
api_key="*******************************************",
base_url="https://api.centml.com/openai/v1"
)
completion = client.chat.completions.create(
model="meta-llama/Llama-3.2-11B-Vision-Instruct",
messages=[{ "role": "system", "content": "You are a helpful assistant." }],
stream=False,
)
print(completion.choices[0].message)
import OpenAI from "openai";
const client = new OpenAI(
api_key="*******************************************",
base_url="https://api.centml.com/openai/v1"
)
async function main() {
const completion = await client.chat.completions.create({
model: "meta-llama/Llama-3.2-11B-Vision-Instruct",
messages: [{ "role": "system", "content": "You are a helpful assistant." }],
stream: false,
});
console.log(completion.choices[0])
}
main()