Llama

Llama 3.2 1B Instruct

meta-llama/Llama-3.2-1B-Instruct

Overview

Llama 3.2 1B Instruct is a compact language model optimized for multilingual dialogue use cases, including agentic retrieval and summarization. Despite its smaller size, it aims to outperform many other open-source and closed chat models on common benchmarks. Utilizing an optimized transformer architecture and fine-tuning with SFT and RLHF, this model is designed for efficient handling of text generation, summarization, and conversation in multiple languages.

Tags

CentML Optimized

Chat

Dedicated

LLM

API

Select a language:
curl -X POST "https://api.centml.com/openai/v1/chat/completions" \
  -H "Authorization: Bearer *******************************************" \
  -H "Content-Type: application/json" \
  -d '{
    "model": "meta-llama/Llama-3.2-1B-Instruct",
    "messages": [{ "role": "system", "content": "You are a helpful assistant." }],
    "stream": false
  }'
from openai import OpenAI

client = OpenAI(
	api_key="*******************************************",
	base_url="https://api.centml.com/openai/v1"
)

completion = client.chat.completions.create(
    model="meta-llama/Llama-3.2-1B-Instruct",
    messages=[{ "role": "system", "content": "You are a helpful assistant." }],
    stream=False,
)

print(completion.choices[0].message)
import OpenAI from "openai";

const client = new OpenAI(
	api_key="*******************************************",
	base_url="https://api.centml.com/openai/v1"
)

async function main() {
  const completion = await client.chat.completions.create({
    model: "meta-llama/Llama-3.2-1B-Instruct",
    messages: [{ "role": "system", "content": "You are a helpful assistant." }],
    stream: false,
  });

  console.log(completion.choices[0])
}

main()