Model Library
Llama Guard 4 12B
Designed to detect inappropriate content in both images and text used as input or generated as output by models. Llama Guard 4 can evaluate text-only and image+text inputs.
Serverless
Deepseek V3 03-24
An iteration of the DeepSeek V3 model with notable improvements in reasoning capabilities across various benchmarks, including MMLU-Pro, GPQA, and AIME.
Serverless
Qwen2-VL-7B-Instruct
An advanced vision-language model. It is purpose-built for tasks that involve both visual and textual understanding, including image captioning, visual question answering, and content generation.
Serverless
Qwen2-VL-2B-Instruct
A powerful multimodal model that excels in visual understanding tasks, including image and video comprehension.
Llama 3.2 1B Instruct
A lightweight, instruction-tuned large language model designed for multilingual dialogue and tasks like agentic retrieval and summarization.
Llama 3.1 8B
An 8-billion-parameter language model, optimized for instruction following and reasoning. It delivers strong performance on benchmarks while maintaining a small, efficient footprint.
QwQ-32B
A 32-billion-parameter reasoning model developed through advanced reinforcement learning, noted for matching the performance of much larger state-of-the-art models on complex reasoning and coding tasks.
Serverless
Qwen2.5-VL-7B-Instruct
A 7B-parameter multimodal model, designed for complex vision-language tasks. Ideal for document understanding, video summarization, and interactive applications requiring both visual perception and language reasoning.
Serverless
Phi-4-mini-instruct
Phi-4-mini-instruct is a lightweight open model built upon synthetic data and filtered publicly available websites - with a focus on high-quality, reasoning dense data.
Serverless
Llama 4 Scout 17B (16E) Instruct
A mixture-of-experts (MoE) language model activating 17 billion parameters out of a total of 109B. Designed for assistant-style interaction and visual reasoning.
Serverless