A Technical Deep Dive into Pipeline Parallel Inference with CentML

Case Studies

A Technical Deep Dive into Pipeline Parallel Inference with CentML

With yesterday’s release of Llama-3.1-405B, we’re excited to announce that CentML’s recent contribution to vLLM, adding pipeline parallel inference support, […]

Read More

Maximizing LLM training and inference efficiency using CentML on OCI

Case Studies

Maximizing LLM training and inference efficiency using CentML on OCI

In partnership with CentML, Oracle has developed innovative solutions to meet the growing demand for high-performance NVIDIA GPUs for machine […]

Read More

GenAI company cuts training costs by 36% with CentML

Case Studies

GenAI company cuts training costs by 36% with CentML

A growing generative AI company partnered with CentML to accelerate their API-as-a-service and iterate with foundational models—all without using top-of-the-line […]

Read More

Get started

Let's make your LLM better! Book a Demo