Author: ermek
![A Technical Deep Dive into Pipeline Parallel Inference with CentML](https://centml.ai/wp-content/uploads/2024/07/Pipeline_parallel_inference_main_image-900x505.jpg)
A Technical Deep Dive into Pipeline Parallel Inference with CentML
With yesterday’s release of Llama-3.1-405B, we’re excited to announce that CentML’s recent contribution to vLLM, adding pipeline parallel inference support, […]
![Hardware Efficiency in the Era of LLM Deployments](https://centml.ai/wp-content/uploads/2024/04/Hardware_Efficiency_img-900x505.jpg)
Hardware Efficiency in the Era of LLM Deployments
How CServe can make LLM deployment easy, efficient and scalable