Click any tag below to further narrow down your results
Links
Kthena is a new system tailored for Kubernetes that optimizes the routing, orchestration, and scheduling of Large Language Model (LLM) inference. It addresses key challenges like resource utilization and latency, offering features such as intelligent routing and production-grade orchestration. This sub-project of Volcano enhances support for AI lifecycle management.
This article explains how to implement large-scale inference for language models using Kubernetes. It covers key concepts like batching strategies, performance metrics, and intelligent routing to optimize GPU usage. Practical deployment examples and challenges in managing inference are also discussed.
KServe v0.15 has been released, enhancing capabilities for serving generative AI models, including support for large language models (LLMs) and advanced caching mechanisms. Key features include integration with Envoy AI Gateway, multi-node inference, and autoscaling with KEDA, aimed at improving performance and scalability for AI workloads. The update also introduces a dedicated documentation section for generative AI and various performance optimizations.
The guide outlines how to deploy large language models (LLMs) at scale using Google Kubernetes Engine (GKE) and the GKE Inference Gateway, which optimizes load balancing by considering AI-specific metrics. It provides a step-by-step walkthrough for setting up an inference pipeline with the vLLM framework, ensuring efficient resource management and performance for AI workloads. Key features include intelligent load balancing, simplified operations, and support for multiple models and hardware configurations.