5 links
tagged with all of: machine-learning + kubernetes
Click any tag below to further narrow down your results
Links
Setting up a local Langfuse server with Kubernetes allows developers to manage traces and metrics for sensitive LLM applications without relying on third-party services. The article details the necessary tools and configurations, including Helm, Kustomize, and Traefik, to successfully deploy and access Langfuse on a local GPU cluster. It also provides insights on managing secrets and testing the setup through a Python container.
Software is transitioning towards genuine autonomy through agentic AI, which utilizes Large Language Models for proactive, goal-driven operations. Kubernetes offers a robust platform engineering foundation to meet the unique demands of agentic workloads, addressing challenges such as dynamic compute, persistent state management, and complex orchestration, while emphasizing the need for a platform-centric approach in deploying agentic AI at scale.
OpenAI leverages Kubernetes and Apache technologies to manage their scalable infrastructure effectively, ensuring that machine learning models can be deployed and maintained seamlessly. The integration of these tools allows for efficient resource management and orchestration, enabling OpenAI to handle complex workloads and enhance their service delivery.
The Kubeflow Trainer project has been integrated into the PyTorch ecosystem, providing a scalable and community-supported solution for running PyTorch on Kubernetes. It simplifies distributed training of AI models and fine-tuning of large language models (LLMs) while optimizing GPU utilization and supporting advanced scheduling capabilities. The integration enhances the deployment of distributed PyTorch applications and offers a streamlined experience for AI practitioners and platform admins alike.
Amazon Web Services has launched AI on EKS, an open source initiative aimed at simplifying the deployment and scaling of AI/ML workloads on Amazon Elastic Kubernetes Service. This project provides deployment-ready blueprints, Terraform templates, and best practices to optimize infrastructure for large language models and other AI tasks, while separating it from the previously established Data on EKS initiative to enhance focus and maintainability.