The fast and the curious: Chasing scalable AI dreams with Kubernetes and k0rdent
This session will explore how k0rdent automates AI infrastructure, GPU provisioning, and model deployment with KServe, while addressing the challenges of scaling GPUs, managing costs, and optimizing performance on Kubernetes.
Deploying AI at scale on Kubernetes can be a complex and costly endeavor, with challenges like GPU provisioning, cost management, and performance optimization. In this session, Bharath Nallapeta will demonstrate how k0rdent automates GPU-ready cluster provisioning, making AI deployment seamless across cloud and on-prem environments.
Attendees will learn how to serve AI models with KServe, dynamically scale GPU resources with Knative auto-scaling, and monitor performance with Prometheus and Grafana. The session will focus on strategies for maximizing compute efficiency and minimizing costs, particularly in a landscape where GPUs are both scarce and expensive.
A live demo will walk through the entire AI deployment workflow, from spinning up clusters to running real-time inference, offering insights into streamlining AI operations with Kubernetes without the typical complexity.
Attendees will learn how to serve AI models with KServe, dynamically scale GPU resources with Knative auto-scaling, and monitor performance with Prometheus and Grafana. The session will focus on strategies for maximizing compute efficiency and minimizing costs, particularly in a landscape where GPUs are both scarce and expensive.
A live demo will walk through the entire AI deployment workflow, from spinning up clusters to running real-time inference, offering insights into streamlining AI operations with Kubernetes without the typical complexity.
The fast and the curious: Chasing scalable AI dreams with Kubernetes and k0rdent
This session will explore how k0rdent automates AI infrastructure, GPU provisioning, and model deployment with KServe, while addressing the challenges of scaling GPUs, managing costs, and optimizing performance on Kubernetes.
Panelist

Panelist

Panelist

Moderator

Bharath Nallapeta
Senior Software Engineer, Mirantis
Deploying AI at scale on Kubernetes can be a complex and costly endeavor, with challenges like GPU provisioning, cost management, and performance optimization. In this session, Bharath Nallapeta will demonstrate how k0rdent automates GPU-ready cluster provisioning, making AI deployment seamless across cloud and on-prem environments.
Attendees will learn how to serve AI models with KServe, dynamically scale GPU resources with Knative auto-scaling, and monitor performance with Prometheus and Grafana. The session will focus on strategies for maximizing compute efficiency and minimizing costs, particularly in a landscape where GPUs are both scarce and expensive.
A live demo will walk through the entire AI deployment workflow, from spinning up clusters to running real-time inference, offering insights into streamlining AI operations with Kubernetes without the typical complexity.
Attendees will learn how to serve AI models with KServe, dynamically scale GPU resources with Knative auto-scaling, and monitor performance with Prometheus and Grafana. The session will focus on strategies for maximizing compute efficiency and minimizing costs, particularly in a landscape where GPUs are both scarce and expensive.
A live demo will walk through the entire AI deployment workflow, from spinning up clusters to running real-time inference, offering insights into streamlining AI operations with Kubernetes without the typical complexity.
The fast and the curious: Chasing scalable AI dreams with Kubernetes and k0rdent
This session will explore how k0rdent automates AI infrastructure, GPU provisioning, and model deployment with KServe, while addressing the challenges of scaling GPUs, managing costs, and optimizing performance on Kubernetes.
Deploying AI at scale on Kubernetes can be a complex and costly endeavor, with challenges like GPU provisioning, cost management, and performance optimization. In this session, Bharath Nallapeta will demonstrate how k0rdent automates GPU-ready cluster provisioning, making AI deployment seamless across cloud and on-prem environments.
Attendees will learn how to serve AI models with KServe, dynamically scale GPU resources with Knative auto-scaling, and monitor performance with Prometheus and Grafana. The session will focus on strategies for maximizing compute efficiency and minimizing costs, particularly in a landscape where GPUs are both scarce and expensive.
A live demo will walk through the entire AI deployment workflow, from spinning up clusters to running real-time inference, offering insights into streamlining AI operations with Kubernetes without the typical complexity.
Attendees will learn how to serve AI models with KServe, dynamically scale GPU resources with Knative auto-scaling, and monitor performance with Prometheus and Grafana. The session will focus on strategies for maximizing compute efficiency and minimizing costs, particularly in a landscape where GPUs are both scarce and expensive.
A live demo will walk through the entire AI deployment workflow, from spinning up clusters to running real-time inference, offering insights into streamlining AI operations with Kubernetes without the typical complexity.
The fast and the curious: Chasing scalable AI dreams with Kubernetes and k0rdent
This session will explore how k0rdent automates AI infrastructure, GPU provisioning, and model deployment with KServe, while addressing the challenges of scaling GPUs, managing costs, and optimizing performance on Kubernetes.
Panelist

Panelist

Panelist

Host

Bharath Nallapeta
Senior Software Engineer, Mirantis
Sign up now

