Blueprints
June 12

Engineering Cloud Native AI Platform

This talk presents various best practices and challenges in building large, efficient, scalable, and reliable AI/ML platforms using cloud-native technologies.

In recent years, advances in ML/AI have made tremendous progress. Yet designing large-scale data science and ML applications still remains challenging. The variety of ML frameworks, hardware accelerators, and cloud vendors, as well as the complexity of data science workflows, brings new challenges to MLOps. For example, it’s non-trivial to build an inference system suitable for models of different sizes, especially for LLMs or large models in general.

This talk presents various best practices and challenges in building large, efficient, scalable, and reliable AI/ML platforms using cloud-native technologies such as Kubernetes, Kubeflow, and KServe. We’ll deep dive into a reference platform dedicated to modern cloud-native AI infrastructure.

Yuan Tang
Principal Engineer, Red Hat; Project Lead at Argo and Kubeflow
Yuan Tang

Register for PlatformCon 2024

Connect with fellow platform practitioners, learn from the best in the industry and engage directly with speakers on Slack.
Community
Join over 20k platform engineers from all over the world
Slack
Share best practices, discuss new trends and tooling with 20k+ platform practitioners
Speakers
Engage with 100+ speakers in dedicated channels or directly in DMs