Session
From Hugging Face to Cloud Native: Igniting the LLM Revolution with Kubernetes and Open Source Tools
In the wave of AI, open-source large language models (LLMs) such as LLaMA, Gemma, and DeepSeek are reshaping the technological landscape, but how to efficiently deploy these models from prototype to production remains a challenge for developers. This presentation will share how to build a scalable and efficient LLMOps pipeline using Kubernetes and open-source tools, covering the entire process from model download to inference optimization. Based on my experience at DaoCloud and open-source projects (such as the Hugging Face model download GUI and Kubernetes configuration tool), I will demonstrate how cloud-native technologies can simplify LLM deployment, including practical cases of automated model distribution, dynamic resource scheduling, and inference acceleration.

Samzong Lu
PM at DaoCloud, AI/LLMOps PM Leader, CNCF Multiple Project Contributors, Open Source Enthusiast
Shanghai, China
Links
Please note that Sessionize is not responsible for the accuracy or validity of the data provided by speakers. If you suspect this profile to be fake or spam, please let us know.
Jump to top