Session
Mastering Massive AI: Next-Level Model & Data Parallelism for Production-Scale Impact
Scaling large AI models is less about raw compute and more about engineering precision. This session distills proven techniques for model and data parallelism drawn from real-world, production-scale deployments. We’ll cover how to partition massive models and datasets to maximize throughput, minimize communication overhead, and maintain accuracy under strict latency and cost constraints. You’ll see practical benchmarks, configuration patterns, and profiling results that reveal what actually works at multi-node, multi-GPU scale - along with the trade-offs that matter when the stakes are high. Whether you’re training in the cloud or optimizing inference pipelines, you’ll leave with an actionable framework for designing, evaluating, and deploying parallelized AI workflows that meet performance targets in production.
Shashank Kapadia
Machine Learning Engineering | Building Scalable AI Solutions | NLP & Personalization | Ethical AI Advocate | Mentor | Writer
Sunnyvale, California, United States
Links
Please note that Sessionize is not responsible for the accuracy or validity of the data provided by speakers. If you suspect this profile to be fake or spam, please let us know.
Jump to top