Session

On-Device LLM Inference for Privacy-First Web Apps

In this hands-on workshop, you will learn to deploy Google's Gemma-3n models directly in web browsers using MediaPipe's LLM Inference API. We'll build privacy-first applications that run completely on-device with WebGPU acceleration, starting with text generation and progressing to Gemma-3n's multimodal capabilities that process text, images, and audio inputs simultaneously.

We will also explore advanced features like LoRA fine-tuning for custom models, all while ensuring user data never leaves the browser for complete privacy and instant responses.

Patty O'Callaghan

Technical Director - Head of AI Architecture and Engineering Group @ Charles River Laboratories | Google Developer Advisory Board | Google Developer Expert in AI/ML

Glasgow, United Kingdom

Actions

Please note that Sessionize is not responsible for the accuracy or validity of the data provided by speakers. If you suspect this profile to be fake or spam, please let us know.

Jump to top