
Shubham Agnihotri
Chief Manager - Generative AI - IDFC Bank
Mumbai, India
Actions
Shubham Agnihotri is a Chief Manager – Generative AI at IDFC FIRST Bank, with 6+ years of experience working across diverse modalities like text, images, audio, video, tabular, and graph data. He specializes in fine-tuning and deploying Large Language Models (LLMs) and multimodal AI systems in production environments, contributing to real-world impact at scale.
Shubham is an active community leader and currently serves as an organizer for the AWS User Group Mumbai, where he helps nurture the cloud and AI ecosystem through meetups, hackathons, and hands-on sessions. Previously, he co-organized TensorFlow User Group (TFUG) Bangalore, leading over a dozen developer events.
He is also the founder of an agriculture drone startup that enabled precision spraying services for farmers. His solution helped save over 1 million liters of drinking water, reduce pesticide usage by 40%, and was recognized as an Emerging Agri-Drone Startup by the Times of India Group. This mission-driven venture drives his continued passion for building tech with societal impact.
🎤 Featured Speaking Engagements
AsiaTech Singapore -
1. https://asiatechxsg.com/techxlr8asia/speakers/shubham-agnihotri/
2. https://asiatechxsg.com/developerxperience-summit/speakers/shubham-agnihotri/
Aws Community Days Bangalore: https://acd.awsugblr.in/
Google Cloud Community Day: https://www.linkedin.com/feed/update/urn:li:activity:7337524380920348674/
TechShow London : https://youtu.be/jx_9SbARV9M?si=6PN1h1TYhgAPeQ4z
AWS Community Day Mumbai : https://youtu.be/XihAhZQZtV4?si=ST8XYvuRTdgoArLq
In addition to community and enterprise work, Shubham is part of the Google AI Community Network and was the founding lead of Cynergy, the official coding group of Ramaiah Institute, where he fostered a peer-to-peer learning culture.
Links
Area of Expertise
Topics
Case study - Unlocking Customer Insights: AI-Powered Conversation Intelligence for Banks
Customer interactions hold a wealth of insights, but manual evaluation is slow, inconsistent, and reactive. This session reveals how AI-powered conversation intelligence transforms raw call audio into actionable business insights. By leveraging Google’s advanced speech-to-text and emotion-analysis technologies, Neo4j graph databases, and Generative AI, we automate the identification of customer pain points, sentiment trends, and agent performance metrics—enabling real-time, scalable quality assurance and continuous coaching.
We’ll walk through a cross-functional pipeline: from live call ingestion and transcription, sentiment and intent analysis, to structured graph storage in Neo4j, and finally visualization of interaction patterns and agent scorecards. Attendees will see how this framework equips managers with real-time dashboards, automated evaluation frameworks, and triggers for root-cause investigations—raising the bar for customer care excellence.
Key Takeaways
- From Voice to Insights via Google’s Audio Tech
# Learn how to use Google’s speech-to-text API, audio emotion detection, and sentiment analysis to extract intent, emotion, and language patterns from live customer calls.
# Understand the importance of accurately capturing both what customers say and how they say it.
- Modeling Conversational Analytics in Neo4j
# See how to structure call metadata, speakers, intents, sentiment flows, and tags within a Neo4j graph
# Discover why relationships (e.g., call → query type → agent response) empower pattern discovery and timely insight alerts .
- Automated Detection of Pain Points & Coaching Needs
# Learn how AI identifies recurring customer complaints, root-causes, and potential service failures automatically
# Dive into agent performance scoring frameworks—measuring factors like empathy, response accuracy, and compliance—at scale.
- Live Quality Scoring & Manager Dashboards
# Understand how real-time dashboards ingest graph updates from Neo4j to reflect emerging trends, agent scoring, and customer sentiment distribution.
# Explore alerting systems that notify managers of low-quality exchanges or rising negative sentiment clusters.
- Scalable, Explainable, and Continuous Improvement
# See how graphs enable explainability: managers can trace a low score to specific call segments and transcripts.
# Learn how the pipeline supports continuous learning—enhancing speech models, conversation flows, and agent training based on data-driven insights.
Autonomous Knowledge Agents: Building a Next-Gen RAG System with AWS Bedrock, Claude, Lambda & S3
Step into the future of AI-powered retrieval with Agentic RAG—where LLMs don’t just fetch data, but think, reason, and act autonomously. In this session, we’ll harness the power of AWS Bedrock, Claude, Lambda, and S3 to build a next-gen self-improving retrieval system.
Key takeaways:
- Claude on Bedrock for dynamic, context-aware generation.
- S3 as a scalable knowledge vault for seamless retrieval.
- Lambda-driven orchestration for real-time, event-driven processing.
- Agentic workflows that refine retrieval and response autonomously.
Accelerating Training with Multi-GPU: Using PyTorch Lightning for PaLI-Gemma VLM
Dive into the world of high-performance model training with multi-GPU setups using PyTorch Lightning, featuring the PaLI-Gemma Vision-Language Model (VLM). This session will cover the essential tools and techniques for setting up efficient distributed training pipelines, allowing you to fully utilize multiple GPUs to speed up and scale model training. We’ll explore PyTorch Lightning’s seamless integration for handling multi-GPU workloads, making it accessible even for those new to distributed training. Attendees will get a step-by-step guide to configuring PaLI-Gemma for accelerated training, from data handling to model synchronization and optimization strategies. By the end, you'll have the skills needed to harness the power of multi-GPU training to scale multimodal AI models effectively.
MultiModel RAGs: Unlocking the Power of Multimodality with PaliGemma
Unlocking the Potential of Multimodality with Pali-Gemma. In this session, I’ll dive into the concepts and applications of multimodality, showcasing how combining multiple forms of data can enhance AI model capabilities. I’ll walk through the process of training on Colab/Kaggle, highlighting each step from data preparation to fine-tuning. Following the training phase, I’ll demonstrate how we can leverage Hugging Face Spaces for inference and deployment, providing a practical approach to deploying powerful Vision-Language Models (VLMs) in real-world applications. This session will serve as a comprehensive guide for anyone interested in building and deploying multimodal AI solutions, showing how PaLI-Gemma, a state-of-the-art model, is trained and implemented. By the end, participants will have a solid understanding of how multimodal RAGs work and how to bring them to production efficiently.
Building Transformers: Code Your Vision
From hugging face transformers to transformers from scratch in python, hands on Workshop with Transformers brush ups to implementing it.
Cloud Community Days 2025 Mumbai Sessionize Event
AWS Community Day Bengaluru 2025 Sessionize Event
Devfest Mumbai, 2023 Sessionize Event

Shubham Agnihotri
Chief Manager - Generative AI - IDFC Bank
Mumbai, India
Links
Actions
Please note that Sessionize is not responsible for the accuracy or validity of the data provided by speakers. If you suspect this profile to be fake or spam, please let us know.
Jump to top