Session
Introduction to evaluating generative AI applications
Demos are easy; dependable AI is not. This session shows how to verify that your LLM apps actually work using Azure AI Foundry and Microsoft's agent framework. We’ll cover what to measure and how to spot and reduce hallucinations with groundedness checks. You’ll see evaluators in action, and learn how to continuously evaluate your AI applications to ensure the quality of the outputs remains up to par across releases.
Jesse Wellenberg
Software Engineer at Xebia
Utrecht, The Netherlands
Links
Please note that Sessionize is not responsible for the accuracy or validity of the data provided by speakers. If you suspect this profile to be fake or spam, please let us know.
Jump to top