Session

Introduction to evaluating generative AI applications

Demos are easy; dependable AI is not. This session shows how to verify that your LLM apps actually work using Azure AI Foundry and Microsoft's agent framework. We’ll cover what to measure and how to spot and reduce hallucinations with groundedness checks. You’ll see evaluators in action, and learn how to continuously evaluate your AI applications to ensure the quality of the outputs remains up to par across releases.

Jesse Wellenberg

Software Engineer at Xebia

Utrecht, The Netherlands

Actions

Please note that Sessionize is not responsible for the accuracy or validity of the data provided by speakers. If you suspect this profile to be fake or spam, please let us know.

Jump to top