One of the biggest barriers enterprises face in scaling generative and agentic AI is ensuring accuracy, reliability, and trustworthiness at every stage of the lifecycle. At the root of this is often a lack of understanding and oversight of the data used.
This session explores how human-in-the-loop evaluation bridges the gap between a proof of concept and a production-grade AI system. We’ll explore how leading companies design evaluation programs that combine a deep understanding of data, human expertise, and automation to improve performance and maintain trust.
You’ll learn:
Check out the incredible speaker line-up to see who will be joining Leo.
Download The Latest Agenda