USING AI & MACHINE LEARNING
IN THE ENTERPRISE

San Francisco + On-Demand | November 7, 2024

Aditya Palnitkar

Software Engineer, Meta

Aditya is a Staff Software Engineer at Meta and has 10+ years of experience working on large scale recommender systems, as well as AI agents.

Watch in-person: November 7

Evals for Supercharging your AI Agents

You wouldn’t dream of deploying software without monitoring or unit testing. However, this is what regularly happens with LLM applications, even though their fickleness and fragility are well known problems. While testing is often seen as a drag on a software team’s productivity, especially in fast moving organisations, the exact opposite is true for LLMs and AI agents. Great monitoring and eval can supercharge your development velocity. A good eval system can help you find the best ROI items to work on, and put together positive feedback loops in your iteration cycle, and help you find the most impactful items to work on- in the next week, month, half or year. Join me as I talk about how to put a world class eval system to work towards your LLM application.