Home
Categories
EXPLORE
True Crime
Comedy
Business
Society & Culture
History
Sports
Health & Fitness
About Us
Contact Us
Copyright
© 2024 PodJoint
00:00 / 00:00
Sign in

or

Don't have an account?
Sign up
Forgot password
https://is1-ssl.mzstatic.com/image/thumb/Podcasts126/v4/ca/ae/47/caae47e5-6b4f-4ccc-085a-6104dab5331e/mza_16644954937649900406.jpg/600x600bb.jpg
Weaviate Podcast
Weaviate
131 episodes
5 days ago
Join Connor Shorten as he interviews machine learning experts and explores Weaviate use cases from users and customers.
Show more...
Technology
RSS
All content for Weaviate Podcast is the property of Weaviate and is served directly from their servers with no modification, redirects, or rehosting. The podcast is not affiliated with or endorsed by Podjoint in any way.
Join Connor Shorten as he interviews machine learning experts and explores Weaviate use cases from users and customers.
Show more...
Technology
https://d3t3ozftmdmh3i.cloudfront.net/staging/podcast_uploaded_nologo/34794122/34794122-1704376570442-9264e631e6abf.jpg
Patronus AI with Anand Kannappan - Weaviate Podcast #122!
Weaviate Podcast
1 hour 1 minute 6 seconds
5 months ago
Patronus AI with Anand Kannappan - Weaviate Podcast #122!

AI agents are getting more complex and harder to debug. How do you know what's happening when your agent makes 20+ function calls? What if you have a Multi-Agent System orchestrating several Agents? Anand Kannappan, co-founder of Patronus AI, reveals how their groundbreaking tool Percival transforms agent debugging and evaluation. Percival can instantly analyze complex agent traces, it pinpoints failures across 60 different modes, and it automatically suggests prompt fixes to improve performance. Anand unpacks several of these common failure modes. This includes the critical challenges of "context explosion" where agents process millions of tokens. He also explains domain adaptation for specific use cases, and the complex challenge of multi-agent orchestration. The paradigm of AI Evals is shifting from static evaluation to dynamic oversight! Also learn how Percival's memory architecture leverages both episodic and semantic knowledge with Weaviate!This conversation explores powerful concepts like process vs. outcome rewards and LLM-as-judge approaches. Anand shares his vision for "agentic supervision" where equally capable AI systems provide oversight for complex agent workflows. Whether you're building AI agents, evaluating LLM systems, or interested in how debugging autonomous systems will evolve, this episode delivers concrete techniques. You'll gain philosophical insights on evaluation and a roadmap for how evaluation must transform to keep pace with increasingly autonomous AI systems.

Weaviate Podcast
Join Connor Shorten as he interviews machine learning experts and explores Weaviate use cases from users and customers.