Home
Categories
EXPLORE
True Crime
Comedy
Society & Culture
Business
Sports
History
Fiction
About Us
Contact Us
Copyright
© 2024 PodJoint
00:00 / 00:00
Sign in

or

Don't have an account?
Sign up
Forgot password
https://is1-ssl.mzstatic.com/image/thumb/Podcasts221/v4/93/44/10/934410f5-e2a5-d5e6-6161-ebe34cf7dff3/mza_16724237815679329034.jpg/600x600bb.jpg
The Chief AI Officer Show
Front Lines
31 episodes
1 week ago
The Chief AI Officer Show bridges the gap between enterprise buyers and AI innovators. Through candid conversations with leading Chief AI Officers and startup founders, we unpack the real stories behind AI deployment and sales. Get practical insights from those pioneering AI adoption and building tomorrow’s breakthrough solutions.
Show more...
Technology
RSS
All content for The Chief AI Officer Show is the property of Front Lines and is served directly from their servers with no modification, redirects, or rehosting. The podcast is not affiliated with or endorsed by Podjoint in any way.
The Chief AI Officer Show bridges the gap between enterprise buyers and AI innovators. Through candid conversations with leading Chief AI Officers and startup founders, we unpack the real stories behind AI deployment and sales. Get practical insights from those pioneering AI adoption and building tomorrow’s breakthrough solutions.
Show more...
Technology
https://pbcdn1.podbean.com/imglogo/ep-logo/pbblog18558349/CAIOPod-Epi_Dark.png
Gentrace’s Doug Safreno on Escaping POC Purgatory with Collaborative AI Evaluation
The Chief AI Officer Show
42 minutes 33 seconds
6 months ago
Gentrace’s Doug Safreno on Escaping POC Purgatory with Collaborative AI Evaluation
The reliability gap between AI models and production-ready applications is where countless enterprise initiatives die in POC purgatory. In this episode of Chief AI Officer, Doug Safreno, Co-founder & CEO of Gentrace, offers the testing infrastructure that helped customers escape the Whac-A-Mole cycle plaguing AI development. Having experienced this firsthand when building an email assistant with GPT-3 in late 2022, Doug explains why traditional evaluation methods fail with generative AI, where outputs can be wrong in countless ways beyond simple classification errors. With Gentrace positioned as a ”collaborative LLM testing environment” rather than just a visualization layer, Doug shares how they’ve transformed companies from isolated engineering testing to cross-functional evaluation that increased velocity 40x and enabled successful production launches. His insights from running monthly dinners with bleeding-edge AI engineers reveal how the industry conversation has evolved from basic product questions to sophisticated technical challenges with retrieval and agentic workflows. Topics discussed: - Why asking LLMs to grade their own outputs creates circular testing failures, and how giving evaluator models access to reference data or expected outcomes the generating model never saw leads to meaningful quality assessment. - How Gentrace’s platform enables subject matter experts, product managers, and educators to contribute to evaluation without coding, increasing test velocity by 40x. - Why aiming for 100% accuracy is often a red flag, and how to determine the right threshold based on recoverability of errors, stakes of the application, and business model considerations. - Testing strategies for multi-step processes where the final output might be an edit to a document rather than text, requiring inspection of entire traces and intermediate decision points. - How engineering discussions have shifted from basic form factor questions (chatbot vs. autocomplete) to specific technical challenges in implementing retrieval with LLMs and agentic workflows. - How converting user feedback on problematic outputs into automated test criteria creates continuous improvement loops without requiring engineering resources. - Using monthly dinners with 10-20 bleeding-edge AI engineers and broader events with 100+ attendees to create learning communities that generate leads while solving real problems. - Why 2024 was about getting basic evaluation in place, while 2025 will expose the limitations of simplistic frameworks that don’t use ”unfair advantages” or collaborative approaches. - How to frame AI reliability differently from traditional software while still providing governance, transparency, and trust across organizations. - Signs a company is ready for advanced evaluation infrastructure: when playing Whac-A-Mole with fixes, when product managers easily break AI systems despite engineering evals, and when lack of organizational trust is blocking deployment.
The Chief AI Officer Show
The Chief AI Officer Show bridges the gap between enterprise buyers and AI innovators. Through candid conversations with leading Chief AI Officers and startup founders, we unpack the real stories behind AI deployment and sales. Get practical insights from those pioneering AI adoption and building tomorrow’s breakthrough solutions.