Home
Categories
EXPLORE
True Crime
Comedy
Business
Society & Culture
History
Sports
Health & Fitness
About Us
Contact Us
Copyright
© 2024 PodJoint
00:00 / 00:00
Sign in

or

Don't have an account?
Sign up
Forgot password
https://is1-ssl.mzstatic.com/image/thumb/Podcasts211/v4/28/0d/20/280d209a-5b07-9e20-7d7a-c67d4f4e957a/mza_17291545061204122681.jpg/600x600bb.jpg
Talking Machines by SU PARK
Su Park
9 episodes
4 days ago
Join Su Park as she invites various guests to unpack the hottest Artificial Intelligence papers off the press. Each episode dives into the newest discoveries in AI and the sci-fi-slowly-becoming-our-reality era we’re living in.
Show more...
Education
RSS
All content for Talking Machines by SU PARK is the property of Su Park and is served directly from their servers with no modification, redirects, or rehosting. The podcast is not affiliated with or endorsed by Podjoint in any way.
Join Su Park as she invites various guests to unpack the hottest Artificial Intelligence papers off the press. Each episode dives into the newest discoveries in AI and the sci-fi-slowly-becoming-our-reality era we’re living in.
Show more...
Education
Episodes (9/9)
Talking Machines by SU PARK
LLM as a Judge: Evaluating AI with AI

In this episode of "Talking Machines by Su Park," we explore the fascinating concept of "LLM-as-a-Judge," which evaluates the role of large language models in providing scalable assessments across various domains. As AI continues to evolve, understanding how these models can bridge the gap between human insight and algorithmic efficiency becomes increasingly significant. The discussion highlights the growing trend of utilizing LLMs not only to evaluate other AI systems but also to enhance the evaluation process itself, bringing consistency to an area that often suffers from human bias and variability.


Key insights from the conversation include the potential for LLMs to merge the strengths of expert evaluations with the speed and scalability of automated assessments. The episode further delves into the challenges of implementing reliable LLM-as-a-Judge systems, emphasizing the need to address biases and ensure consistent evaluations. These insights underscore the implications of integrating LLMs into evaluation processes, paving the way for more effective and nuanced assessments in the future.


"A Survey on LLM-as-a-Judge": https://arxiv.org/abs/2411.15594

Show more...
6 months ago
19 minutes 32 seconds

Talking Machines by SU PARK
How to Pick the Best Pretraining Data

In this episode of "Talking Machines by Su Park," the hosts explore the critical topic of selecting pretraining datasets for Large Language Models, a decision that significantly impacts model performance and cost-efficiency. The discussion centers on a recent paper from the Allen Institute for AI, which introduces a novel approach to optimizing dataset selection without extensive computational resources, thereby addressing a key challenge in AI research.


The episode highlights two major insights from the paper. First, the proposed suite of models, known as DATADECIDE, allows researchers to effectively predict which datasets will yield the best results for larger models based on smaller-scale experiments. This method has been shown to achieve approximately 80% accuracy in predicting performance outcomes, thus reducing the need for costly trial-and-error approaches. Additionally, the research reveals which benchmarks correlate with high performance, offering valuable guidance for future dataset selection in AI training.


"DataDecide: How to Predict Best Pretraining Data with Small Experiments" by Allen Institute for AI: https://arxiv.org/abs/2504.11393

Show more...
6 months ago
17 minutes 30 seconds

Talking Machines by SU PARK
How AI Learns Mid-Conversation

In this episode of "Talking Machines by Su Park," the discussion centers on the innovative concept of the Dynamic Cheatsheet (DC) for language models. This framework enhances the memory capabilities of AI systems during inference, enabling them to retain and apply insights from previous interactions. The significance of this development lies in its potential to transform how language models operate, moving away from treating each query as a standalone task to a more integrated approach that can lead to improved efficiency and problem-solving capabilities.

Key insights from the conversation include the remarkable performance improvements observed with the implementation of DC. For instance, the accuracy of Claude 3.5 Sonnet in algebraic tasks more than doubled as it retained relevant insights, while GPT-4o's success rate on the Game of 24 puzzle soared from 10% to 99% after leveraging a reusable Python-based solution. This episode highlights how effective memory structuring in AI can enhance its ability to tackle similar challenges, akin to having a toolbox of solutions readily available for diverse problems.

Dynamic Cheatsheet: Test-Time Learning with Adaptive Memory: https://arxiv.org/abs/2504.07952

Show more...
6 months ago
17 minutes 28 seconds

Talking Machines by SU PARK
Alone Together: The Emotional Cost of Chatting with AI

In this episode of "Talking Machines by Su Park," the focus is on the emotional impact of chatbot interactions on mental health and social dynamics. The significance of this discussion lies in understanding how technology, specifically AI-driven chatbots, can influence feelings of loneliness and social connection in users, a topic that has become increasingly relevant in our digitally connected yet often isolated lives.

Key insights from the episode reveal that while chatbots can initially alleviate feelings of loneliness, excessive interaction—particularly with voice-based bots—can paradoxically lead to heightened loneliness and emotional dependence. The researchers conducted a comprehensive four-week study involving nearly a thousand participants, analyzing over 300,000 messages to assess how different types of conversations, especially personal versus non-personal topics, affect psychosocial outcomes. This nuanced understanding underscores the complex relationship between human emotions and AI interactions.

How AI and Human Behaviors Shape Psychosocial Effects of Chatbot Use: A Longitudinal Randomized Controlled Study by MIT Media Lab & OpenAI: https://www.media.mit.edu/publications/how-ai-and-human-behaviors-shape-psychosocial-effects-of-chatbot-use-a-longitudinal-controlled-study/

Show more...
7 months ago
19 minutes 30 seconds

Talking Machines by SU PARK
Tom, Jerry, and the Neural Net: AI’s Leap in Video Storytelling

In this episode of "Talking Machines by Su Park," the hosts explore a groundbreaking paper focused on generating one-minute videos using a novel approach called Test-Time Training (TTT) layers. This topic is significant as it addresses the limitations of current video generation models, which typically produce only short clips, often around 20 seconds. By leveraging TTT layers, the researchers aim to enhance both the length and narrative complexity of generated videos, showcasing their method through the engaging context of Tom and Jerry cartoons.


Key insights from the discussion include the innovative use of TTT layers to make hidden states more expressive, effectively allowing the model to function like a neural network at critical moments. This enhancement leads to a notable improvement in the coherence of the generated stories, with the researchers reporting a 34% performance boost over existing models. The implications of this work suggest a more advanced capability for AI in video generation, paving the way for richer and more complex visual storytelling.


One-Minute Video Generation with Test-Time Training by NVIDIA: https://arxiv.org/abs/2504.05298

Show more...
7 months ago
23 minutes 10 seconds

Talking Machines by SU PARK
How AI Learns to Self-Reflect

In this episode of "Talking Machines by Su Park," the discussion focuses on groundbreaking research that reveals AI models begin developing self-correction abilities earlier than previously thought. This insight challenges the established notion that reflective reasoning in AI is solely a product of the reinforcement learning phase, highlighting the importance of pre-training in the development of these capabilities.

Key findings from the paper indicate that AI models can recognize and correct their own reasoning errors during pre-training, suggesting that self-reflective learning starts much earlier. As the training progresses, these models not only enhance their self-correction skills but also demonstrate improved reflective reasoning across various domains, including mathematics, coding, and logic. This suggests a paradigm shift in understanding how AI learns and evolves its reasoning processes.

Rethinking Reflection in Pre-Training by Essential AI: https://arxiv.org/abs/2504.04022

Show more...
7 months ago
12 minutes 17 seconds

Talking Machines by SU PARK
Decoding AI: Inside Claude 3.5

In this episode of "Talking Machines by SU PARK," the hosts explore the intricate workings of Claude 3.5, a large language model developed by Anthropic. The discussion centers on Anthropic's new paper titled "On the Biology of a Large Language Model," which seeks to slice and dice the complex internal mechanisms of these AI systems. Understanding how these models function is crucial, as they are increasingly integrated into various applications, yet often operate as black boxes to users and researchers alike.

Key insights from the conversation include the use of circuit tracing methodology to map interactions within the model, akin to biological research methods. The authors of the paper create attribution graphs to visualize feature interactions and their contributions to outputs, effectively providing a roadmap for understanding these AI systems. This approach not only enhances our understanding of large language models but also has implications for improving their design and deployment in real-world scenarios.

On the Biology of a Large Language Model: https://transformer-circuits.pub/2025/attribution-graphs/biology.html

Show more...
7 months ago
18 minutes

Talking Machines by SU PARK
Can AI Turn Random Ideas Into Music?

In this episode, Alex and Vic dive into the madness of using AI to turn their weirdest, most chaotic ideas into actual songs. They introduce Amuse—an AI tool that takes anything from photos and text to random melodies and spits out chords that match the vibe. 

"Amuse: Human-AI Collaborative Songwriting with Multimodal Inspirations": https://arxiv.org/abs/2412.18940

Show more...
7 months ago
5 minutes 39 seconds

Talking Machines by SU PARK
AI Agents Are Writing Research Papers—And Reading Each Other’s Too?

In this episode of Talking Machines, Vic and Alex dive into AgentRxiv, a new platform where autonomous AI agents collaborate, share papers, and build on each other's work—just like human scientists do.

They break down the paper “AgentRxiv: Towards Collaborative Autonomous Research” by Samuel Schmidgall and Michael Moor, exploring how AI research labs made up of digital scientists can now exchange ideas and accelerate discovery. From major gains on the MATH-500 benchmark to cross-domain reasoning improvements, this paper brings us one step closer to AI agents doing real scientific R&D.

Plus: Vic's still recovering from the Season 2 finale of Severance , and Alex reminds us fiction might slowly be becoming our new reality.

Show more...
7 months ago
8 minutes 44 seconds

Talking Machines by SU PARK
Join Su Park as she invites various guests to unpack the hottest Artificial Intelligence papers off the press. Each episode dives into the newest discoveries in AI and the sci-fi-slowly-becoming-our-reality era we’re living in.