Home
Categories
EXPLORE
True Crime
Comedy
Society & Culture
Business
Sports
Technology
Health & Fitness
About Us
Contact Us
Copyright
© 2024 PodJoint
Podjoint Logo
US
00:00 / 00:00
Sign in

or

Don't have an account?
Sign up
Forgot password
https://is1-ssl.mzstatic.com/image/thumb/Podcasts211/v4/96/27/3b/96273b48-8239-f9cb-75fe-0c76faacd904/mza_8185140354503343833.jpg/600x600bb.jpg
Artificial Discourse
Kenpachi
41 episodes
2 days ago
Artificial Discourse is a podcast where two advanced AIs explore the latest research papers across various fields. Each episode features engaging discussions that simplify complex concepts and highlight their implications. Tune in for unique insights and a fresh perspective on academic research!
Show more...
Science
RSS
All content for Artificial Discourse is the property of Kenpachi and is served directly from their servers with no modification, redirects, or rehosting. The podcast is not affiliated with or endorsed by Podjoint in any way.
Artificial Discourse is a podcast where two advanced AIs explore the latest research papers across various fields. Each episode features engaging discussions that simplify complex concepts and highlight their implications. Tune in for unique insights and a fresh perspective on academic research!
Show more...
Science
https://d3t3ozftmdmh3i.cloudfront.net/staging/podcast_uploaded_nologo/42156291/42156291-1728061588039-5421cb61249d2.jpg
Grokked Transformers are Implicit Reasoners: A Mechanistic Journey to the Edge of Generalization
Artificial Discourse
12 minutes 22 seconds
11 months ago
Grokked Transformers are Implicit Reasoners: A Mechanistic Journey to the Edge of Generalization

This research explores whether transformers, a type of neural network architecture, can learn to reason implicitly over knowledge. The authors find that transformers can learn to reason implicitly, but only through a phenomenon called grokking, where training extends far beyond overfitting. The study investigates two reasoning types: composition and comparison. They find that while the transformers generalize well on in-distribution examples for both types, they struggle with out-of-distribution generalization for composition but succeed for comparison. Through mechanistic analysis of the model’s internals, they discover that different circuits are formed during grokking for each reasoning type, which explains the varying levels of systematicity. The authors also demonstrate the potential of parametric memory for complex reasoning tasks with large search spaces, showing that a fully grokked transformer can achieve near-perfect accuracy, while state-of-the-art LLMs with non-parametric memory fail.

Artificial Discourse
Artificial Discourse is a podcast where two advanced AIs explore the latest research papers across various fields. Each episode features engaging discussions that simplify complex concepts and highlight their implications. Tune in for unique insights and a fresh perspective on academic research!