Home
Categories
EXPLORE
True Crime
Comedy
Society & Culture
Sports
History
Education
Business
About Us
Contact Us
Copyright
© 2024 PodJoint
Loading...
0:00 / 0:00
Podjoint Logo
US
Sign in

or

Don't have an account?
Sign up
Forgot password
https://is1-ssl.mzstatic.com/image/thumb/Podcasts221/v4/75/db/da/75dbda7a-9c02-a923-c9a8-ac50c4a94f59/mza_7528638632772517919.jpg/600x600bb.jpg
Data Science Decoded
Mike E
29 episodes
5 days ago
We discuss seminal mathematical papers (sometimes really old 😎 ) that have shaped and established the fields of machine learning and data science as we know them today. The goal of the podcast is to introduce you to the evolution of these fields from a mathematical and slightly philosophical perspective. We will discuss the contribution of these papers, not just from pure a math aspect but also how they influenced the discourse in the field, which areas were opened up as a result, and so on. Our podcast episodes are also available on our youtube: https://youtu.be/wThcXx_vXjQ?si=vnMfs
Show more...
Mathematics
Science
RSS
All content for Data Science Decoded is the property of Mike E and is served directly from their servers with no modification, redirects, or rehosting. The podcast is not affiliated with or endorsed by Podjoint in any way.
We discuss seminal mathematical papers (sometimes really old 😎 ) that have shaped and established the fields of machine learning and data science as we know them today. The goal of the podcast is to introduce you to the evolution of these fields from a mathematical and slightly philosophical perspective. We will discuss the contribution of these papers, not just from pure a math aspect but also how they influenced the discourse in the field, which areas were opened up as a result, and so on. Our podcast episodes are also available on our youtube: https://youtu.be/wThcXx_vXjQ?si=vnMfs
Show more...
Mathematics
Science
https://d3t3ozftmdmh3i.cloudfront.net/staging/podcast_uploaded_nologo/41505637/41505637-1720347263425-80b9b83d77589.jpg
Data Science #19 - The Kullback–Leibler divergence paper (1951)
Data Science Decoded
52 minutes 41 seconds
5 months ago
Data Science #19 - The Kullback–Leibler divergence paper (1951)

In this episode with go over the Kullback-Leibler (KL) divergence paper, "On Information and Sufficiency" (1951). It introduced a measure of the difference between two probability distributions, quantifying the cost of assuming one distribution when another is true.


This concept, rooted in Shannon's information theory (which we reviewed in previous episodes), became fundamental in hypothesis testing, model evaluation, and statistical inference. KL divergence has profoundly impacted data science and AI, forming the basis for techniques like maximum likelihood estimation, Bayesian inference, and generative models such as variational autoencoders (VAEs).


It measures distributional differences, enabling optimization in clustering, density estimation, and natural language processing. In AI, KL divergence ensures models generalize well by aligning training and real-world data distributions. Its role in probabilistic reasoning and adaptive decision-making bridges theoretical information theory and practical machine learning, cementing its relevance in modern technologies.

Data Science Decoded
We discuss seminal mathematical papers (sometimes really old 😎 ) that have shaped and established the fields of machine learning and data science as we know them today. The goal of the podcast is to introduce you to the evolution of these fields from a mathematical and slightly philosophical perspective. We will discuss the contribution of these papers, not just from pure a math aspect but also how they influenced the discourse in the field, which areas were opened up as a result, and so on. Our podcast episodes are also available on our youtube: https://youtu.be/wThcXx_vXjQ?si=vnMfs