Home
Categories
EXPLORE
Comedy
Society & Culture
True Crime
Business
Fiction
History
Science
About Us
Contact Us
Copyright
© 2024 PodJoint
Loading...
0:00 / 0:00
Podjoint Logo
MX
Sign in

or

Don't have an account?
Sign up
Forgot password
https://is1-ssl.mzstatic.com/image/thumb/Podcasts211/v4/cb/94/dd/cb94dd8c-3061-a5fc-079f-a47a75bb1a24/mza_8328803272468307441.jpg/600x600bb.jpg
Argmax
Vahe Hagopian, Taka Hasegawa, Farrukh Rahman
17 episodes
6 months ago
In this episode we talk about the paper "Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer" by Noam Shazeer, Azalia Mirhoseini, Krzysztof Maziarz, Andy Davis, Quoc Le, Geoffrey Hinton, Jeff Dean.
Show more...
Mathematics
Science
RSS
All content for Argmax is the property of Vahe Hagopian, Taka Hasegawa, Farrukh Rahman and is served directly from their servers with no modification, redirects, or rehosting. The podcast is not affiliated with or endorsed by Podjoint in any way.
In this episode we talk about the paper "Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer" by Noam Shazeer, Azalia Mirhoseini, Krzysztof Maziarz, Andy Davis, Quoc Le, Geoffrey Hinton, Jeff Dean.
Show more...
Mathematics
Science
Episodes (17/17)
Argmax
Mixture of Experts
In this episode we talk about the paper "Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer" by Noam Shazeer, Azalia Mirhoseini, Krzysztof Maziarz, Andy Davis, Quoc Le, Geoffrey Hinton, Jeff Dean.
Show more...
9 months ago
54 minutes

Argmax
LoRA
We talk about Low Rank Approximation for fine tuning Transformers. We are also on YouTube now! Check out the video here: https://youtu.be/lLzHr0VFi3Y
Show more...
1 year ago
1 hour 2 minutes

Argmax
15: InstructGPT
In this episode we discuss the paper "Training language models to follow instructions with human feedback" by Ouyang et al (2022). We discuss the RLHF paradigm and how important RL is to tuning GPT.
Show more...
2 years ago
57 minutes

Argmax
14: Whisper
This week we talk about Whisper. It is a weakly supervised speech recognition model.
Show more...
2 years ago
49 minutes

Argmax
13: AlphaTensor
We talk about AlphaTensor, and how researchers were able to find a new algorithm for matrix multiplication.
Show more...
2 years ago
49 minutes

Argmax
12: SIRENs
In this episode we talked about "Implicit Neural Representations with Periodic Activation Functions" and the strength of periodic non-linearities.
Show more...
2 years ago
54 minutes

Argmax
11: CVPR Workshop on Autonomous Driving Keynote by Ashok Elluswamy, a Tesla engineer
In this episode we discuss this video: https://youtu.be/jPCV4GKX9DwHow Tesla approaches collision detection with novel methods.
Show more...
2 years ago
48 minutes

Argmax
10: Outracing champion Gran Turismo drivers with deep reinforcement learning
We discuss Sony AI's accomplishment of creating a novel AI agent that can beat professional racers in Gran Turismo. Some topics include:- The crafting of rewards to make the agent behave nicely- What is QR-SAC?- How to deal with "rare" experiences in the replay bufferLink to paper: https://www.nature.com/articles/s41586-021-04357-7
Show more...
2 years ago
54 minutes

Argmax
8: GATO (A Generalist Agent)
Today we talk about GATO, a multi-modal, multi-task, multi-embodiment generalist agent.
Show more...
2 years ago
44 minutes

Argmax
9: Heads-Up Limit Hold'em Poker Is Solved
Today we talk about recent AI advances in Poker; specifically the use of counterfactual regret minimization to solve the game of 2-player Limit Texas Hold'em.
Show more...
2 years ago
47 minutes

Argmax
7: Deep Unsupervised Learning Using Nonequilibrium Thermodynamics (Diffusion Models)
We start talking about diffusion models as a technique for generative deep learning.
Show more...
3 years ago
30 minutes

Argmax
6: Deep Reinforcement Learning at the Edge of the Statistical Precipice
We discuss NeurIPS outstanding paper award winning paper, talking about important topics surrounding metrics and reproducibility.
Show more...
3 years ago
1 hour 1 minute

Argmax
5: QMIX
We talk about QMIX https://arxiv.org/abs/1803.11485 as an example of Deep Multi-agent RL.
Show more...
3 years ago
42 minutes

Argmax
4: Can Neural Nets Learn the Same Model Twice?
Todays paper: Can Neural Nets Learn the Same Model Twice? Investigating Reproducibilityand Double Descent from the Decision Boundary Perspective (https://arxiv.org/pdf/2203.08124.pdf)Summary:A discussion of reproducibility and double descent through visualizations of decision boundaries.Highlights of the discussion:Relationship between model performance and reproducibilityWhich models are robust and reproducibleHow they calculate the various scores
Show more...
3 years ago
55 minutes

Argmax
3: VICReg
Todays paper: VICReg (https://arxiv.org/abs/2105.04906)Summary of the paperVICReg prevents representation collapse using a mixture of variance, invariance and covariance when calculating the loss. It does not require negative samples and achieves great performance on downstream tasks.Highlights of discussionThe VICReg architecture (Figure 1)Sensitivity to hyperparameters (Table 7)Top 5 metric usefulness
Show more...
3 years ago
44 minutes

Argmax
2: data2vec
Todays paper: data2vec (https://arxiv.org/abs/2202.03555)Summary of the paperA multimodal SSL algorithm that predicts latent representation of different types of input.Highlights of discussionWhat are the motivations of SSL and multimodalHow does the student teacher learning work?What are similarities and differences between ViT, BYOL, and Reinforcement Learning algorithms.
Show more...
3 years ago
53 minutes

Argmax
1: Reward is Enough
This is the first episode of Argmax! We talk about our motivations for doing a podcast, and what we hope listeners will get out of it.Todays paper: Reward is Enough Summary of the paperThe authors present the Reward is Enough hypothesis: Intelligence, and its associated abilities, can be understood as subserving the maximisation of reward by an agent acting in its environment.Highlights of discussionHigh level overview of Reinforcement LearningHow evolution can be encoded as a reward maximiza...
Show more...
3 years ago
54 minutes

Argmax
In this episode we talk about the paper "Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer" by Noam Shazeer, Azalia Mirhoseini, Krzysztof Maziarz, Andy Davis, Quoc Le, Geoffrey Hinton, Jeff Dean.