Explorez tous les épisodes de Argmax
Date | Titre | Durée | |
---|---|---|---|
21 Feb 2022 | 1: Reward is Enough | 00:54:36 | |
This is the first episode of Argmax! We talk about our motivations for doing a podcast, and what we hope listeners will get out of it. Highlights of discussion
| |||
07 Mar 2022 | 2: data2vec | 00:53:23 | |
Todays paper: data2vec (https://arxiv.org/abs/2202.03555) Highlights of discussion
| |||
21 Mar 2022 | 3: VICReg | 00:44:46 | |
Todays paper: VICReg (https://arxiv.org/abs/2105.04906)
| |||
06 Apr 2022 | 4: Can Neural Nets Learn the Same Model Twice? | 00:55:23 | |
Todays paper: Can Neural Nets Learn the Same Model Twice? Investigating Reproducibility
| |||
26 Apr 2022 | 5: QMIX | 00:42:06 | |
We talk about QMIX https://arxiv.org/abs/1803.11485 as an example of Deep Multi-agent RL. | |||
06 Jun 2022 | 6: Deep Reinforcement Learning at the Edge of the Statistical Precipice | 01:01:08 | |
We discuss NeurIPS outstanding paper award winning paper, talking about important topics surrounding metrics and reproducibility. | |||
14 Jun 2022 | 7: Deep Unsupervised Learning Using Nonequilibrium Thermodynamics (Diffusion Models) | 00:30:55 | |
We start talking about diffusion models as a technique for generative deep learning. | |||
29 Jul 2022 | 8: GATO (A Generalist Agent) | 00:44:51 | |
Today we talk about GATO, a multi-modal, multi-task, multi-embodiment generalist agent. | |||
29 Jul 2022 | 9: Heads-Up Limit Hold'em Poker Is Solved | 00:47:55 | |
Today we talk about recent AI advances in Poker; specifically the use of counterfactual regret minimization to solve the game of 2-player Limit Texas Hold'em. | |||
23 Aug 2022 | 10: Outracing champion Gran Turismo drivers with deep reinforcement learning | 00:54:50 | |
We discuss Sony AI's accomplishment of creating a novel AI agent that can beat professional racers in Gran Turismo. Some topics include: | |||
30 Sep 2022 | 11: CVPR Workshop on Autonomous Driving Keynote by Ashok Elluswamy, a Tesla engineer | 00:48:51 | |
In this episode we discuss this video: https://youtu.be/jPCV4GKX9Dw | |||
25 Oct 2022 | 12: SIRENs | 00:54:17 | |
In this episode we talked about "Implicit Neural Representations with Periodic Activation Functions" and the strength of periodic non-linearities. | |||
11 Mar 2023 | 13: AlphaTensor | 00:49:05 | |
We talk about AlphaTensor, and how researchers were able to find a new algorithm for matrix multiplication. | |||
17 Mar 2023 | 14: Whisper | 00:49:14 | |
This week we talk about Whisper. It is a weakly supervised speech recognition model. | |||
28 Mar 2023 | 15: InstructGPT | 00:57:27 | |
In this episode we discuss the paper "Training language models to follow instructions with human feedback" by Ouyang et al (2022). We discuss the RLHF paradigm and how important RL is to tuning GPT. | |||
02 Sep 2023 | LoRA | 01:02:56 | |
We talk about Low Rank Approximation for fine tuning Transformers. We are also on YouTube now! Check out the video here: https://youtu.be/lLzHr0VFi3Y | |||
08 Oct 2024 | Mixture of Experts | 00:54:46 | |
In this episode we talk about the paper "Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer" by Noam Shazeer, Azalia Mirhoseini, Krzysztof Maziarz, Andy Davis, Quoc Le, Geoffrey Hinton, Jeff Dean. |