logo
episode-header-image
Apr 2025
51m 45s

Teaching LLMs to Self-Reflect with Reinf...

Sam Charrington
About this episode
Today, we're joined by Maohao Shen, PhD student at MIT to discuss his paper, “Satori: Reinforcement Learning with Chain-of-Action-Thought Enhances LLM Reasoning via Autoregressive Search.” We dig into how Satori leverages reinforcement learning to improve language model reasoning—enabling model self-reflection, self-correction, and exploration of alternative ... Show More
Up next
Oct 7
Recurrence and Attention for Long-Context Transformers with Jacob Buckman - #750
Today, we're joined by Jacob Buckman, co-founder and CEO of Manifest AI to discuss achieving long context in transformers. We discuss the bottlenecks of scaling context length and recent techniques to overcome them, including windowed attention, grouped query attention, and laten ... Show More
57m 23s
Sep 30
The Decentralized Future of Private AI with Illia Polosukhin - #749
In this episode, Illia Polosukhin, a co-author of the seminal "Attention Is All You Need" paper and co-founder of Near AI, joins us to discuss his vision for building private, decentralized, and user-owned AI. Illia shares his unique journey from developing the Transformer archit ... Show More
1h 5m
Sep 23
Inside Nano Banana 🍌 and the Future of Vision-Language Models with Oliver Wang - #748
Today, we’re joined by Oliver Wang, principal scientist at Google DeepMind and tech lead for Gemini 2.5 Flash Image—better known by its code name, “Nano Banana.” We dive into the development and capabilities of this newly released frontier vision-language model, beginning with th ... Show More
1h 3m
Recommended Episodes
Aug 2023
Cuttlefish Model Tuning
Hongyi Wang, a Senior Researcher at the Machine Learning Department at Carnegie Mellon University, joins us. His research is in the intersection of systems and machine learning. He discussed his research paper, Cuttlefish: Low-Rank Model Training without All the Tuning, on today’ ... Show More
27m 8s
Apr 2022
How The Pros Train (And What We Can Learn From It)
From top level cyclists to world-class runners and speedskaters, pro training regimes offer a fascinating look into what it takes to count yourself among the world's best. Recent research among top-level athletes is also leading us to question entrenched training methods as sport ... Show More
54m 53s
Oct 2024
Dr. Checcucci and Ass. Prof. Puliatti discuss new tech and standardisation trends in surgical training
In 'Episode 3' of the series "Voices of tomorrow's urologists (ESRU): Journeying through Europe's resident perspectives", Dr. Enrico Checcucci (IT) and Assoc. Prof. Stefano Puliatti (IT) discuss "Surgical training unveiled: Exploring new tech and standardisati ... Show More
15m 8s
Nov 2024
How to Improve at Learning Using Neuroscience & AI | Dr. Terry Sejnowski
In this episode, my guest is Dr. Terry Sejnowski, Ph.D., professor of computational neurobiology at the Salk Institute for Biological Studies. He is world-renowned for exploring how our brain processes and stores information and, with that understanding, for developing tools that ... Show More
2h 34m
Jan 2025
Erik Bernhardsson on Creating Tools That Make AI Feel Effortless
Today on No Priors, Elad chats with Erik Bernhardsson, founder and CEO of Modal Labs, a platform simplifying ML workflows by providing a serverless infrastructure designed to streamline deployment, scaling, and development for AI engineers. Erik talks about his early work on Spot ... Show More
23m 36s
Nov 2024
AI and the Future of Math, with DeepMind’s AlphaProof Team
In this week’s episode of No Priors, Sarah and Elad sit down with the Google DeepMind team behind AlphaProof, a new reinforcement learning-based system for formal math reasoning that recently reached a silver-medal standard in solving International Mathematical Olympiad problems. ... Show More
39m 21s
Feb 2025
Grok 3: The New AI Challenger
In this episode, Jaeden discusses the launch of Grok 3, the latest AI model from X AI, highlighting its capabilities, training methods, and performance benchmarks compared to competitors like OpenAI's ChatGPT. He shares personal experiences using Grok 3, including its reasoni ... Show More
16m 45s
Feb 2025
LLMs and Graphs Synergy
In this episode, Garima Agrawal, a senior researcher and AI consultant, brings her years of experience in data science and artificial intelligence. Listeners will learn about the evolving role of knowledge graphs in augmenting large language models (LLMs) for domain-specific task ... Show More
34m 47s
Jun 2022
Focus – Learning Language
How do we learn to talk? When do babies first understand language? And how do scientists study language learning in the lab? In this Focus episode of How We’re Wired, join producer Dr Eva Higginbotham as she visits the Oxford Brookes Babylab to understand how babies make sense of ... Show More
19m 55s