PodcastsTechnologyMachine Learning Street Talk (MLST)

Machine Learning Street Talk (MLST)

Machine Learning Street Talk (MLST)
Machine Learning Street Talk (MLST)
Latest episode

249 episodes

  • Machine Learning Street Talk (MLST)

    When AI Discovers The Next Transformer - Robert Lange (Sakana)

    2026/03/13 | 1h 18 mins.
    Robert Lange, founding researcher at Sakana AI, joins Tim to discuss *Shinka Evolve* — a framework that combines LLMs with evolutionary algorithms to do open-ended program search. The core claim: systems like AlphaEvolve can optimize solutions to fixed problems, but real scientific progress requires co-evolving the problems themselves.

    GTC is coming, the premier AI conference, great opportunity to learn about AI. NVIDIA and partners will showcase breakthroughs in physical AI, AI factories, agentic AI, and inference, exploring the next wave of AI innovation for developers and researchers. Register for virtual GTC for free, using my link and win NVIDIA DGX Spark (https://nvda.ws/4qQ0LMg)

    • Why AlphaEvolve gets stuck — it needs a human to hand it the right problem. Shinka tries to invent new problems automatically, drawing on ideas from POET, PowerPlay, and MAP-Elites quality-diversity search.

    • The *architecture* of Shinka: an archive of programs organized as islands, LLMs used as mutation operators, and a UCB bandit that adaptively selects between frontier models (GPT-5, Sonnet 4.5, Gemini) mid-run. The credit-assignment problem across models turns out to be genuinely hard.

    • Concrete results — state-of-the-art circle packing with dramatically fewer evaluations, second place in an AtCoder competitive programming challenge, evolved load-balancing loss functions for mixture-of-experts models, and agent scaffolds for AIME math benchmarks.

    • Are these systems actually thinking outside the box, or are they parasitic on their starting conditions? When LLMs run autonomously, "nothing interesting happens." Robert pushes back with the stepping-stone argument — evolution doesn't need to extrapolate, just recombine usefully.

    • The AI Scientist question: can automated research pipelines produce real science, or just workshop-level slop that passes surface-level review? Robert is honest that the current version is more co-pilot than autonomous researcher.

    • Where this lands in 5-20 years — Robert's prediction that scientific research will be fundamentally transformed, and Tim's thought experiment about alien mathematical artifacts that no human could have conceived.

    Robert Lange: https://roberttlange.com/

    ---
    TIMESTAMPS:
    00:00:00 Introduction: Robert Lange, Sakana AI and Shinka Evolve
    00:04:15 AlphaEvolve's Blind Spot: Co-Evolving Problems with Solutions
    00:09:05 Unknown Unknowns, POET, and Auto-Curricula for AI Science
    00:14:20 MAP-Elites and Quality-Diversity: Shinka's Evolutionary Architecture
    00:28:00 UCB Bandits, Mutations and the Vibe Research Vision
    00:40:00 Scaling Shinka: Meta-Evolution, Democratisation and the Three-Axis Model
    00:47:10 Applications, ARC-AGI and the Future of Work
    00:57:00 The AI Scientist and the Human Co-Pilot: Who Steers the Search?
    01:06:00 AI Scientist v2, Slop Critique and the Future of Scientific Publishing

    ---
    REFERENCES:
    paper:
    [00:03:30] ShinkaEvolve: Towards Open-Ended And Sample-Efficient Program Evolution
    https://arxiv.org/abs/2509.19349
    [00:04:15] AlphaEvolve: A Coding Agent for Scientific and Algorithmic Discovery
    https://arxiv.org/abs/2506.13131
    [00:06:30] Darwin Godel Machine: Open-Ended Evolution of Self-Improving Agents
    https://arxiv.org/abs/2505.22954
    [00:09:05] Paired Open-Ended Trailblazer (POET)
    https://arxiv.org/abs/1901.01753
    [00:10:00] PowerPlay: Training an Increasingly General Problem Solver by Continually Searching for the Simplest Still Unsolvable Problem
    https://arxiv.org/abs/1112.5309
    [00:10:40] Automated Capability Discovery via Foundation Model Self-Exploration
    https://arxiv.org/abs/2502.07577
    [00:15:30] Illuminating Search Spaces by Mapping Elites (MAP-Elites)
    https://arxiv.org/abs/1504.04909
    [00:47:10] Automated Design of Agentic Systems (ADAS)
    https://arxiv.org/abs/2408.08435
    <trunc, see ReScript/YT>

    PDF : https://app.rescript.info/api/sessions/b8a9dcf60623657c/pdf/download
    Transcript: https://app.rescript.info/public/share/SDOD_3oXOcli3zTqcAtR8eibT5U3gam84oo4KRtI-Vk
  • Machine Learning Street Talk (MLST)

    "Vibe Coding is a Slot Machine" - Jeremy Howard

    2026/03/03 | 1h 26 mins.
    Dive into the realities of AI-assisted coding, the origins of modern fine-tuning, and the cognitive science behind machine learning with fast.ai founder Jeremy Howard. In this episode, we unpack why AI might be turning software engineering into a slot machine and how to maintain true technical intuition in the age of large language models.

    GTC is coming, the premier AI conference, great opportunity to learn about AI. NVIDIA and partners will showcase breakthroughs in physical AI, AI factories, agentic AI, and inference, exploring the next wave of AI innovation for developers and researchers. Register for virtual GTC for free, using my link and win NVIDIA DGX Spark (https://nvda.ws/4qQ0LMg)

    Jeremy Howard is a renowned data scientist, researcher, entrepreneur, and educator. As the co-founder of fast.ai, former President of Kaggle, and the creator of ULMFiT, Jeremy has spent decades democratizing deep learning. His pioneering work laid the foundation for modern transfer learning and the pre-training and fine-tuning paradigm that powers today's language models.

    Key Topics and Main Insights Discussed:

    - The Origins of ULMFiT and Fine-Tuning
    - The Vibe Coding Illusion and Software Engineering
    - Cognitive Science, Friction, and Learning
    - The Future of Developers

    RESCRIPT: https://app.rescript.info/public/share/BhX5zP3b0m63srLOQDKBTFTooSzEMh_ARwmDG_h_izk

    Jeremy Howard:
    https://x.com/jeremyphoward
    https://www.answer.ai/

    ---
    TIMESTAMPS (fixed):
    00:00:00 Introduction & GTC Sponsor
    00:04:30 ULMFiT & The Birth of Fine-Tuning
    00:12:00 Intuition & The Mechanics of Learning
    00:18:30 Abstraction Hierarchies & AI Creativity
    00:23:00 Claude Code & The Interpolation Illusion
    00:27:30 Coding vs. Software Engineering
    00:30:00 Cosplaying Intelligence: Dennett vs. Searle
    00:36:30 Automation, Radiology & Desirable Difficulty
    00:42:30 Organizational Knowledge & The Slope
    00:48:00 Vibe Coding as a Slot Machine
    00:54:00 The Erosion of Control in Software
    01:01:00 Interactive Programming & REPL Environments
    01:05:00 The Notebook Debate & Exploratory Science
    01:17:30 AI Existential Risk & Power Centralization
    01:24:20 Current Risks, Privacy & Enfeeblement

    ---
    REFERENCES:
    Blog Post:
    [00:03:00] fast.ai Blog: Self-Supervised Learning
    https://www.fast.ai/posts/2020-01-13-self_supervised.html
    [00:13:30] DeepMind Blog: Gemini Deep Think
    https://deepmind.google/blog/accelerating-mathematical-and-scientific-discovery-with-gemini-deep-think/
    [00:19:30] Modular Blog: Claude C Compiler analysis
    https://www.modular.com/blog/the-claude-c-compiler-what-it-reveals-about-the-future-of-software
    [00:19:45] Anthropic Engineering Blog: Building C Compiler
    https://www.anthropic.com/engineering/building-c-compiler
    [00:48:00] Cursor Blog: Scaling Agents
    https://cursor.com/blog/scaling-agents
    [01:05:15] fast.ai Blog: NB Dev Merged Driver
    https://www.fast.ai/posts/2022-08-25-jupyter-git.html
    [01:17:30] Jeremy Howard: Response to AI Risk Letter
    https://www.normaltech.ai/p/is-avoiding-extinction-from-ai-really
    Book:
    [00:08:30] M. Chirimuuta: The Brain Abstracted
    https://mitpress.mit.edu/9780262548045/the-brain-abstracted/
    [00:30:00] Daniel Dennett: Consciousness Explained
    https://www.amazon.com/Consciousness-Explained-Daniel-C-Dennett/dp/0316180661
    [00:42:30] Cesar Hidalgo: Infinite Alphabet / Laws of Knowledge
    https://www.amazon.com/Infinite-Alphabet-Laws-Knowledge/dp/0241655676
    Archive Article:
    [00:13:45] MLST Archive: Why Creativity Cannot Be Interpolated
    https://archive.mlst.ai/read/why-creativity-cannot-be-interpolated
    Research Study:
    [00:24:30] METR Study: AI OS Development
    https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/
    Paper:
    [00:24:45] Fred Brooks: No Silver Bullet
    https://www.cs.unc.edu/techreports/86-020.pdf
    [00:30:15] John Searle: Minds, Brains, and Programs
    https://www.cambridge.org/core/journals/behavioral-and-brain-sciences/article/minds-brains-and-programs/DC644B47A4299C637C89772FACC2706A
  • Machine Learning Street Talk (MLST)

    Evolution "Doesn't Need" Mutation - Blaise Agüera y Arcas

    2026/02/16 | 55 mins.
    What if life itself is just a really sophisticated computer program that wrote itself into existence?

    Blaise Agüera y Arcas presenting at ALife 2025 — the most technically detailed public walkthrough of the ideas in his *What is Life?* and *What is Intelligence?* books that we've come across.He covers the BFF experiments (self-replicating programs emerging spontaneously from random noise), the mathematical framework connecting Lotka-Volterra population dynamics with Smoluchowski coagulation, eigenvalue analysis of cooperation matrices, and his central claim that symbiogenesis — not mutation — is the primary engine of evolutionary novelty.The experimental results are genuinely striking: complex self-replicating code arising from random byte strings with zero mutation, a sharp phase transition that looks like gelation, and a proof that blocking deep symbiogenetic ancestry trees prevents the transition entirely.A few things worth flagging for critical viewers:— The substrate is more carefully engineered than the framing sometimes suggests. The choice of language, tape length, interaction protocol, and step limits all shape what emerges. Their own SUBLEQ counterexample (where self-replicators *don't* arise despite being theoretically possible) highlights that these design choices matter substantially — and a general theory of which substrates support this transition is still missing.— The leap from "self-replicating programs on fixed-length tapes" to "life was computational and intelligent from the start" involves significant philosophical extrapolation beyond what the experiments directly demonstrate.— The Bedau et al. (2000) open problems paper he references at the start actually sets a higher bar for Challenge 3.2 than BFF currently meets: it asks that "the internal organization of these 'organisms' and the boundaries separating them from their environment arise and be sustained through the activities of lower-level primitives" — whereas BFF's tape boundaries are fixed by design, not emergent.
    ---
    TIMESTAMPS:
    00:00:00 Introduction: From Noise to Programs & ALife History
    00:03:15 Defining Life: Function as the "Spirit"
    00:05:45 Von Neumann's Insight: Life is Embodied Computation
    00:09:15 Physics of Computation: Irreversibility & Fallacies
    00:15:00 The BFF Experiment: Spontaneous Generation of Code
    00:23:45 The Mystery: Complexity Growth Without Mutation
    00:27:00 Symbiogenesis: The Engine of Novelty
    00:33:15 Mathematical Proof: Blocking Symbiosis Stops Life
    00:40:15 Evolutionary Implications: It's Symbiogenesis All The Way Down
    00:44:30 Intelligence as Modeling Others
    00:46:49 Q&A: Levels of Abstraction & Definitions

    ---
    REFERENCES:
    Paper:
    [00:01:16] Open Problems in Artificial Life
    https://direct.mit.edu/artl/article/6/4/363/2354/Open-Problems-in-Artificial-Life
    [00:09:30] When does a physical system compute?
    https://arxiv.org/abs/1309.7979
    [00:15:00] Computational Life
    https://arxiv.org/abs/2406.19108
    [00:27:30] On the Origin of Mitosing Cells
    https://pubmed.ncbi.nlm.nih.gov/11541392/
    [00:42:00] The Major Evolutionary Transitions
    https://www.nature.com/articles/374227a0
    [00:44:00] The ARC gene
    https://www.nih.gov/news-events/news-releases/memory-gene-goes-viral
    Person:
    [00:05:45] Alan Turing
    https://plato.stanford.edu/entries/turing/
    [00:07:30] John von Neumann
    https://en.wikipedia.org/wiki/John_von_Neumann
    [00:11:15] Hector Zenil
    https://hectorzenil.net/
    [00:12:00] Robert Sapolsky
    https://profiles.stanford.edu/robert-sapolsky

    ---
    LINKS:
    RESCRIPT: https://app.rescript.info/public/share/ff7gb6HpezOR3DF-gr9-rCoMFzzEgUjLQK6voV5XVWY
  • Machine Learning Street Talk (MLST)

    VAEs Are Energy-Based Models? [Dr. Jeff Beck]

    2026/01/25 | 46 mins.
    What makes something truly *intelligent?* Is a rock an agent? Could a perfect simulation of your brain actually *be* you? In this fascinating conversation, Dr. Jeff Beck takes us on a journey through the philosophical and technical foundations of agency, intelligence, and the future of AI.

    Jeff doesn't hold back on the big questions. He argues that from a purely mathematical perspective, there's no structural difference between an agent and a rock – both execute policies that map inputs to outputs. The real distinction lies in *sophistication* – how complex are the internal computations? Does the system engage in planning and counterfactual reasoning, or is it just a lookup table that happens to give the right answers?

    *Key topics explored in this conversation:*

    *The Black Box Problem of Agency* – How can we tell if something is truly planning versus just executing a pre-computed response? Jeff explains why this question is nearly impossible to answer from the outside, and why the best we can do is ask which model gives us the simplest explanation.

    *Energy-Based Models Explained* – A masterclass on how EBMs differ from standard neural networks. The key insight: traditional networks only optimize weights, while energy-based models optimize *both* weights and internal states – a subtle but profound distinction that connects to Bayesian inference.

    *Why Your Brain Might Have Evolved from Your Nose* – One of the most surprising moments in the conversation. Jeff proposes that the complex, non-smooth nature of olfactory space may have driven the evolution of our associative cortex and planning abilities.

    *The JEPA Revolution* – A deep dive into Yann LeCun's Joint Embedding Prediction Architecture and why learning in latent space (rather than predicting every pixel) might be the key to more robust AI representations.

    *AI Safety Without Skynet Fears* – Jeff takes a refreshingly grounded stance on AI risk. He's less worried about rogue superintelligences and more concerned about humans becoming "reward function selectors" – couch potatoes who just approve or reject AI outputs. His proposed solution? Use inverse reinforcement learning to derive AI goals from observed human behavior, then make *small* perturbations rather than naive commands like "end world hunger."

    Whether you're interested in the philosophy of mind, the technical details of modern machine learning, or just want to understand what makes intelligence *tick,* this conversation delivers insights you won't find anywhere else.

    ---
    TIMESTAMPS:
    00:00:00 Geometric Deep Learning & Physical Symmetries
    00:00:56 Defining Agency: From Rocks to Planning
    00:05:25 The Black Box Problem & Counterfactuals
    00:08:45 Simulated Agency vs. Physical Reality
    00:12:55 Energy-Based Models & Test-Time Training
    00:17:30 Bayesian Inference & Free Energy
    00:20:07 JEPA, Latent Space, & Non-Contrastive Learning
    00:27:07 Evolution of Intelligence & Modular Brains
    00:34:00 Scientific Discovery & Automated Experimentation
    00:38:04 AI Safety, Enfeeblement & The Future of Work

    ---
    REFERENCES:
    Concept:
    [00:00:58] Free Energy Principle (FEP)
    https://en.wikipedia.org/wiki/Free_energy_principle
    [00:06:00] Monte Carlo Tree Search
    https://en.wikipedia.org/wiki/Monte_Carlo_tree_search
    Book:
    [00:09:00] The Intentional Stance
    https://mitpress.mit.edu/9780262540537/the-intentional-stance/
    Paper:
    [00:13:00] A Tutorial on Energy-Based Learning (LeCun 2006)
    http://yann.lecun.com/exdb/publis/pdf/lecun-06.pdf
    [00:15:00] Auto-Encoding Variational Bayes (VAE)
    https://arxiv.org/abs/1312.6114
    [00:20:15] JEPA (Joint Embedding Prediction Architecture)
    https://openreview.net/forum?id=BZ5a1r-kVsf
    [00:22:30] The Wake-Sleep Algorithm
    https://www.cs.toronto.edu/~hinton/absps/ws.pdf
    <trunc, see rescript>

    ---
    RESCRIPT:
    https://app.rescript.info/public/share/DJlSbJ_Qx080q315tWaqMWn3PixCQsOcM4Kf1IW9_Eo
    PDF:
    https://app.rescript.info/api/public/sessions/0efec296b9b6e905/pdf
  • Machine Learning Street Talk (MLST)

    Abstraction & Idealization: AI's Plato Problem [Mazviita Chirimuuta]

    2026/01/23 | 53 mins.
    Professor Mazviita Chirimuuta joins us for a fascinating deep dive into the philosophy of neuroscience and what it really means to understand the mind.*What can neuroscience actually tell us about how the mind works?* In this thought-provoking conversation, we explore the hidden assumptions behind computational theories of the brain, the limits of scientific abstraction, and why the question of machine consciousness might be more complicated than AI researchers assume.Mazviita, author of *The Brain Abstracted,* brings a unique perspective shaped by her background in both neuroscience research and philosophy. She challenges us to think critically about the metaphors we use to understand cognition — from the reflex theory of the late 19th century to today's dominant view of the brain as a computer.*Key topics explored:**The problem of oversimplification* — Why scientific models necessarily leave things out, and how this can sometimes lead entire fields astray. The cautionary tale of reflex theory shows how elegant explanations can blind us to biological complexity.*Is the brain really a computer?* — Mazviita unpacks the philosophical assumptions behind computational neuroscience and asks: if we can model anything computationally, what makes brains special? The answer might challenge everything you thought you knew about AI.*Haptic realism* — A fresh way of thinking about scientific knowledge that emphasizes interaction over passive observation. Knowledge isn't about reading the "source code of the universe" — it's something we actively construct through engagement with the world.*Why embodiment matters for understanding* — Can a disembodied language model truly understand? Mazviita makes a compelling case that human cognition is deeply entangled with our sensory-motor engagement and biological existence in ways that can't simply be abstracted away.*Technology and human finitude* — Drawing on Heidegger, we discuss how the dream of transcending our physical limitations through technology might reflect a fundamental misunderstanding of what it means to be a knower.This conversation is essential viewing for anyone interested in AI, consciousness, philosophy of mind, or the future of cognitive science. Whether you're skeptical of strong AI claims or a true believer in machine consciousness, Mazviita's careful philosophical analysis will give you new tools for thinking through these profound questions.---TIMESTAMPS:00:00:00 The Problem of Generalizing Neuroscience00:02:51 Abstraction vs. Idealization: The "Kaleidoscope"00:05:39 Platonism in AI: Discovering or Inventing Patterns?00:09:42 When Simplification Fails: The Reflex Theory00:12:23 Behaviorism and the "Black Box" Trap00:14:20 Haptic Realism: Knowledge Through Interaction00:20:23 Is Nature Protean? The Myth of Converging Truth00:23:23 The Computational Theory of Mind: A Useful Fiction?00:27:25 Biological Constraints: Why Brains Aren't Just Neural Nets00:31:01 Agency, Distal Causes, and Dennett's Stances00:37:13 Searle's Challenge: Causal Powers and Understanding00:41:58 Heidegger's Warning & The Experiment on Children---REFERENCES:Book:[00:01:28] The Brain Abstractedhttps://mitpress.mit.edu/9780262548045/the-brain-abstracted/[00:11:05] The Integrated Action of the Nervous Systemhttps://www.amazon.sg/integrative-action-nervous-system/dp/9354179029[00:18:15] The Quest for Certainty (Dewey)https://www.amazon.com/Quest-Certainty-Relation-Knowledge-Lectures/dp/0399501916[00:19:45] Realism for Realistic People (Chang)https://www.cambridge.org/core/books/realism-for-realistic-people/ACC93A7F03B15AA4D6F3A466E3FC5AB7<truncated, see ReScript>---RESCRIPT:https://app.rescript.info/public/share/A6cZ1TY35p8ORMmYCWNBI0no9ChU3-Kx7dPXGJURvZ0PDF Transcript:https://app.rescript.info/api/public/sessions/0fb7767e066cf712/pdf

More Technology podcasts

About Machine Learning Street Talk (MLST)

Welcome! We engage in fascinating discussions with pre-eminent figures in the AI field. Our flagship show covers current affairs in AI, cognitive science, neuroscience and philosophy of mind with in-depth analysis. Our approach is unrivalled in terms of scope and rigour – we believe in intellectual diversity in AI, and we touch on all of the main ideas in the field with the hype surgically removed. MLST is run by Tim Scarfe, Ph.D (https://www.linkedin.com/in/ecsquizor/) and features regular appearances from MIT Doctor of Philosophy Keith Duggar (https://www.linkedin.com/in/dr-keith-duggar/).
Podcast website

Listen to Machine Learning Street Talk (MLST), The Sixth Bureau and many other podcasts from around the world with the radio.net app

Get the free radio.net app

  • Stations and podcasts to bookmark
  • Stream via Wi-Fi or Bluetooth
  • Supports Carplay & Android Auto
  • Many other app features
Social
v8.7.2 | © 2007-2026 radio.de GmbH
Generated: 3/14/2026 - 1:27:23 PM