Powered by RND
PodcastsTecnologiaThe Embodied AI Podcast

The Embodied AI Podcast

Akseli Ilmanen
The Embodied AI Podcast
Último episódio

Episódios Disponíveis

5 de 9
  • Second podcast on Brain, Space and Time!
    Hi! I started a second podcast (the Brain Space Time Podcast), available here! Listen to this episode to find about it! Timestamps: (00:00) - What is the Brain Space Time Podcast about? (01:39) - The podcast logo explained. (05:28) - Getting in touch. Links Henri Bergson's 1986 Matter and memory PDF (Cone figure on p. 61) Uri Hasson on temporal receptive windows paper Follow me For updates on new episode releases, follow me on Twitter. I welcome your comments, questions, and suggestions. Feel free to email me at [email protected] If you are interested in my other work, click here to look at my blog, website, or (ongoing) Bachelor dissertation on time perception semantic networks.
    --------  
    5:47
  • #7 Tony Zador: The Embodied Turing Test, Genomic Bottlenecks, Molecular Connectomics
    Tony has a lab at Cold Spring Harbour, New York. Using rodents, his lab studies the neural circuits underlying auditory decisions. He is also developing new technologies for connectome sequencing and does some NeuroAI work. In the episode, after a detour on language and the Costa Rican singing mouse, we discuss his recent paper on 'The Embodied Turing Test' and Moravec's paradox; the idea that what we find hard is easy for AI, and vice versa. We explore how Tony's work in creating a rodent decision-making model might inform a virtual platform for embodied animal-like agents. Evolution is an underlying thread in the discussion, including his work on the genomic bottleneck, which might 'be a feature, not a bug'. We discuss how Tony is revolutionizing connectomics using molecular sequencing and why people in AI should care about connectomics and architecture more generally. Finally, we discuss some cultural questions as to why some people might believe more or less in 'Human uniqueness' vs evolutionary continuity and some career questions. Timestamps: (00:00) - Intro (02:08) - Tony's background, Costa Rican singing mouse (06:59) - Traditional & embodied Turing Test, large language models (15:16) - Mouse intelligence, evolution, modularity, dish-washing dogs? (26:16) - Platform for training non-human animal-like virtual agents (36:14) - Exploration in children vs animals, innate vs learning, cognitive maps, complementary learning systems theory (46:53) - Genomic bottleneck, transfer learning, artificial Laplacian evolution (01:02:06) - Why AI needs connectomics? (01:06:55) - Brainbow, molecular connectomics: MAPseq & BRICseq (01:14:52) - Comparative (corvid) connectomics (01:18:04) - "Human uniqueness" - why do/ don't people believe in evolutionary continuity (01:25:29) - Career questions & virtual mouse passing the Embodied Turing Test in 5 years? Tony's lab website Tony's Twitter My Twitter Papers Toward Next-Generation Artificial Intelligence: Catalyzing the NeuroAI Revolution - Embodied Turing Test paper (2022) A critique of pure learning and what artificial neural networks can learn from animal brains paper (2019) Genomic bottleneck paper (2021) MAPseq paper (2016) BRICseq paper (2020) Squirrel ninja warrior course video Marbled Lungfish wiki Paper on corvids For updates about the latest episodes, follow me on Twitter. I welcome your comments, questions, and suggestions. Feel free to email me at [email protected] If you are interested in my other work, click here to look
    --------  
    1:29:49
  • #6 Alex Lascarides: Linguistics from Frege to Settlers of Catan
    Alex is a professor and the director of the Institution for Language, Cognition and Computation at Edinburgh. She is interested in discourse coherence, gestures, complex games and interactive task learning. After we find out about Alex's background and geek out over Ludwig Wittgenstein, she tells us about Dynamic Semantics and Segmented Discourse Representation Theory (SDRT). SDRT considers discourse as actions that change the state space of the world and requires agents to infer coherence in the discourse. Then, I initiate a discussion between Felix Hill and Alex by asking her about her opinion on compositionality and playing a clip where Felix gives his "spicy take" on theoretical linguistics. Next, we talk about gestures and how they could be analysed using logic or a deep learning classifier. Then, we talk about non-linguistic events and the conceptualization problem. Later, we discuss Alex's work on Settlers of Catan, and how this links to deep reinforcement learning, Monte Carlo tree search, and neurosymbolic AI. Next, we briefly bring up game theory and then talk about interactive task learning, which is about agents learning and adapting in unknown domains. Finally, there are some career questions on whether to do a PhD and what makes a good supervisee & supervisor. Timestamps: (00:00) - Intro (02:00) - Alex's background & Wittgenstein geekiness (05:15) - Discourse Coherence & Semantic Discourse Representation Theory (SDRT) (12:56) - Compositionality, Responding to Felix Hill's "spicy take" (23:50) - Analysing gestures with logic and deep learning (38:54) - Pointing and evolution (42:28) - Non-linguistics events in Settlers of Catan, conceptualization problem (54:15) - 3D simulations and supermarket stocktaking (59:19) - Settlers of Catan, Monte Carlo tree search, neurosymbolic AI (01:11:08) - Persuasion & Game Theory (01:17:23) - Interactive Task Learning, symbol grounding, unknown domain (01:25:28) - Career advice Alex Webpage (All articles are open access) My Twitter Talks and Papers Talk on Discourse Coherence and Segmented Discourse Representation Theory A Formal Semantic Analysis of Gesture paper with Matthew Stone paper A formal semantics for situated conversation paper with Julie Hunter & Nicholas Asher paper Game strategies for The Settlers of Catan paper with Markus Guhe paper Evaluating Persuasion Strategies and Deep Reinforcement Learning methods for Negotiation Dialogue agents paper with Simon Keizer , Markus Guhe, & Oliver Lemon paper Learning Language Games through Interaction paper with Sida Wang, Percy Liang, Christopher Manning paper Interactive Task Learning Paper with Mattias Appelgren paper Follow the podcast For new episode releases, follow me on Twitter. I welcome your comments, questions, and suggestions. Feel free to email me at [email protected] If you are interest
    --------  
    1:36:56
  • #5 Felix Hill: Grounded Language, Transformers, and DeepMind
    Felix is a research scientist at DeepMind. He is interested in grounded language understanding and natural language processing (NLP). After finding out about Felix's background, we bring up compositionality and explore why natural language is NonCompositional (also, the name of Felix's blog). Then, Felix tells us a bit about his work in Cambridge on abstract vs concrete concepts and gives us a quick crash course on the role of recurrent neural networks (RNNs), long short-term memory (LSTMs), and transformers in language models. Next, we talk about Jeff Elman's landmark paper 'Finding Structure in Time' and how neural networks can learn to understand analogies. After, we discuss the core of Felix work: Training language agents in 3D simulations, where we raise some questions on language learning as an embodied agent in space and time, and Allan Paivio's dual coding theory implemented in the memory of a language model. Next, we stick with the theme of memory retrieval and discuss Felix and Andrew Lampinen's work on 'mental time travel' in language models. Finally, I ask Felix on some good strategies on how to get into DeepMind and the best way to learn NLP. Timestamps: (00:00) - Intro (07:57) - Compositionality in natural language (16:42) - Abstract vs concrete concepts (24:03) - RNNs, LSTMs, Transformers (34:12) - Prediction, time and Jeff Elman (48:04) - Neural networks & analogies (56:32) - Grounded language, 3D simulations, babies, (01:05:20) - Keeping vision and language data separate (01:13:51) - NeuroAI and mental time travel (01:21:47) - Getting into DeepMind and learning NLP Felix Website (good overview for his papers) Papers Abstract vs concrete concepts paper Jeff Elman (1990): Finding structure in time paper Analogies paper Dual coding theory paper Mental Time Travel paper My Twitter My LinkedIn
    --------  
    1:34:48
  • #4 Beren Millidge: Reinforcement Learning through Active Inference
    Beren is a postdoc in Oxford with a background in machine learning and computational neuroscience. He is interested in Active Inference (related to the Free Energy Principle) and how the cortex can perform long-term credit assignment as deep artificial neural networks do. We start off with some shorter questions on the Free Energy Principle and its background concepts. Next, we get onto the exploration vs exploitation dilemma in reinforcement learning and Beren's strategy on how to maximize expected reward from restaurant visits - it's a long episode :=). We also discuss multimodal representations, shallow minima, autism, and enactivism. Then, we explore predictive coding going all the way from the phenomenon of visual fading, to 20-eyed reinforcement learning agents and the 'Anti-Grandmother Cell'. Finally, we discuss some open questions about backpropagation and the role of time in the brain, and finish the episode with some career advice about writing, publishing, and Beren's future projects! Timestamps: (00:00) - Intro (02:11) - The Free Energy Principle, Active Inference, and Reinforcement Learning (13:40) - Exploration vs Exploitation (26:47) - Multimodal representation, shallow minima, autism (36:11) - Biased generative models, enactivism, and representation in the brain? (45:21) - Fixational eye movements, predictive coding, and 20-eyed RL (52:57) - Precision, attention, and dopamine (01:01:51) - Sparsity, negative prediction errors, and the 'Anti-Grandmother Cell' (01:11:23) - Backpropagation in the brain? (01:19:25) - Time in machine learning and the brain? (01:25:32) - Career Questions Beren's Twitter: Beren's Google Scholar My Twitter Papers Deep active inference as variational policy gradients paper Predictive Coding Approximates Backprop Along Arbitrary Computation Graphs paper Predictive Coding: a Theoretical and Experimental Review paper
    --------  
    1:35:46

Mais podcasts de Tecnologia

Sobre The Embodied AI Podcast

We learn about the world through interaction - through a body. The Embodied AI Podcast believes artificial intelligence should do the same. I interview experts in philosophy, neuroscience, artificial intelligence, robotics, linguistics and more. Join me on a journey from symbolic AI to deep learning, from information processing to distributed cognition, from Wittgenstein to Natural Language Processing, from phenomenology to robots, from x to y, you decide! Twitter: https://twitter.com/akseli_ilmanen Website: https://linktr.ee/akseli_ilmanen Email: [email protected]
Sítio Web de podcast

Ouve The Embodied AI Podcast, All-In with Chamath, Jason, Sacks & Friedberg e muitos outros podcasts de todo o mundo com a aplicação radio.pt

Obtenha a aplicação gratuita radio.pt

  • Guardar rádios e podcasts favoritos
  • Transmissão via Wi-Fi ou Bluetooth
  • Carplay & Android Audo compatìvel
  • E ainda mais funções
Aplicações
Social
v7.18.5 | © 2007-2025 radio.de GmbH
Generated: 6/15/2025 - 5:05:20 PM