top of page

#5 Felix Hill: Grounded Language, Transformers, and DeepMind

Writer's picture: Akseli IlmanenAkseli Ilmanen

Updated: Feb 6, 2023




Felix is a research scientist at DeepMind. He is interested in grounded language understanding and natural language processing (NLP). After finding out about Felix's background, we bring up compositionality and explore why natural language is NonCompositional (also, the name of Felix's blog). Then, Felix tells us a bit about his work in Cambridge on abstract vs concrete concepts and gives us a quick crash course on the role of recurrent neural networks (RNNs), long short-term memory (LSTMs), and transformers in language models. Next, we talk about Jeff Elman's landmark paper 'Finding Structure in Time' and how neural networks can learn to understand analogies. After, we discuss the core of Felix work: Training language agents in 3D simulations, where we raise some questions on language learning as an embodied agent in space and time, and Allan Paivio's dual coding theory implemented in the memory of a language model. Next, we stick with the theme of memory retrieval and discuss Felix and Andrew Lampinen's work on 'mental time travel' in language models. Finally, I ask Felix on some good strategies on how to get into DeepMind and the best way to learn NLP.


Timestamps:

(00:00) - Intro

(07:57) - Compositionality in natural language

(16:42) - Abstract vs concrete concepts

(24:03) - RNNs, LSTMs, Transformers

(34:12) - Prediction, time and Jeff Elman

(48:04) - Neural networks & analogies

(56:32) - Grounded language, 3D simulations, babies,

(01:05:20) - Keeping vision and language data separate

(01:13:51) - NeuroAI and mental time travel

(01:21:47) - Getting into DeepMind and learning NLP


Felix Website (good overview for his papers)


Papers

Abstract vs concrete concepts paper

Jeff Elman (1990): Finding structure in time paper

Analogies paper

Dual coding theory paper

Mental Time Travel paper

Recent Posts

See All

Comments


bottom of page