Sveriges mest populära poddar

The Gradient: Perspectives on AI

Joel Lehman: Open-Endedness and Evolution through Large Models

99 min • 22 september 2022

Have suggestions for future podcast guests (or other feedback)? Let us know here!

In episode 42 of The Gradient Podcast, Daniel Bashir speaks to Joel Lehman.

Joel is a machine learning scientist interested in AI safety, reinforcement learning, and creative open-ended search algorithms. Joel has spent time at Uber AI Labs and OpenAI and is the co-author of the book Why Greatness Cannot be Planned: The Myth of the Objective

Subscribe to The Gradient Podcast:  Apple Podcasts  | Spotify | Pocket Casts | RSSFollow The Gradient on Twitter

Outline:

* (00:00) Intro

* (01:40) From game development to AI

* (03:20) Why evolutionary algorithms

* (10:00) Abandoning Objectives: Evolution Through the Search for Novelty Alone

* (24:10) Measuring a desired behavior post-hoc vs optimizing for that behavior

* (27:30) Neuroevolution through Augmenting Topologies (NEAT), Evolving a Diversity of Virtual Creatures

* (35:00) Humans are an inefficient solution to evolution’s objectives

* (47:30) Is embodiment required for understanding? Today’s LLMs as practical thought experiments in disembodied understanding

* (51:15) Evolution through Large Models (ELM)

* (1:01:07) ELM: Quality Diversity Algorithms, MAP-Elites, bootstrapping training data

* (1:05:25) Dimensions of Diversity in MAP-Elites, what is “interesting”?

* (1:12:30) ELM: Fine-tuning the language model

* (1:18:00) Results of invention in ELM, complexity in creatures

* (1:20:20) Future work building on ELM, key challenges in open-endedness

* (1:24:30) How Joel’s research affects his approach to life and work

* (1:28:30) Balancing novelty and exploitation in work

* (1:34:10) Intense competition in AI, Joel’s advice for people considering ML research

* (1:38:45) Daniel isn’t the worst interviewer ever

* (1:38:50) Outro

Links:

* Joel’s webpage

* Evolution through Large Models: The Tweet

* Papers:

* Abandoning Objectives: Evolution through the search for novelty alone

* Evolving a diversity of virtual creatures through novelty search and local competition

* Designing neural networks through neuroevolution

* Evolution through Large Models

* Resources for (aspiring) ML researchers!

* Cohere for AI

* ML Collective



Get full access to The Gradient at thegradientpub.substack.com/subscribe
Förekommer på
00:00 -00:00