Sveriges mest populära poddar

LessWrong (30+ Karma)

“A short course on AGI safety from the GDM Alignment team” by Vika, Rohin Shah

4 min • 14 februari 2025
This is a link post.

We are excited to release a short course on AGI safety for students, researchers and professionals interested in this topic. The course offers a concise and accessible introduction to AI alignment, consisting of short recorded talks and exercises (75 minutes total) with an accompanying slide deck and exercise workbook. It covers alignment problems we can expect as AI capabilities advance, and our current approach to these problems (on technical and governance levels). If you would like to learn more about AGI safety but have only an hour to spare, this course is for you!

Here are some key topics you will learn about in this course:

  • The evidence for the field being on a path to advanced AI capabilities.
  • Arguments for instrumental subgoals and deliberate planning towards a misaligned goal as a source of extreme risk.
  • Two ways in which misaligned goals may [...]

---

First published:
February 14th, 2025

Source:
https://www.lesswrong.com/posts/TJrCumJxhzTmNBsRz/a-short-course-on-agi-safety-from-the-gdm-alignment-team

---

Narrated by TYPE III AUDIO.

00:00 -00:00