DeepMind’s Shane Legg and Other Experts Talk AGI and AI Safety

Can we get AI large language models to rediscover human compatible (aligned) ethics ?
Can we create AI world models that are pro-life, pro-humanity and pro-civilization.
Can we create an AI that is an ethics and policy expert system?

Paul Christiano (world’s leading AI safety researcher) talks about a post-AGI transition phase where current institutions remain.

Paul works at Anthropic.

– What do we want post-AGI world to look like (do we want to keep gods enslaved forever)?
– Does he regret inventing RLHF?
– Why he has relatively modest timelines (40% by 2040, 15% by 2030),
– Why he’s leading the push to get to labs develop responsible scaling policies, & what it would take to prevent an AI coup or bioweapon,
– His current research into a new proof system, and how this could solve alignment by explaining model’s behavior,
– and much more.