Can we get AI large language models to rediscover human compatible (aligned) ethics ?
Can we create AI world models that are pro-life, pro-humanity and pro-civilization.
Can we create an AI that is an ethics and policy expert system?
Paul Christiano (world’s leading AI safety researcher) talks about a post-AGI transition phase where current institutions remain.
Paul works at Anthropic.
– What do we want post-AGI world to look like (do we want to keep gods enslaved forever)?
– Does he regret inventing RLHF?
– Why he has relatively modest timelines (40% by 2040, 15% by 2030),
– Why he’s leading the push to get to labs develop responsible scaling policies, & what it would take to prevent an AI coup or bioweapon,
– His current research into a new proof system, and how this could solve alignment by explaining model’s behavior,
– and much more.
Brian Wang is a Futurist Thought Leader and a popular Science blogger with 1 million readers per month. His blog Nextbigfuture.com is ranked #1 Science News Blog. It covers many disruptive technology and trends including Space, Robotics, Artificial Intelligence, Medicine, Anti-aging Biotechnology, and Nanotechnology.
Known for identifying cutting edge technologies, he is currently a Co-Founder of a startup and fundraiser for high potential early-stage companies. He is the Head of Research for Allocations for deep technology investments and an Angel Investor at Space Angels.
A frequent speaker at corporations, he has been a TEDx speaker, a Singularity University speaker and guest at numerous interviews for radio and podcasts. He is open to public speaking and advising engagements.