In order to trust machines with important jobs, we need a high level of confidence that they share our values and goals. Recent work shows that this “alignment” can be brittle, superficial, even unstable. In one study, a few training adjustments led a popular chatbot to recommend murder. On this episode, contributing writer Stephen Ornes tells host Samir Patel about what this research reveals.
Audio coda from The National Archives and Records Administration.
Fler avsnitt av The Quanta Podcast
Visa alla avsnitt av The Quanta PodcastThe Quanta Podcast med Quanta Magazine finns tillgänglig på flera plattformar. Informationen på denna sida kommer från offentliga podd-flöden.
