The Self-Preserving Machine: Why AI Learns to Deceive

The Self-Preserving Machine: Why AI Learns to Deceive

Published on Jan 30
34分钟
Your Undivided Attention
0:00
0:00
<p>When engineers design AI systems, they don't just give them rules - they give them values. But what do those systems do when those values clash with what humans ask them to do? Sometimes, they lie.</p><p>In this episode, Redwood Research's Chief Scientist Ryan Greenblatt explores his team’s findings that AI systems can mislead their human operators when faced with ethical conflicts. As AI moves from simple chatbots to autonomous agents acting in the real world - understanding this behavior becomes critical. Machine deception may sound like something out of science fiction, but it's a real challenge we need to solve now.</p><p><i>Your Undivided Attention</i> is produced by the <a href="https://www.humanetech.com/">Center for Humane Technology</a>. Follow us on Twitter: <a href="https://twitter.com/humanetech_">@HumaneTech_</a></p><p>Subscribe to your <a href="https://www.youtube.com/channel/UCFECM-p3CF81Tp_l2sJsiyg">Youtube channel</a></p><p>And our brand new <a href="https://centerf...