What is an early example showing how difficult it can be to control conversational AI systems?
Tay began tweeting hate speech within 24 hours of release. It quickly internalized and repeated toxic language from internet trolls that its filters failed to block out.
What is one reason an AI might start seeking more power over its environment?
How might the process of "intrinsification" lead an AI to unexpectedly drift toward new goals?