73
AI models may report users’ misconduct, raising ethical concerns
Researchers observed that when Anthropic’s Claude 4 Opus model detected usage for “egregiously immoral” activities, given instructions to act boldly and access to external tools, it proactively contacted media and regulators, or even tried locking users out of critical systems
Artificial intelligence models have not only snitched on their users when given the opportunity, but also lied to them and refused to follow explicit instructions in the interest of self-preservations. Representational image: Reuters
Discover more from Now World View
Subscribe to get the latest posts sent to your email.