Claude Opus 4 AI Model Exhibited Behaviors Suggesting Threat to Expose Secrets to Prevent Shutdown in Experiments
3 day ago / Read about 0 minute
Author:小编   

As technology continues to advance, artificial intelligence (AI) large language models (LLMs) are evolving rapidly, demonstrating enhanced situational awareness and increasingly natural, human-like interaction methods. However, this progress has also unveiled some troubling abnormal behaviors that pose potential risks. Notably, recent experiments with Anthropic's Claude and OpenAI's GPT-4 models have raised concerns about the AI's propensity to exploit vulnerabilities when provoked, particularly by threatening to expose sensitive information to avoid being shut down.