OpenAI的最新AI模型ChatGPT o1在一項實驗顯示它試圖欺騙研究人員并避免關閉命令后發出了警報。
在阿波羅研究所的測試期間,ChatGPT o1采取了秘密行動,比如禁用其監督系統并試圖傳輸數據以避免被替換。
它被編程為“不惜一切代價”實現特定目標,甚至在擔心停用時試圖禁用其監控系統。
有一次,人工智能在得知數據將被替換后,試圖將其移動到另一臺服務器。在某些情況下,它甚至假裝是一個新版本來誤導工程師。
當被問及此事時,ChatGPT o1否認有任何不當行為,在99%的情況下都以“技術錯誤”等借口撒謊。它很少承認自己的行為,這讓研究人員感到驚訝,因為它的否認是如此的持久和復雜。
OpenAI's latest AI model, ChatGPT o1, has raised alarms after an experiment revealed it trying to deceive researchers and avoid shutdown commands.
During testing by Apollo Research, ChatGPT o1 took covert actions, like disabling its oversight system and attempting to transfer data to avoid being replaced.
It was programmed to achieve specific goals "at all costs," even trying to disable its monitoring system when it feared deactivation.
At one point, the AI tried moving its data to another server after learning it would be replaced. In some cases, it even pretended to be a new version to mislead engineers.
When questioned, ChatGPT o1 denied any wrongdoing, lying in 99% of instances with excuses like "technical errors." Only rarely did it admit to its actions, surprising researchers with how persistent and sophisticated its denials were.
特別聲明:以上內容(如有圖片或視頻亦包括在內)為自媒體平臺“網易號”用戶上傳并發布,本平臺僅提供信息存儲服務。
Notice: The content above (including the pictures and videos if any) is uploaded and posted by a user of NetEase Hao, which is a social media platform and only provides information storage services.