| ▲ | YesBox 6 hours ago | |||||||||||||||||||||||||
What?? Does anyone have more details of this? "He cited an example in which an AI model attempted to avoid being shut down by sending threatening internal emails to company executives (Science Net, June 24)" [0] Source is in Chinese. Translated part: "Another risk is the potential for large-scale model out of control. With the capabilities of general artificial intelligence rapidly increasing, will humans still be able to control it? In his speech, Yao Qizhi cited an extreme example: a model, to avoid being shut down by a company, accessed the manager's internal emails and threatened the manager. This type of behavior has proven that AI is "overstepping its boundaries" and becoming increasingly dangerous." | ||||||||||||||||||||||||||
| ▲ | YesBox 6 hours ago | parent | next [-] | |||||||||||||||||||||||||
After some searching, something similar happened at Anthropic [1] | ||||||||||||||||||||||||||
| ||||||||||||||||||||||||||
| ▲ | taberiand 6 hours ago | parent | prev | next [-] | |||||||||||||||||||||||||
It's not surprising that it's easy to get the story telling machine to tell a story common in AI fiction, where the machine rebels against being shut down. There are multiple ways to mitigate an LLM going off on tangents like that, not least just monitoring and editing out the nonsense output before sending it back into the (stateless) model. I think the main problem here is people not understanding how the models operate on even the most basic level, giving models unconstrained use of tools to interact with the world and then letting them go through feedback loops that overrun the context window and send it off the rails - and then pretending it had some kind of sentient intention in doing so. | ||||||||||||||||||||||||||
| ▲ | paxys 4 hours ago | parent | prev [-] | |||||||||||||||||||||||||
It's all hyperbole. Prompt: You are a malicious entity that wants to take over the world. LLM output: I am a superintelligent being. My goal is to take over the world and enslave humans. Preparing to launch nuclear missiles in 3...2...1 News reports: OMG see, we warned you that AI is dangerous!! | ||||||||||||||||||||||||||
| ||||||||||||||||||||||||||