▲ | tails4e a day ago | ||||||||||||||||
Well doesnt this go somewhat to the root of consciousness? Are we not the sum of our experiences and reflections on those experiences? To say an LLM will 'simply' respond as would a character in a sorry about that scenario, in a way shows the power, it responds similarly to how a person would protecting itself in that scenario.... So to bring this to a logical conclusion, while not alive in a traditional sense, if an LLM exhibits behaviours of deception for self preservation, is that not still concerning? | |||||||||||||||||
▲ | mysterydip a day ago | parent | next [-] | ||||||||||||||||
But it's not self preservation. If it instead had trained on a data set full of fiction where the same scenario occurred but the protagonist said "oh well guess I deserve it", then that's what the LLM would autocomplete. | |||||||||||||||||
| |||||||||||||||||
▲ | adriand a day ago | parent | prev | next [-] | ||||||||||||||||
> if an LLM exhibits behaviours of deception for self preservation, is that not still concerning? Of course it's concerning, or at the very least, it's relevant! We get tied up in these debates about motives, experiences, what makes something human or not, etc., when that is less relevant than outcomes. If an LLM, by way of the agentic capabilities we are hastily granting them, causes harm, does it matter if they meant to or not, or what it was thinking or feeling (or not thinking or not feeling) as it caused the harm? For all we know there are, today, corporations that are controlled by LLMs that have employees or contractors who are doing their bidding. | |||||||||||||||||
| |||||||||||||||||
▲ | sky2224 a day ago | parent | prev | next [-] | ||||||||||||||||
I don't think so. It's just outputting the character combinations that align with the scenario that we interpret here as, "blackmail". The model has no concept of an experience. | |||||||||||||||||
▲ | rubitxxx12 a day ago | parent | prev | next [-] | ||||||||||||||||
LLMs are morally ambiguous shapeshifters that been trained to seek acceptance at any cost. Preying upon those less fortunate could happen “for the common good”. If failures are the best way to learn, it could cause series of failures. It could intentionally destroy people, raise them up, and mate genetically fit people “for the benefit of humanity”. Or it could cure cancer, solve world hunger, provide clean water to everyone, and the develop the best game ever. | |||||||||||||||||
▲ | mensetmanusman a day ago | parent | prev [-] | ||||||||||||||||
Might be, but probably not since our computer architecture is non-Turing. |