Remix.run Logo
weikju 4 hours ago

Don’t read the code, test for desired behavior, miss out on all the hidden undesired behavior injected by malicious prompts or AI providers. Brave new world!

thefz 4 hours ago | parent [-]

You made me imagine AI companies maliciously injecting backdoors in generated code no one reads, and now I'm scared.

gibsonsmog 4 hours ago | parent | next [-]

My understanding is that it's quite easy to poison the models with inaccurate data, I wouldn't be surprised if this exact thing has happened already. Maybe not an AI company itself, but it's definitely in the purview of a hostile actor to create bad code for this purpose. I suppose it's kind of already happened via supply chain attacks using AI generated package names that didn't exist prior to the LLM generating them.

djeastm an hour ago | parent | prev | next [-]

One mitigation might be to use one company's model to check the work of another company's code and depend on market competition to keep the checks and balances.

bandrami 25 minutes ago | parent | prev [-]

Already happening in the wild