Remix.run Logo
amluto 4 hours ago

There is a major issue with current AI tools that they want to effectively grant access to everything their user has access to. The whole sandbox structure is wrong (although various people have vibe coded assorted improvements).

yonatan8070 3 hours ago | parent | next [-]

Another issue I've noticed is they're sometimes very resourceful. For example when Codex can't directly edit file due to sandboxing restrictions, rather than asking "hey can I apply this diff on the file", it'd ask for permission to run a `cat EOF` command to re-write the whole file, which the UI doesn't surface properly (just shows the first line...).

This sounds similar to what's described in the "Claude deleted my DB post", it decided "I need to do X", then searched for whatever would let it do X, regardless of intended purpose.

amluto 2 hours ago | parent [-]

I amused myself by removing codex-rs’s web search tool and then asking it to search for “foo”. It wrote a Python script to do the search.

traderj0e 2 hours ago | parent | prev | next [-]

If you pretend you have an intern with their own machine and run the AI agents on that machine, you have the same separation.

zahlman 3 hours ago | parent | prev [-]

If you want them to be able to write code and then run tests on that code, it can be a bit difficult to restrict access meaningfully....

amluto 2 hours ago | parent [-]

Only for code that can’t be tested in an isolated environment, and designing code that can’t be tested in an isolated environment is generally a mistake for quite a few reasons.