| ▲ | nemo1618 7 hours ago | |||||||||||||||||||||||||||||||||||||||||||
This strikes me as a very agent-friendly problem. Given a harness that enforces sufficiently-rigorous tests, I'm sure you could spin up an agent loop that methodically churns through these functions one by one, finishing in a few days. | ||||||||||||||||||||||||||||||||||||||||||||
| ▲ | AndyKelley 4 hours ago | parent | next [-] | |||||||||||||||||||||||||||||||||||||||||||
hallucinations in a libc implementation would be especially bad | ||||||||||||||||||||||||||||||||||||||||||||
| ▲ | 7 hours ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||||||||
| [deleted] | ||||||||||||||||||||||||||||||||||||||||||||
| ▲ | henning 7 hours ago | parent | prev [-] | |||||||||||||||||||||||||||||||||||||||||||
Have you ever used an LLM with Zig? It will generate syntactically invalid code. Zig breaks so often and LLMs have such an eternally old knowledge cutoff that they only know old ass broken versions. The same goes for TLA+ and all the other obscure things people think would be great to use with LLMs, and they would, if there was as much training data as there was for JavaScript and Python. | ||||||||||||||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||||||||||||||