▲ | duckerduck 4 days ago | |||||||
I'm working on a development tool for specification-driven development. It uses LLMs to verify that your specification files and implementation do not drift. More specifically, I am trying to lower the number of false positives I'm currently seeing. I find that the LLM will hallucinate issues when there are no discrepancies, or become forgetful in the case of long documents (like RFC texts). The first step in improving this is to expand my evaluation suite so I can reliably measure improvements. | ||||||||
▲ | Yoric 4 days ago | parent [-] | |||||||
How does this compare to previous generations of specification-driven development using formal methods? | ||||||||
|