| ▲ | jameson 7 hours ago | |
> "In combination with other prompt changes, it hurt coding quality, and was reverted on April 20" Do researchers know correlation between various aspects of a prompt and the response? LLM, to me at least, appears to be a wildly random function that it's difficult to rely on. Traditional systems have structured inputs and outputs, and we can know how a system returned the output. This doesn't appear to be the case for LLM where inputs and outputs are any texts. Anecdotally, I had a difficult time working with open source models at a social media firm, and something as simple as wrapping the example of JSON structure with ```, adding a newline or wording I used wildly changed accuracy. | ||