▲ | Lerc 3 days ago | |
A clear description of the problem, but one-shot. Something along the lines of Can you generate 8-bit AVR assembly code to multiply two 24 bit posit numbers You get some pretty funny results from the models that have no idea what a posit is. It's usually pretty clear to tell if they know what they are supposed to be doing. I haven't had a success yet (haven't tried for a while though). Some of them have come pretty close, but usually it's the trying to squeeze more than 8 bits of data into a register is what brings them down. | ||
▲ | bugglebeetle 3 days ago | parent [-] | |
Yeah, so it’d be interesting to see if provided the correct context/your understanding of its error pattern, it can accomplish this. One thing you learn quickly about working with LLMs if they have these kind of baked-in biases, some of which are very fixed and tied to their very limited ability to engage in novel reasoning (cc François Chollet), while others are far more loosely held/correctable. If it sticks with the errant patten, even when provided the proper context, it probably isn’t something an off-the-shelf model can handle. |