| ▲ | written-beyond 7 hours ago | |||||||||||||
That reduces the quality of the response though. | ||||||||||||||
| ▲ | debugnik 6 hours ago | parent | next [-] | |||||||||||||
As opposed to emitting non-JSON tokens and having to throw away the answer? | ||||||||||||||
| ||||||||||||||
| ▲ | Der_Einzige 5 hours ago | parent | prev [-] | |||||||||||||
THIS IS LIES: https://blog.dottxt.ai/say-what-you-mean.html I will die on this hill and I have a bunch of other Arxiv links from better peer reviewed sources than yours to back my claim up (i.e. NeurIPS caliber papers with more citations than yours claiming it does harm the outputs) Any actual impact of structured/constrained generation on the outputs is a SAMPLER problem, and you can fix what little impact may exist with things like https://arxiv.org/abs/2410.01103 Decoding is intentionally nerfed/kept to top_k/top_p by model providers because of a conspiracy against high temperature sampling: https://gist.github.com/Hellisotherpeople/71ba712f9f899adcb0... | ||||||||||||||
| ||||||||||||||