| ▲ | gmuslera 6 hours ago | |||||||||||||||||||||||||
Reminded me of pi filesystem (https://github.com/philipl/pifs), with enough digits of pi precalculated you might be able to do a decent compression program. The trick is in the amount of reasonable digits for that, if it’s smaller or bigger than that trained LLM. | ||||||||||||||||||||||||||
| ▲ | GuB-42 5 hours ago | parent | next [-] | |||||||||||||||||||||||||
I suspect that the length of the offset of your input data in pi is equal to the length of the input data itself, plus or minus a few bytes at most, regardless of the size of the input data. That is: no compression, but it won't make things worse either. Unless the input data is the digits of pi, obviously, or the result of some computation involving pi. | ||||||||||||||||||||||||||
| ||||||||||||||||||||||||||
| ▲ | sltkr 5 hours ago | parent | prev [-] | |||||||||||||||||||||||||
I'm going to be the nerd that points out that it has not been mathematically proven that pi contains every substring, so the pifs might not work even in theory (besides being utterly impractical, of course). On a more serious note, as far as I understand these compression competitions require that static data is included in the size computation. So if you compress 1000 MB into 500 MB, but to decompress you need a 1 MB binary and a 100 MB initial dictionary, your score would be 500 + 100 + 1 = 601 MB, not 500 MB. The relevance to this discussion is that the LLM weights would have to be included as static data, since the only way to regenerate them is from the initial training data, which is much larger than the resulting model. By comparison, pi based compression is the other way around: since pi is a natural constant, if your decompressor requires (say) a trillion digits of pi, you could write a relatively small program (a few kb) to generate them. It would be terribly slow, but it wouldn't affect your compression ratio much. | ||||||||||||||||||||||||||
| ||||||||||||||||||||||||||