Remix.run Logo
simianwords 3 days ago

Thanks - still not clear what they did really. Some inference time hacks?

FergusArgyll 3 days ago | parent | next [-]

That would imply the model always had a 1m token context but they limited it in the api and app? That's strange because they can just charge more for every token past 250k (like google does, I believe).

But if not shouldn't it have to be completely retrained model? it's clearly not that - good question!

otabdeveloper4 3 days ago | parent | prev | next [-]

Most likely still 32k tokens under the hood, but with some context slicing/averaging hacks to make inference not error out on infinite input.

(That's what I do locally with llama.cpp)

Aeolun 3 days ago | parent | prev [-]

They already had 0.5M context window on the enteprise version.