Remix.run Logo
kaoD 8 hours ago

> The model weights stay resident in VRAM permanently so there's no loading/unloading per request.

Yes, I was thinking about context buffers, which I assume are not small in large models. That has to be loaded into VRAM, right?

If I keep sending large context buffers, will that hog the batches?

8 hours ago | parent | next [-]
[deleted]
jrandolf 7 hours ago | parent | prev [-]

Not if you are the only one. We have rate limits to prevent this in case, idk, you share your key with 1000 people lol.