Remix.run Logo
mort96 5 days ago

I mean when it comes time to output the image from the GPU, I don't want to add a hundred milliseconds of network latency...

whalesalad 5 days ago | parent [-]

This is re gpu for compute not graphics.

mattnewton 5 days ago | parent | next [-]

Still undesirable latency for a lot of compute use cases, like image or video editing; it’s really only negligible for LLMs.

Since that’s definitely a big enough use case all on its own, I wonder if such a product should really just double down on LLMs.

serf 4 days ago | parent [-]

remote GPU compute payloads have been around a lot longer than LLMs, they're just few and far between.

folding@home and other such asynchronous "get this packet of work done and get back to me' style of operations rarely care much about latency.

Remote transcoding efforts can usually adjust whatever buffer needed to cover huge latency gaps , a lot of sim and render suites can do remote work regardless of machine to machine latency..

I just sort of figure the industry will trend more async when latency becomes a bigger issue than compute. Won't work in some places, but I think we tend to avoid thinking that way right now due to a lack of real need to do so; but latency is one of those numbers that trends down slowly.

mort96 5 days ago | parent | prev [-]

Oh. Weird use for a graphics unit.

nkrisc 5 days ago | parent | next [-]

Using GPU for compute is nothing new or unusual these days, not for quite a while.

userbinator 4 days ago | parent | next [-]

I've heard it phrased thus: The "G" in "GPU" stands for "general-purpose".

mort96 4 days ago | parent | prev [-]

No, but its primary purpose remains graphics

nkrisc 4 days ago | parent [-]

Arguably that’s no longer true.

lostlogin 5 days ago | parent | prev [-]

It’s what’s driven nearly the entire AI boom.