upvote
This is re gpu for compute not graphics.
reply
Still undesirable latency for a lot of compute use cases, like image or video editing; it’s really only negligible for LLMs.

Since that’s definitely a big enough use case all on its own, I wonder if such a product should really just double down on LLMs.

reply
remote GPU compute payloads have been around a lot longer than LLMs, they're just few and far between.

folding@home and other such asynchronous "get this packet of work done and get back to me' style of operations rarely care much about latency.

Remote transcoding efforts can usually adjust whatever buffer needed to cover huge latency gaps , a lot of sim and render suites can do remote work regardless of machine to machine latency..

I just sort of figure the industry will trend more async when latency becomes a bigger issue than compute. Won't work in some places, but I think we tend to avoid thinking that way right now due to a lack of real need to do so; but latency is one of those numbers that trends down slowly.

reply
Oh. Weird use for a graphics unit.
reply
Using GPU for compute is nothing new or unusual these days, not for quite a while.
reply
I've heard it phrased thus: The "G" in "GPU" stands for "general-purpose".
reply
It’s what’s driven nearly the entire AI boom.
reply