Quote
Use a message queue
No durable message queue is optimized for low latency. The original request was for setting up a system that spreads load across GPUs for real-time processing.
Presumably it's better to lose a few frames, than to have to process through old and outdated frames, when the system hiccups.
Quote
If a frame has multiple stages of processing, it might be wiser to do all of the stages on a single machine.
The original poster was not saying that multiple steps are spread across machines. The original requirement stated was that the results of the computer vision type analysis would be exported to some other machine.