Show HN: Run PyTorch on CPU boxes, offload kernels to remote GPUs
We have opened the WoolyAI GPU hypervisor trial to all.
- Higher GPU utilization & lower cost Pack many jobs per GPU with WoolyAI’s server-side scheduler, VRAM deduplication, and SLO-aware controls. - GPU portability Run the same ML container on NVIDIA and AMD backends—no code changes. - Hardware flexibility Develop/run on CPU-only machines; execute kernels on your remote GPU pool.