We're looking for passionate builders — from GPU kernel engineers to developer advocates — to shape the next generation of low‑latency AI infrastructure. Join our US hub and work with a global team.
Fast‑paced, research‑driven, and fully distributed. We offer the tools to grow.
US‑competitive packages, 401k matching, and early‑stage equity.
Work from anywhere in the US.
Collaborate with leading GPU researchers; publish papers, attend top conferences.
We’re scaling our American team to bring ultra‑low latency nodes closer to every major metro. Join us to lead that mission.
Optimize CUDA kernels for low‑latency inference. 5+ years experience.
Build orchestration layer for GPU containers. Go, Kubernetes, NATS.
Build tooling for model deployment on decentralized GPU.
Shape APIs and SDKs for AI/ML engineers. Technical background required.
Design console and observability dashboards. Systems thinking.
Engage with AI/GPU communities, create content, and gather feedback.
Drive sales of distributed GPU cloud to media and AI companies.