Running Large-Scale GPU Workloads on Kubernetes with Slurm
Mirrored from NVIDIA Developer Blog for archival readability. Support the source by reading on the original site.
Slurm is an open source cluster management and job scheduling system for Linux. It manages job scheduling for over 65% of TOP500 systems. Most organizations...
Slurm is an open source cluster management and job scheduling system for Linux. It manages job scheduling for over 65% of TOP500 systems. Most organizations running large-scale AI training have years of investment in Slurm job scripts, fair-share policies, and accounting workflows. The challenge is getting Slurm scheduling capabilities onto Kubernetes—the standard platform for managing GPU…
More from NVIDIA Developer Blog
-
Accelerated X-Ray Analysis for Nanoscale Imaging (XANI) of Novel Materials
May 13
-
Transform Video Into Instantly Searchable, Actionable Intelligence with AI Agents and Skills
May 13
-
Google DeepMind paper: reinforcement learning at scale
May 13
-
How to Eliminate Pipeline Friction in AI Model Serving
May 12
Discussion (0)
Sign in to join the discussion. Free account, 30 seconds — email code or GitHub.
Sign in →No comments yet. Sign in and be the first to say something.