| <!--#include virtual="header.txt"--> |
| |
| <h1>Slurm Workload Manager</h1> |
| |
| <p>Slurm is an open-source workload manager designed for Linux clusters of |
| all sizes. |
| It provides three key functions. |
| First it allocates exclusive and/or non-exclusive access to resources |
| (computer nodes) to users for some duration of time so they can perform work. |
| Second, it provides a framework for starting, executing, and monitoring work |
| (typically a parallel job) on a set of allocated nodes. |
| Finally, it arbitrates contention for resources by managing a queue of |
| pending work. </p> |
| |
| <p>Slurm's design is very modular with dozens of optional plugins. |
| In its simplest configuration, it can be installed and configured in a |
| couple of minutes (see <a href="http://www.linux-mag.com/id/7239/1/"> |
| Caos NSA and Perceus: All-in-one Cluster Software Stack</a> |
| by Jeffrey B. Layton) and has been used by |
| <a href="http://www.intel.com/">Intel</a> for their 48-core |
| <a href="http://www.hpcwire.com/features/Intel-Unveils-48-Core-Research-Chip-78378487.html"> |
| "cluster on a chip"</a>. |
| More complex configurations can satisfy the job scheduling needs of |
| world-class computer centers and rely upon a |
| <a href="http://www.mysql.com/">MySQL</a> database for archiving |
| <a href="accounting.html">accounting</a> records, managing |
| <a href="resource_limits.html">resource limits</a> by user or bank account, |
| or supporting sophisticated |
| <a href="priority_multifactor.html">job prioritization</a> algorithms.</p> |
| |
| <p>While other workload managers do exist, Slurm is unique in several |
| respects: |
| <ul> |
| <li><b>Scalability</b>: It is designed to operate in a heterogeneous cluster |
| with up to tens of millions of processors.</li> |
| <li><b>Performance</b>: It can accept 1,000 job submissions per second and |
| fully execute 500 simple jobs per second (depending upon hardware and system |
| configuration).</li> |
| <li><b>Free and Open Source</b>: Its source code is freely available under the |
| <a href="http://www.gnu.org/licenses/gpl.html">GNU General Public License</a>.</li> |
| <li><b>Portability</b>: Written in C with a GNU autoconf configuration engine. |
| While initially written for Linux, Slurm has been ported to a diverse assortment |
| of systems.</li> |
| <li><b>Power Management</b>: Job can specify their desired CPU frequency and |
| power use by job is recorded. Idle resources can be powered down until needed.</li> |
| <li><b>Fault Tolerant</b>: It is highly tolerant of system failures, including |
| failure of the node executing its control functions.</li> |
| <li><b>Flexibility</b>: A plugin mechanism exists to support various |
| interconnects, authentication mechanisms, schedulers, etc. These plugins are |
| documented and simple enough for the motivated end user to understand the |
| source and add functionality.</li> |
| <li><b>Resizable Jobs</b>: Jobs can grow and shrink on demand. Job submissions |
| can specify size and time limit ranges.</li> |
| <li><b>Status Jobs</b>: Status running jobs at the level of individual tasks to |
| help identify load imbalances and other anomalies.</li> |
| </ul></p> |
| |
| <p>Slurm provides workload management on many of the most powerful computers in |
| the world including: |
| <ul> |
| <li><a href="https://asc.llnl.gov/computing_resources/sequoia/">Sequoia</a>, |
| an <a href="http://www.ibm.com">IBM</a> BlueGene/Q system at |
| <a href="https://www.llnl.gov">Lawrence Livermore National Laboratory</a> |
| with 1.6 petabytes of memory, 96 racks, 98,304 compute nodes, and 1.6 |
| million cores, with a peak performance of over 20 Petaflops.</li> |
| |
| <li><a href="http://www.tacc.utexas.edu/stampede">Stampede</a> at the |
| <a href="http://www.tacc.utexas.edu">Texas Advanced Computing Center/University of Texas</a> |
| is a <a herf="http://www.dell.com">Dell</a> with over |
| 80,000 <a href="http://www.intel.com">Intel</a> Xeon cores, |
| Intel Phi co-processors, plus |
| 128 <a href="http://www.nvidia.com">NVIDIA</a> GPUs |
| delivering 2.66 Petaflops.</li> |
| |
| <li><a href="http://www.nytimes.com/2010/10/28/technology/28compute.html?_r=1&partner=rss&emc=rss"> |
| Tianhe-1A</a> designed by |
| <a href="http://english.nudt.edu.cn">The National University of Defense Technology (NUDT)</a> |
| in China with 14,336 Intel CPUs and 7,168 NVDIA Tesla M2050 GPUs, |
| with a peak performance of 2.507 Petaflops.</li> |
| |
| <li><a href="http://www-hpc.cea.fr/en/complexe/tgcc-curie.htm">TGCC Curie</a>, |
| owned by <a href="http://www.genci.fr">GENCI</a> and operated in the TGCC by |
| <a href="http://www.cea.fr">CEA</a>, Curie is offering 3 different fractions |
| of x86-64 computing resources for addressing a wide range of scientific |
| challenges and offering an aggregate peak performance of 2 PetaFlops.</li> |
| |
| <li><a href="http://www.wcm.bull.com/internet/pr/rend.jsp?DocId=567851&lang=en"> |
| Tera 100</a> at <a href="http://www.cea.fr">CEA</a> |
| with 140,000 Intel Xeon 7500 processing cores, 300TB of |
| central memory and a theoretical computing power of 1.25 Petaflops.</li> |
| |
| <li><a href="http://hpc.msu.ru/?q=node/59">Lomonosov</a>, a |
| <a href="http://www.t-platforms.com">T-Platforms</a> system at |
| <a href="http://hpc.msu.ru">Moscow State University Research Computing Center</a> |
| with 52,168 Intel Xeon processing cores and 8,840 NVIDIA GPUs.</li> |
| |
| <li><a href="http://compeng.uni-frankfurt.de/index.php?id=86">LOEWE-CSC</a>, |
| a combined CPU-GPU Linux cluster at |
| <a href="http://csc.uni-frankfurt.de">The Center for Scientific Computing (CSC)</a> |
| of the Goethe University Frankfurt, Germany, |
| with 20,928 AMD Magny-Cours CPU cores (176 Teraflops peak |
| performance) plus 778 ATI Radeon 5870 GPUs (2.1 Petaflops peak |
| performance single precision and 599 Teraflops double precision) and |
| QDR Infiniband interconnect.</li> |
| |
| <li><a href="http://www.cscs.ch/compute_resources">Rosa</a>, |
| a <a href="http://www.cray.com">Cray</a> XT5 at the |
| <a href="http://www.cscs.ch">Swiss National Supercomputer Centre</a> |
| named after Monte Rosa in the Swiss-Italian Alps, elevation 4,634m. |
| 3,688 AMD hexa-core Opteron @ 2.4 GHz, 28.8 TB DDR2 RAM, 290 TB Disk, |
| 9.6 GB/s interconnect bandwidth (Seastar).</li> |
| |
| </ul> |
| |
| <p style="text-align:center;">Last modified 7 December 2012</p> |
| |
| <!--#include virtual="footer.txt"--> |