| #!/usr/bin/env expect |
| ############################################################################ |
| # Purpose: Test of Slurm functionality |
| # Test of srun --spread-job option. |
| ############################################################################ |
| # Copyright (C) SchedMD LLC. |
| # |
| # This file is part of Slurm, a resource management program. |
| # For details, see <https://slurm.schedmd.com/>. |
| # Please also read the included file: DISCLAIMER. |
| # |
| # Slurm is free software; you can redistribute it and/or modify it under |
| # the terms of the GNU General Public License as published by the Free |
| # Software Foundation; either version 2 of the License, or (at your option) |
| # any later version. |
| # |
| # Slurm is distributed in the hope that it will be useful, but WITHOUT ANY |
| # WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS |
| # FOR A PARTICULAR PURPOSE. See the GNU General Public License for more |
| # details. |
| # |
| # You should have received a copy of the GNU General Public License along |
| # with Slurm; if not, write to the Free Software Foundation, Inc., |
| # 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. |
| ############################################################################ |
| source ./globals |
| |
| set node_cnt 0 |
| set job_id 0 |
| set jobid_excl 0 |
| |
| proc cleanup {} { |
| global job_id jobid_excl |
| |
| cancel_job [list $job_id $jobid_excl] |
| } |
| |
| proc run_spread_job { task_cnt } { |
| global max_job_delay srun scontrol bin_printenv number job_id |
| |
| set timeout $max_job_delay |
| spawn $srun -n $task_cnt --spread-job $bin_printenv SLURM_JOB_ID |
| expect { |
| -re "($number)" { |
| set job_id $expect_out(1,string) |
| exp_continue |
| } |
| timeout { |
| fail "srun not responding" |
| } |
| eof { |
| wait |
| } |
| } |
| if {$job_id == 0} { |
| fail "srun job ID not found" |
| } |
| |
| wait_for_job -fail $job_id "DONE" |
| |
| set timeout 10 |
| set num_nodes 0 |
| spawn $scontrol show job $job_id |
| expect { |
| -re "NumNodes=($number)" { |
| set num_nodes $expect_out(1,string) |
| exp_continue |
| } |
| timeout { |
| fail "scontrol not responding" |
| } |
| eof { |
| wait |
| } |
| } |
| if {$num_nodes != $task_cnt} { |
| fail "Invalid node count for job ($job_id) ($num_nodes != $task_cnt)" |
| } |
| } |
| |
| if {![check_config_select "cons_tres"]} { |
| skip "Test is only compatible with a config of SelectType=select/cons_tres" |
| } |
| |
| set nodes [get_nodes_by_state] |
| set node_cnt [llength $nodes] |
| if { $node_cnt < 3} { |
| skip "Insufficient nodes for test in default partition ($node_cnt < 3)" |
| } |
| |
| # Allocate the second node to force reproducing bug 8517 |
| set node_exclusive [lindex $nodes 1] |
| set output [run_command_output -fail "$sbatch --exclusive -w $node_exclusive --wrap 'sleep 300'"] |
| if {![regexp {Submitted batch job (\d+)} $output {} jobid_excl]} { |
| fail "Job submission of exclusive job failed" |
| } |
| wait_for_job -fail $jobid_excl "RUNNING" |
| |
| if {$node_cnt > 8} { |
| set max_task_cnt 8 |
| } else { |
| set max_task_cnt $node_cnt |
| } |
| |
| for {set inx 2} {$inx < $max_task_cnt} {incr inx 2} { |
| run_spread_job $inx |
| sleep 1 |
| } |