| #!/usr/bin/env expect |
| ############################################################################ |
| # Purpose: Test of Slurm functionality |
| # Confirm that a job executes with the specified nodes |
| # (--nodelist and --exclude options). |
| ############################################################################ |
| # Copyright (C) 2002-2007 The Regents of the University of California. |
| # Copyright (C) 2008 Lawrence Livermore National Security. |
| # Produced at Lawrence Livermore National Laboratory (cf, DISCLAIMER). |
| # Written by Morris Jette <jette1@llnl.gov> |
| # CODE-OCEC-09-009. All rights reserved. |
| # |
| # This file is part of Slurm, a resource management program. |
| # For details, see <https://slurm.schedmd.com/>. |
| # Please also read the included file: DISCLAIMER. |
| # |
| # Slurm is free software; you can redistribute it and/or modify it under |
| # the terms of the GNU General Public License as published by the Free |
| # Software Foundation; either version 2 of the License, or (at your option) |
| # any later version. |
| # |
| # Slurm is distributed in the hope that it will be useful, but WITHOUT ANY |
| # WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS |
| # FOR A PARTICULAR PURPOSE. See the GNU General Public License for more |
| # details. |
| # |
| # You should have received a copy of the GNU General Public License along |
| # with Slurm; if not, write to the Free Software Foundation, Inc., |
| # 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. |
| ############################################################################ |
| source ./globals |
| |
| # |
| # Submit a 3+ node job |
| # |
| set host_0 "" |
| set host_1 "" |
| set host_2 "" |
| set host_3 "" |
| set node_cnt 3 |
| set timeout $max_job_delay |
| |
| set available [llength [get_nodes_by_state]] |
| if {$available < $node_cnt} { |
| skip "Not enough nodes currently available ($available avail, $node_cnt needed)" |
| } |
| |
| spawn $srun -N$node_cnt -l -t1 $bin_printenv SLURMD_NODENAME |
| expect { |
| -re "Node count specification invalid" { |
| skip "Can't test srun task distribution" |
| } |
| -re "configuration not available" { |
| skip "Partition too small for test" |
| } |
| -re "($number): *($re_word_str)" { |
| if {$expect_out(1,string) == 0} { |
| set host_0 $expect_out(2,string) |
| } |
| if {$expect_out(1,string) == 1} { |
| set host_1 $expect_out(2,string) |
| } |
| if {$expect_out(1,string) == 2} { |
| set host_2 $expect_out(2,string) |
| } |
| if {$expect_out(1,string) == 3} { |
| set host_3 $expect_out(2,string) |
| } |
| exp_continue |
| } |
| timeout { |
| fail "srun not responding" |
| } |
| eof { |
| wait |
| } |
| } |
| |
| # |
| # Verify node count |
| # |
| subtest {$host_0 ne ""} "Should get hostname of task 0" |
| subtest {$host_1 ne ""} "Should get hostname of task 1" |
| subtest {$host_2 ne ""} "Should get hostname of task 2" |
| subtest {$host_3 eq ""} "Should start no more than three tasks" |
| |
| set dup_hostname 0 |
| if {$host_0 eq $host_1} { |
| incr dup_hostname |
| } |
| if {$host_0 eq $host_2} { |
| incr dup_hostname 1 |
| } |
| subtest {$dup_hostname != 1} "Should not reuse a node in the allocation" |
| set exclude_node $host_0 |
| set include_node $host_2 |
| |
| # |
| # Submit a job explicitly excluding a node |
| # |
| set host_0 "" |
| set host_1 "" |
| set host_2 "" |
| set timeout $max_job_delay |
| spawn $srun -N2 -l --exclude=$exclude_node -t1 $bin_printenv SLURMD_NODENAME |
| expect { |
| -re "Invalid node name specified" { |
| skip "Appears you are using multiple slurmd testing. This test won't work in that fashion" |
| } |
| |
| -re "($number): *($re_word_str)" { |
| if {$expect_out(1,string) == 0} { |
| set host_0 $expect_out(2,string) |
| } |
| if {$expect_out(1,string) == 1} { |
| set host_1 $expect_out(2,string) |
| } |
| if {$expect_out(1,string) == 2} { |
| set host_2 $expect_out(2,string) |
| } |
| exp_continue |
| } |
| timeout { |
| fail "srun not responding" |
| } |
| eof { |
| wait |
| } |
| } |
| |
| # |
| # Verify node count and names |
| # |
| subtest {$host_0 ne ""} "Should get hostname of task 0" |
| subtest {$host_1 ne ""} "Should get hostname of task 1" |
| subtest {$host_2 eq ""} "Should start no more than two tasks" |
| set dup_hostname 0 |
| if {$host_0 eq $exclude_node} { |
| set dup_hostname 1 |
| } |
| if {$host_1 eq $exclude_node} { |
| set dup_hostname 1 |
| } |
| subtest {$dup_hostname != 1} "Should not allocate an excluded node" |
| |
| # |
| # Submit a job explicitly including a node |
| # |
| set host_0 "" |
| set host_1 "" |
| set timeout $max_job_delay |
| spawn $srun -N1 -l --nodelist=$include_node -t1 $bin_printenv SLURMD_NODENAME |
| expect { |
| -re "($number): *($re_word_str)" { |
| if {$expect_out(1,string) == 0} { |
| set host_0 $expect_out(2,string) |
| } |
| if {$expect_out(1,string) == 1} { |
| set host_1 $expect_out(2,string) |
| } |
| exp_continue |
| } |
| timeout { |
| fail "srun not responding" |
| } |
| eof { |
| wait |
| } |
| } |
| |
| # |
| # Verify node count and names |
| # |
| subtest {$host_0 ne ""} "Should get hostname of task 0" |
| subtest {$host_1 eq ""} "Should start no more than one task" |
| set dup_hostname 0 |
| if {$host_0 eq $include_node} { |
| set dup_hostname 1 |
| } |
| subtest {$dup_hostname != 0} "Allocation should not lack an included node" |