| /****************************************************************************\ |
| * update_config.c - request that slurmctld update its configuration |
| ***************************************************************************** |
| * Copyright (C) 2002-2007 The Regents of the University of California. |
| * Copyright (C) 2008-2010 Lawrence Livermore National Security. |
| * Produced at Lawrence Livermore National Laboratory (cf, DISCLAIMER). |
| * Written by Morris Jette <jette1@llnl.gov> and Kevin Tew <tew1@llnl.gov>. |
| * CODE-OCEC-09-009. All rights reserved. |
| * |
| * This file is part of Slurm, a resource management program. |
| * For details, see <https://slurm.schedmd.com/>. |
| * Please also read the included file: DISCLAIMER. |
| * |
| * Slurm is free software; you can redistribute it and/or modify it under |
| * the terms of the GNU General Public License as published by the Free |
| * Software Foundation; either version 2 of the License, or (at your option) |
| * any later version. |
| * |
| * In addition, as a special exception, the copyright holders give permission |
| * to link the code of portions of this program with the OpenSSL library under |
| * certain conditions as described in each individual source file, and |
| * distribute linked combinations including the two. You must obey the GNU |
| * General Public License in all respects for all of the code used other than |
| * OpenSSL. If you modify file(s) with this exception, you may extend this |
| * exception to your version of the file(s), but you are not obligated to do |
| * so. If you do not wish to do so, delete this exception statement from your |
| * version. If you delete this exception statement from all source files in |
| * the program, then also delete it here. |
| * |
| * Slurm is distributed in the hope that it will be useful, but WITHOUT ANY |
| * WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS |
| * FOR A PARTICULAR PURPOSE. See the GNU General Public License for more |
| * details. |
| * |
| * You should have received a copy of the GNU General Public License along |
| * with Slurm; if not, write to the Free Software Foundation, Inc., |
| * 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. |
| \*****************************************************************************/ |
| |
| #include <errno.h> |
| #include <stdio.h> |
| #include <stdlib.h> |
| #include <string.h> |
| |
| #include "slurm/slurm.h" |
| |
| #include "src/common/slurm_protocol_api.h" |
| #include "src/common/xstring.h" |
| |
| static int _slurm_update (void * data, slurm_msg_type_t msg_type); |
| |
| /* |
| * slurm_update_job - issue RPC to a job's configuration per request, |
| * only usable by user root or (for some parameters) the job's owner |
| * IN job_msg - description of job updates |
| * RET SLURM_SUCCESS on success, otherwise return SLURM_ERROR with errno set |
| */ |
| int |
| slurm_update_job (job_desc_msg_t * job_msg) |
| { |
| if (job_msg->job_id_str) { |
| error("Use slurm_update_job2() rather than slurm_update_job() " |
| "with job_msg->job_id_str to get multiple error codes " |
| "for various job array task and avoid memory leaks"); |
| } |
| return _slurm_update ((void *) job_msg, REQUEST_UPDATE_JOB); |
| } |
| |
| /* |
| * slurm_update_job2 - issue RPC to a job's configuration per request, |
| * only usable by user root or (for some parameters) the job's owner |
| * IN job_msg - description of job updates |
| * OUT resp - per task response to the request, |
| * free using slurm_free_job_array_resp() |
| * RET SLURM_SUCCESS on success, otherwise return SLURM_ERROR with errno set |
| */ |
| extern int |
| slurm_update_job2 (job_desc_msg_t * job_msg, job_array_resp_msg_t **resp) |
| { |
| int rc = SLURM_SUCCESS; |
| slurm_msg_t req_msg, resp_msg; |
| slurmdb_cluster_rec_t *save_working_cluster_rec = working_cluster_rec; |
| |
| slurm_msg_t_init(&req_msg); |
| req_msg.msg_type = REQUEST_UPDATE_JOB; |
| req_msg.data = job_msg; |
| |
| tryagain: |
| slurm_msg_t_init(&resp_msg); |
| |
| rc = slurm_send_recv_controller_msg(&req_msg, &resp_msg, |
| working_cluster_rec); |
| switch (resp_msg.msg_type) { |
| case RESPONSE_SLURM_REROUTE_MSG: |
| { |
| reroute_msg_t *rr_msg = (reroute_msg_t *)resp_msg.data; |
| |
| /* Don't expect multiple hops but in the case it does |
| * happen, free the previous rr cluster_rec. */ |
| if (working_cluster_rec && |
| working_cluster_rec != save_working_cluster_rec) |
| slurmdb_destroy_cluster_rec( |
| working_cluster_rec); |
| |
| working_cluster_rec = rr_msg->working_cluster_rec; |
| slurmdb_setup_cluster_rec(working_cluster_rec); |
| rr_msg->working_cluster_rec = NULL; |
| slurm_free_msg_data(resp_msg.msg_type, resp_msg.data); |
| resp_msg.data = NULL; |
| goto tryagain; |
| } |
| case RESPONSE_JOB_ARRAY_ERRORS: |
| *resp = (job_array_resp_msg_t *) resp_msg.data; |
| break; |
| case RESPONSE_SLURM_RC: |
| rc = ((return_code_msg_t *) resp_msg.data)->return_code; |
| if (rc) |
| errno = rc; |
| slurm_free_msg_data(resp_msg.msg_type, resp_msg.data); |
| resp_msg.data = NULL; |
| break; |
| default: |
| errno = SLURM_UNEXPECTED_MSG_ERROR; |
| } |
| |
| if (working_cluster_rec != save_working_cluster_rec) { |
| slurmdb_destroy_cluster_rec(working_cluster_rec); |
| working_cluster_rec = save_working_cluster_rec; |
| } |
| |
| return rc; |
| } |
| |
| int slurm_create_node(update_node_msg_t *node_msg) |
| { |
| return _slurm_update(node_msg, REQUEST_CREATE_NODE); |
| } |
| |
| /* |
| * slurm_update_node - issue RPC to a node's configuration per request, |
| * only usable by user root |
| * IN node_msg - description of node updates |
| * RET SLURM_SUCCESS on success, otherwise return SLURM_ERROR with errno set |
| */ |
| int |
| slurm_update_node ( update_node_msg_t * node_msg) |
| { |
| return _slurm_update ((void *) node_msg, REQUEST_UPDATE_NODE); |
| } |
| |
| /* |
| * slurm_delete_node - issue RPC to delete a node, only usable by user root |
| * IN node_msg - use to pass nodelist of names to delete |
| * RET SLURM_SUCCESS on success, otherwise return SLURM_ERROR with errno set |
| */ |
| int slurm_delete_node(update_node_msg_t *node_msg) |
| { |
| return _slurm_update(node_msg, REQUEST_DELETE_NODE); |
| } |
| |
| /* |
| * slurm_create_partition - create a new partition, only usable by user root |
| * IN part_msg - description of partition configuration |
| * RET SLURM_SUCCESS on success, otherwise return SLURM_ERROR with errno set |
| */ |
| int |
| slurm_create_partition ( update_part_msg_t * part_msg) |
| { |
| return _slurm_update ((void *) part_msg, REQUEST_CREATE_PARTITION); |
| } |
| |
| /* |
| * slurm_update_partition - issue RPC to a partition's configuration per |
| * request, only usable by user root |
| * IN part_msg - description of partition updates |
| * RET SLURM_SUCCESS on success, otherwise return SLURM_ERROR with errno set |
| */ |
| int |
| slurm_update_partition ( update_part_msg_t * part_msg ) |
| { |
| return _slurm_update ((void *) part_msg, REQUEST_UPDATE_PARTITION); |
| } |
| |
| /* |
| * slurm_delete_partition - issue RPC to delete a partition, only usable |
| * by user root |
| * IN part_msg - description of partition to delete |
| * RET SLURM_SUCCESS on success, otherwise return SLURM_ERROR with errno set |
| */ |
| int |
| slurm_delete_partition ( delete_part_msg_t * part_msg ) |
| { |
| return _slurm_update ((void *) part_msg, REQUEST_DELETE_PARTITION); |
| } |
| |
| /* |
| * slurm_create_reservation - create a new reservation, only usable by user root |
| * IN resv_msg - description of reservation |
| * RET name of reservation on success (caller must free the memory), |
| * otherwise return NULL and set errno to indicate the error |
| */ |
| char * |
| slurm_create_reservation (resv_desc_msg_t * resv_msg) |
| { |
| int rc; |
| char *resv_name = NULL; |
| slurm_msg_t req_msg; |
| slurm_msg_t resp_msg; |
| reservation_name_msg_t *resp; |
| |
| slurm_msg_t_init(&req_msg); |
| slurm_msg_t_init(&resp_msg); |
| |
| req_msg.msg_type = REQUEST_CREATE_RESERVATION; |
| req_msg.data = resv_msg; |
| |
| rc = slurm_send_recv_controller_msg(&req_msg, &resp_msg, |
| working_cluster_rec); |
| if (rc) |
| errno = rc; |
| switch (resp_msg.msg_type) { |
| case RESPONSE_CREATE_RESERVATION: |
| resp = (reservation_name_msg_t *) resp_msg.data; |
| if (resp->name) |
| resv_name = strdup(resp->name); |
| break; |
| case RESPONSE_SLURM_RC: |
| rc = ((return_code_msg_t *) resp_msg.data)->return_code; |
| if (rc) |
| errno = rc; |
| break; |
| default: |
| errno = SLURM_UNEXPECTED_MSG_ERROR; |
| } |
| slurm_free_msg_data(resp_msg.msg_type, resp_msg.data); |
| return resv_name; |
| } |
| |
| /* |
| * slurm_update_reservation - modify an existing reservation, only usable by |
| * user root |
| * IN resv_msg - description of reservation |
| * RET SLURM_SUCCESS on success, otherwise return SLURM_ERROR with errno set |
| */ |
| extern int slurm_update_reservation (resv_desc_msg_t * resv_msg) |
| { |
| return _slurm_update ((void *) resv_msg, REQUEST_UPDATE_RESERVATION); |
| } |
| |
| /* |
| * slurm_delete_reservation - issue RPC to delete a reservation, only usable |
| * by user root |
| * IN resv_msg - description of reservation to delete |
| * RET SLURM_SUCCESS on success, otherwise return SLURM_ERROR with errno set |
| */ |
| int |
| slurm_delete_reservation (reservation_name_msg_t * resv_msg) |
| { |
| return _slurm_update ((void *) resv_msg, REQUEST_DELETE_RESERVATION); |
| } |
| |
| /* Update the time limit of a job step, |
| * step_id == NO_VAL updates all job steps of the specified job_id |
| * RET 0 or -1 on error */ |
| int |
| slurm_update_step (step_update_request_msg_t * step_msg) |
| { |
| return _slurm_update ((void *) step_msg, REQUEST_UPDATE_JOB_STEP); |
| } |
| |
| /* |
| * slurm_update_suspend_exc_nodes - issue RPC to set SuspendExcNodes |
| * IN nodes - string to set |
| * IN mode - Whether to set, append or remove nodes from the setting |
| * RET SLURM_SUCCESS on success, otherwise return SLURM_ERROR with errno set |
| */ |
| extern int slurm_update_suspend_exc_nodes(char *nodes, update_mode_t mode) |
| { |
| suspend_exc_update_msg_t req = { .update_str = nodes, .mode = mode }; |
| int rc = SLURM_SUCCESS; |
| |
| /* Check for bad input before sending rpc. */ |
| if (xstrchr(nodes, ':') && (mode != UPDATE_SET)) { |
| error("Append and remove from SuspendExcNodes with ':' is not supported. Please use direct assignment instead."); |
| slurm_seterrno_ret(ESLURM_INVALID_NODE_NAME); |
| return SLURM_SUCCESS; |
| } |
| |
| rc = _slurm_update((void *) &req, REQUEST_SET_SUSPEND_EXC_NODES); |
| |
| /* |
| * We don't know what SuspendExcNodes currently is in the slurmctld. |
| * Give user feedback after rpc. |
| */ |
| if (errno == ESLURM_INVALID_NODE_NAME) |
| error("SuspendExcNodes may have ':' in it. Append and remove are not supported in this case. Please use direct assignment instead."); |
| |
| return rc; |
| } |
| |
| /* |
| * slurm_update_suspend_exc_parts - issue RPC to set SuspendExcParts |
| * IN parts - string to set |
| * IN mode - Whether to set, append or remove partitions from the setting |
| * RET SLURM_SUCCESS on success, otherwise return SLURM_ERROR with errno set |
| */ |
| extern int slurm_update_suspend_exc_parts(char *parts, update_mode_t mode) |
| { |
| suspend_exc_update_msg_t req = { .update_str = parts, .mode = mode }; |
| return _slurm_update((void *) &req, REQUEST_SET_SUSPEND_EXC_PARTS); |
| } |
| |
| /* |
| * slurm_update_suspend_exc_states - issue RPC to set SuspendExcStates |
| * IN states - string to set |
| * IN mode - Whether to set, append or remove nodes from the setting |
| * RET SLURM_SUCCESS on success, otherwise return SLURM_ERROR with errno set |
| */ |
| extern int slurm_update_suspend_exc_states(char *states, update_mode_t mode) |
| { |
| suspend_exc_update_msg_t req = { .update_str = states, .mode = mode }; |
| return _slurm_update((void *) &req, REQUEST_SET_SUSPEND_EXC_STATES); |
| } |
| |
| /* |
| * Move the specified job ID to the top of the queue for a given user ID, |
| * partition, account, and QOS. |
| * IN job_id_str - a job id |
| * RET 0 or -1 on error */ |
| extern int |
| slurm_top_job(char *job_id_str) |
| { |
| int rc = SLURM_SUCCESS; |
| top_job_msg_t top_job_req; |
| slurm_msg_t req_msg; |
| |
| slurm_msg_t_init(&req_msg); |
| memset(&top_job_req, 0, sizeof(top_job_req)); |
| top_job_req.job_id_str = job_id_str; |
| req_msg.msg_type = REQUEST_TOP_JOB; |
| req_msg.data = &top_job_req; |
| |
| if (slurm_send_recv_controller_rc_msg(&req_msg, &rc, |
| working_cluster_rec) < 0) |
| return SLURM_ERROR; |
| |
| errno = rc; |
| return rc; |
| } |
| |
| /* _slurm_update - issue RPC for all update requests */ |
| static int |
| _slurm_update (void *data, slurm_msg_type_t msg_type) |
| { |
| int rc; |
| slurm_msg_t req_msg; |
| |
| slurm_msg_t_init(&req_msg); |
| req_msg.msg_type = msg_type; |
| req_msg.data = data; |
| |
| if (slurm_send_recv_controller_rc_msg(&req_msg, &rc, |
| working_cluster_rec) < 0) |
| return SLURM_ERROR; |
| |
| if (rc != SLURM_SUCCESS) |
| slurm_seterrno_ret(rc); |
| |
| return SLURM_SUCCESS; |
| } |