| /*****************************************************************************\ |
| * node_select.c - node selection plugin wrapper. |
| * |
| * NOTE: The node selection plugin itself is intimately tied to slurmctld |
| * functions and data structures. Some related functions (e.g. data structure |
| * un/packing, environment variable setting) are required by most SLURM |
| * commands. Since some of these commands must be executed on the BlueGene |
| * front-end nodes, the functions they require are here rather than within |
| * the plugin. This is because functions required by the plugin can not be |
| * resolved on the front-end nodes, so we can't load the plugins there. |
| ***************************************************************************** |
| * Copyright (C) 2002-2007 The Regents of the University of California. |
| * Copyright (C) 2008-2009 Lawrence Livermore National Security. |
| * Produced at Lawrence Livermore National Laboratory (cf, DISCLAIMER). |
| * Written by Morris Jette <jette1@llnl.gov>. |
| * CODE-OCEC-09-009. All rights reserved. |
| * |
| * This file is part of SLURM, a resource management program. |
| * For details, see <https://computing.llnl.gov/linux/slurm/>. |
| * Please also read the included file: DISCLAIMER. |
| * |
| * SLURM is free software; you can redistribute it and/or modify it under |
| * the terms of the GNU General Public License as published by the Free |
| * Software Foundation; either version 2 of the License, or (at your option) |
| * any later version. |
| * |
| * In addition, as a special exception, the copyright holders give permission |
| * to link the code of portions of this program with the OpenSSL library under |
| * certain conditions as described in each individual source file, and |
| * distribute linked combinations including the two. You must obey the GNU |
| * General Public License in all respects for all of the code used other than |
| * OpenSSL. If you modify file(s) with this exception, you may extend this |
| * exception to your version of the file(s), but you are not obligated to do |
| * so. If you do not wish to do so, delete this exception statement from your |
| * version. If you delete this exception statement from all source files in |
| * the program, then also delete it here. |
| * |
| * SLURM is distributed in the hope that it will be useful, but WITHOUT ANY |
| * WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS |
| * FOR A PARTICULAR PURPOSE. See the GNU General Public License for more |
| * details. |
| * |
| * You should have received a copy of the GNU General Public License along |
| * with SLURM; if not, write to the Free Software Foundation, Inc., |
| * 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. |
| \*****************************************************************************/ |
| |
| #ifdef HAVE_CONFIG_H |
| # include "config.h" |
| #endif |
| |
| #include <pthread.h> |
| #include <dirent.h> |
| |
| #include "src/common/list.h" |
| #include "src/common/slurm_protocol_api.h" |
| #include "src/common/xstring.h" |
| #include "src/slurmctld/slurmctld.h" |
| #include "src/common/node_select.h" |
| |
| static int select_context_cnt = -1; |
| static int select_context_default = -1; |
| /* If there is a new select plugin, list it here */ |
| static slurm_select_context_t * select_context = NULL; |
| static pthread_mutex_t select_context_lock = |
| PTHREAD_MUTEX_INITIALIZER; |
| |
| /* |
| * Locate and load the appropriate plugin |
| */ |
| static int _select_get_ops(char *select_type, |
| slurm_select_context_t *c) |
| { |
| /* |
| * Must be synchronized with slurm_select_ops_t in node_select.h. |
| */ |
| static const char *syms[] = { |
| "plugin_id", |
| "select_p_state_save", |
| "select_p_state_restore", |
| "select_p_job_init", |
| "select_p_node_init", |
| "select_p_block_init", |
| "select_p_job_test", |
| "select_p_job_begin", |
| "select_p_job_ready", |
| "select_p_job_resized", |
| "select_p_job_fini", |
| "select_p_job_suspend", |
| "select_p_job_resume", |
| "select_p_pack_select_info", |
| "select_p_select_nodeinfo_pack", |
| "select_p_select_nodeinfo_unpack", |
| "select_p_select_nodeinfo_alloc", |
| "select_p_select_nodeinfo_free", |
| "select_p_select_nodeinfo_set_all", |
| "select_p_select_nodeinfo_set", |
| "select_p_select_nodeinfo_get", |
| "select_p_select_jobinfo_alloc", |
| "select_p_select_jobinfo_free", |
| "select_p_select_jobinfo_set", |
| "select_p_select_jobinfo_get", |
| "select_p_select_jobinfo_copy", |
| "select_p_select_jobinfo_pack", |
| "select_p_select_jobinfo_unpack", |
| "select_p_select_jobinfo_sprint", |
| "select_p_select_jobinfo_xstrdup", |
| "select_p_update_block", |
| "select_p_update_sub_node", |
| "select_p_get_info_from_plugin", |
| "select_p_update_node_config", |
| "select_p_update_node_state", |
| "select_p_alter_node_cnt", |
| "select_p_reconfigure", |
| }; |
| int n_syms = sizeof( syms ) / sizeof( char * ); |
| |
| c->select_type = xstrdup(select_type); |
| c->plugin_list = NULL; |
| c->cur_plugin = PLUGIN_INVALID_HANDLE; |
| c->select_errno = SLURM_SUCCESS; |
| |
| /* Find the correct plugin. */ |
| c->cur_plugin = plugin_load_and_link(c->select_type, n_syms, syms, |
| (void **) &c->ops); |
| if ( c->cur_plugin != PLUGIN_INVALID_HANDLE ) |
| return SLURM_SUCCESS; |
| |
| if(errno != EPLUGIN_NOTFOUND) { |
| error("Couldn't load specified plugin name for %s: %s", |
| c->select_type, plugin_strerror(errno)); |
| return SLURM_ERROR; |
| } |
| |
| error("Couldn't find the specified plugin name for %s " |
| "looking at all files", |
| c->select_type); |
| |
| /* Get plugin list. */ |
| if ( c->plugin_list == NULL ) { |
| char *plugin_dir; |
| c->plugin_list = plugrack_create(); |
| if ( c->plugin_list == NULL ) { |
| error( "cannot create plugin manager" ); |
| return SLURM_ERROR; |
| } |
| plugrack_set_major_type( c->plugin_list, "select" ); |
| plugrack_set_paranoia( c->plugin_list, |
| PLUGRACK_PARANOIA_NONE, |
| 0 ); |
| plugin_dir = slurm_get_plugin_dir(); |
| plugrack_read_dir( c->plugin_list, plugin_dir ); |
| xfree(plugin_dir); |
| } |
| |
| c->cur_plugin = plugrack_use_by_type( c->plugin_list, c->select_type ); |
| if ( c->cur_plugin == PLUGIN_INVALID_HANDLE ) { |
| error( "cannot find node selection plugin for %s", |
| c->select_type ); |
| |
| return SLURM_ERROR; |
| } |
| |
| /* Dereference the API. */ |
| if ( plugin_get_syms( c->cur_plugin, |
| n_syms, |
| syms, |
| (void **) &c->ops ) < n_syms ) { |
| error( "incomplete node selection plugin detected" ); |
| return SLURM_ERROR; |
| } |
| |
| return SLURM_SUCCESS; |
| } |
| |
| /* |
| * Destroy a node selection context |
| */ |
| static int _select_context_destroy( slurm_select_context_t *c ) |
| { |
| int rc = SLURM_SUCCESS; |
| /* |
| * Must check return code here because plugins might still |
| * be loaded and active. |
| */ |
| if ( c->plugin_list ) { |
| if ( plugrack_destroy( c->plugin_list ) != SLURM_SUCCESS ) { |
| rc = SLURM_ERROR; |
| } |
| } else { |
| plugin_unload(c->cur_plugin); |
| } |
| |
| xfree(c->select_type); |
| |
| return rc; |
| } |
| |
| /* |
| * Initialize context for node selection plugin |
| */ |
| extern int slurm_select_init(bool only_default) |
| { |
| int retval = SLURM_SUCCESS; |
| char *select_type = NULL; |
| int i, j, rc, len; |
| DIR *dirp; |
| struct dirent *e; |
| char *dir_array = NULL, *head = NULL; |
| |
| slurm_mutex_lock( &select_context_lock ); |
| |
| if ( select_context ) |
| goto done; |
| |
| select_type = slurm_get_select_type(); |
| if(working_cluster_rec) { |
| /* just ignore warnings here */ |
| } else { |
| #ifdef HAVE_XCPU |
| if(strcasecmp(select_type, "select/linear")) { |
| error("%s is incompatible with XCPU use", select_type); |
| fatal("Use SelectType=select/linear"); |
| } |
| #endif |
| #ifdef HAVE_BG |
| # ifdef HAVE_BGQ |
| if(strcasecmp(select_type, "select/bgq")) { |
| error("%s is incompatible with BlueGene/Q", |
| select_type); |
| fatal("Use SelectType=select/bgq"); |
| } |
| # else |
| if(strcasecmp(select_type, "select/bluegene")) { |
| error("%s is incompatible with BlueGene", select_type); |
| fatal("Use SelectType=select/bluegene"); |
| } |
| # endif |
| #endif |
| |
| #ifdef HAVE_CRAY |
| if(strcasecmp(select_type, "select/cray")) { |
| error("%s is incompatible with Cray", select_type); |
| fatal("Use SelectType=select/cray"); |
| } |
| #endif |
| } |
| |
| select_context_cnt = 0; |
| if(only_default) { |
| select_context = xmalloc(sizeof(slurm_select_context_t)); |
| rc = _select_get_ops(select_type, select_context); |
| if (rc == SLURM_SUCCESS) { |
| select_context_default = 0; |
| select_context_cnt++; |
| } |
| goto skip_load_all; |
| } |
| |
| if(!(dir_array = slurm_get_plugin_dir())) { |
| error("plugin_load_and_link: No plugin dir given"); |
| goto done; |
| } |
| |
| head = dir_array; |
| for (i=0; ; i++) { |
| bool got_colon = 0; |
| if (dir_array[i] == ':') { |
| dir_array[i] = '\0'; |
| got_colon = 1; |
| } else if(dir_array[i] != '\0') |
| continue; |
| |
| /* Open the directory. */ |
| if(!(dirp = opendir(head))) { |
| error("cannot open plugin directory %s", head); |
| goto done; |
| } |
| |
| while (1) { |
| char full_name[128]; |
| |
| if(!(e = readdir( dirp ))) |
| break; |
| /* Check only files with select_ in them. */ |
| if (strncmp(e->d_name, "select_", 7)) |
| continue; |
| |
| len = strlen(e->d_name)-3; |
| /* Check only shared object files */ |
| if (strcmp(e->d_name+len, ".so")) |
| continue; |
| |
| /* add one for the / */ |
| len++; |
| xassert(len<sizeof(full_name)); |
| snprintf(full_name, len, "select/%s", e->d_name+7); |
| for (j=0; j<select_context_cnt; j++) { |
| if (!strcmp(full_name, |
| select_context[j].select_type)) |
| break; |
| } |
| if (j < select_context_cnt) { |
| error("Duplicate plugin %s ignored", |
| select_context[j].select_type); |
| } else { |
| xrealloc(select_context, |
| (sizeof(slurm_select_context_t) * |
| (select_context_cnt + 1))); |
| |
| rc = _select_get_ops( |
| full_name, |
| select_context + select_context_cnt); |
| |
| /* only add the ones this system has */ |
| if (rc == SLURM_SUCCESS) { |
| /* set the default */ |
| if (!strcmp(full_name, select_type)) |
| select_context_default = |
| select_context_cnt; |
| select_context_cnt++; |
| } |
| } |
| } |
| |
| closedir(dirp); |
| |
| if (got_colon) { |
| head = dir_array + i + 1; |
| } else |
| break; |
| } |
| |
| skip_load_all: |
| if(select_context_default == -1) |
| fatal("Can't find plugin for %s", select_type); |
| |
| /* Insure that plugin_id is valid and unique */ |
| for (i=0; i<select_context_cnt; i++) { |
| for (j=i+1; j<select_context_cnt; j++) { |
| if (*(select_context[i].ops.plugin_id) != |
| *(select_context[j].ops.plugin_id)) |
| continue; |
| fatal("SelectPlugins: Duplicate plugin_id %u for " |
| "%s and %s", |
| *(select_context[i].ops.plugin_id), |
| select_context[i].select_type, |
| select_context[j].select_type); |
| } |
| if (*(select_context[i].ops.plugin_id) < 100) { |
| fatal("SelectPlugins: Invalid plugin_id %u (<100) %s", |
| *(select_context[i].ops.plugin_id), |
| select_context[i].select_type); |
| } |
| |
| } |
| |
| done: |
| slurm_mutex_unlock( &select_context_lock ); |
| xfree(select_type); |
| xfree(dir_array); |
| return retval; |
| } |
| |
| extern int slurm_select_fini(void) |
| { |
| int rc = SLURM_SUCCESS, i, j; |
| |
| slurm_mutex_lock(&select_context_lock); |
| if (!select_context) |
| goto fini; |
| |
| for (i=0; i<select_context_cnt; i++) { |
| j = _select_context_destroy(select_context + i); |
| if (j != SLURM_SUCCESS) |
| rc = j; |
| } |
| xfree(select_context); |
| select_context_cnt = -1; |
| |
| fini: slurm_mutex_unlock(&select_context_lock); |
| return rc; |
| } |
| |
| extern int select_get_plugin_id_pos(uint32_t plugin_id) |
| { |
| int i; |
| |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| for (i=0; i<select_context_cnt; i++) { |
| if(*(select_context[i].ops.plugin_id) == plugin_id) |
| break; |
| } |
| if(i >= select_context_cnt) |
| return SLURM_ERROR; |
| return i; |
| } |
| |
| extern int select_get_plugin_id() |
| { |
| if (slurm_select_init(0) < 0) |
| return 0; |
| |
| return *(select_context[select_context_default].ops.plugin_id); |
| } |
| |
| /* |
| * Save any global state information |
| * IN dir_name - directory into which the data can be stored |
| */ |
| extern int select_g_state_save(char *dir_name) |
| { |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| return (*(select_context[select_context_default].ops.state_save)) |
| (dir_name); |
| } |
| |
| /* |
| * Initialize context for node selection plugin and |
| * restore any global state information |
| * IN dir_name - directory from which the data can be restored |
| */ |
| extern int select_g_state_restore(char *dir_name) |
| { |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| return (*(select_context[select_context_default].ops.state_restore)) |
| (dir_name); |
| } |
| |
| /* |
| * Note the initialization of job records, issued upon restart of |
| * slurmctld and used to synchronize any job state. |
| */ |
| extern int select_g_job_init(List job_list) |
| { |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| return (*(select_context[select_context_default].ops.job_init)) |
| (job_list); |
| } |
| |
| /* |
| * Note re/initialization of node record data structure |
| * IN node_ptr - current node data |
| * IN node_count - number of node entries |
| */ |
| extern int select_g_node_init(struct node_record *node_ptr, int node_cnt) |
| { |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| return (*(select_context[select_context_default].ops.node_init)) |
| (node_ptr, node_cnt); |
| } |
| |
| |
| /* |
| * Note re/initialization of block record data structure |
| * IN block_list - list of partition records |
| */ |
| extern int select_g_block_init(List block_list) |
| { |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| return (*(select_context[select_context_default].ops.block_init)) |
| (block_list); |
| } |
| |
| /* |
| * Select the "best" nodes for given job from those available |
| * IN/OUT job_ptr - pointer to job being considered for initiation, |
| * set's start_time when job expected to start |
| * IN/OUT bitmap - map of nodes being considered for allocation on input, |
| * map of nodes actually to be assigned on output |
| * IN min_nodes - minimum number of nodes to allocate to job |
| * IN max_nodes - maximum number of nodes to allocate to job |
| * IN req_nodes - requested (or desired) count of nodes |
| * IN mode - SELECT_MODE_RUN_NOW: try to schedule job now |
| * SELECT_MODE_TEST_ONLY: test if job can ever run |
| * SELECT_MODE_WILL_RUN: determine when and where job can run |
| * IN preemptee_candidates - List of pointers to jobs which can bee preempted |
| * IN/OUT preemptee_job_list - Pointer to list of job pointers. These are the |
| * jobs to be preempted to initiate the pending job. Not set |
| * if mode=SELECT_MODE_TEST_ONLY or input pointer is NULL. |
| * Existing list is appended to. |
| * RET zero on success, EINVAL otherwise |
| */ |
| extern int select_g_job_test(struct job_record *job_ptr, bitstr_t *bitmap, |
| uint32_t min_nodes, uint32_t max_nodes, |
| uint32_t req_nodes, uint16_t mode, |
| List preemptee_candidates, |
| List *preemptee_job_list) |
| { |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| return (*(select_context[select_context_default].ops.job_test)) |
| (job_ptr, bitmap, |
| min_nodes, max_nodes, |
| req_nodes, mode, |
| preemptee_candidates, |
| preemptee_job_list); |
| } |
| |
| /* |
| * Note initiation of job is about to begin. Called immediately |
| * after select_g_job_test(). Executed from slurmctld. |
| * IN job_ptr - pointer to job being initiated |
| */ |
| extern int select_g_job_begin(struct job_record *job_ptr) |
| { |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| return (*(select_context[select_context_default].ops.job_begin)) |
| (job_ptr); |
| } |
| |
| /* |
| * determine if job is ready to execute per the node select plugin |
| * IN job_ptr - pointer to job being tested |
| * RET: -2 fatal error, -1 try again, 1 if ready to execute, |
| * 0 not ready to execute |
| */ |
| extern int select_g_job_ready(struct job_record *job_ptr) |
| { |
| if (slurm_select_init(0) < 0) |
| return -1; |
| |
| return (*(select_context[select_context_default].ops.job_ready)) |
| (job_ptr); |
| } |
| |
| /* |
| * Modify internal data structures for a job that has changed size |
| * Only support jobs shrinking now. |
| * RET: 0 or an error code |
| */ |
| extern int select_g_job_resized(struct job_record *job_ptr, |
| struct node_record *node_ptr) |
| { |
| if (slurm_select_init(0) < 0) |
| return -1; |
| |
| return (*(select_context[select_context_default].ops.job_resized)) |
| (job_ptr, node_ptr); |
| } |
| |
| /* |
| * Note termination of job is starting. Executed from slurmctld. |
| * IN job_ptr - pointer to job being terminated |
| */ |
| extern int select_g_job_fini(struct job_record *job_ptr) |
| { |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| return (*(select_context[select_context_default].ops.job_fini)) |
| (job_ptr); |
| } |
| |
| /* |
| * Suspend a job. Executed from slurmctld. |
| * IN job_ptr - pointer to job being suspended |
| * RET SLURM_SUCCESS or error code |
| */ |
| extern int select_g_job_suspend(struct job_record *job_ptr) |
| { |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| return (*(select_context[select_context_default].ops.job_suspend)) |
| (job_ptr); |
| } |
| |
| /* |
| * Resume a job. Executed from slurmctld. |
| * IN job_ptr - pointer to job being resumed |
| * RET SLURM_SUCCESS or error code |
| */ |
| extern int select_g_job_resume(struct job_record *job_ptr) |
| { |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| return (*(select_context[select_context_default].ops.job_resume)) |
| (job_ptr); |
| } |
| |
| extern int select_g_pack_select_info(time_t last_query_time, |
| uint16_t show_flags, Buf *buffer, |
| uint16_t protocol_version) |
| { |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| return (*(select_context[select_context_default].ops.pack_select_info)) |
| (last_query_time, show_flags, buffer, protocol_version); |
| } |
| |
| extern int select_g_select_nodeinfo_pack(dynamic_plugin_data_t *nodeinfo, |
| Buf buffer, |
| uint16_t protocol_version) |
| { |
| void *data = NULL; |
| uint32_t plugin_id; |
| |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| if(nodeinfo) { |
| data = nodeinfo->data; |
| plugin_id = nodeinfo->plugin_id; |
| } else |
| plugin_id = select_context_default; |
| |
| if(protocol_version >= SLURM_2_2_PROTOCOL_VERSION) |
| pack32(*(select_context[plugin_id].ops.plugin_id), |
| buffer); |
| |
| return (*(select_context[plugin_id].ops. |
| nodeinfo_pack))(data, buffer, protocol_version); |
| } |
| |
| extern int select_g_select_nodeinfo_unpack(dynamic_plugin_data_t **nodeinfo, |
| Buf buffer, |
| uint16_t protocol_version) |
| { |
| dynamic_plugin_data_t *nodeinfo_ptr = NULL; |
| |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| nodeinfo_ptr = xmalloc(sizeof(dynamic_plugin_data_t)); |
| *nodeinfo = nodeinfo_ptr; |
| |
| if(protocol_version >= SLURM_2_2_PROTOCOL_VERSION) { |
| int i; |
| uint32_t plugin_id; |
| safe_unpack32(&plugin_id, buffer); |
| for (i=0; i<select_context_cnt; i++) |
| if(*(select_context[i].ops.plugin_id) == plugin_id) { |
| nodeinfo_ptr->plugin_id = i; |
| break; |
| } |
| if (i >= select_context_cnt) { |
| error("we don't have this plugin type %u", plugin_id); |
| goto unpack_error; |
| } |
| } else |
| nodeinfo_ptr->plugin_id = select_context_default; |
| |
| return (*(select_context[nodeinfo_ptr->plugin_id].ops.nodeinfo_unpack)) |
| ((select_nodeinfo_t **)&nodeinfo_ptr->data, buffer, |
| protocol_version); |
| unpack_error: |
| error("select_g_select_nodeinfo_unpack: unpack error"); |
| return SLURM_ERROR; |
| } |
| |
| extern dynamic_plugin_data_t *select_g_select_nodeinfo_alloc(uint32_t size) |
| { |
| dynamic_plugin_data_t *nodeinfo_ptr = NULL; |
| uint32_t plugin_id; |
| |
| if (slurm_select_init(0) < 0) |
| return NULL; |
| |
| plugin_id = working_cluster_rec ? |
| working_cluster_rec->plugin_id_select : select_context_default; |
| |
| nodeinfo_ptr = xmalloc(sizeof(dynamic_plugin_data_t)); |
| nodeinfo_ptr->plugin_id = plugin_id; |
| nodeinfo_ptr->data = (*(select_context[plugin_id].ops. |
| nodeinfo_alloc))(size); |
| return nodeinfo_ptr; |
| } |
| |
| extern int select_g_select_nodeinfo_free(dynamic_plugin_data_t *nodeinfo) |
| { |
| int rc = SLURM_SUCCESS; |
| |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| if(nodeinfo) { |
| if(nodeinfo->data) |
| rc = (*(select_context[nodeinfo->plugin_id].ops. |
| nodeinfo_free))(nodeinfo->data); |
| xfree(nodeinfo); |
| } |
| return rc; |
| } |
| |
| extern int select_g_select_nodeinfo_set_all(time_t last_query_time) |
| { |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| return (*(select_context[select_context_default].ops.nodeinfo_set_all)) |
| (last_query_time); |
| } |
| |
| extern int select_g_select_nodeinfo_set(struct job_record *job_ptr) |
| { |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| return (*(select_context[select_context_default].ops.nodeinfo_set)) |
| (job_ptr); |
| } |
| |
| extern int select_g_select_nodeinfo_get(dynamic_plugin_data_t *nodeinfo, |
| enum select_nodedata_type dinfo, |
| enum node_states state, |
| void *data) |
| { |
| void *nodedata = NULL; |
| uint32_t plugin_id; |
| |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| if(nodeinfo) { |
| nodedata = nodeinfo->data; |
| plugin_id = nodeinfo->plugin_id; |
| } else |
| plugin_id = select_context_default; |
| |
| return (*(select_context[plugin_id].ops.nodeinfo_get)) |
| (nodedata, dinfo, state, data); |
| } |
| |
| extern dynamic_plugin_data_t *select_g_select_jobinfo_alloc(void) |
| { |
| dynamic_plugin_data_t *jobinfo_ptr = NULL; |
| uint32_t plugin_id; |
| |
| if (slurm_select_init(0) < 0) |
| return NULL; |
| |
| plugin_id = working_cluster_rec ? |
| working_cluster_rec->plugin_id_select : select_context_default; |
| |
| jobinfo_ptr = xmalloc(sizeof(dynamic_plugin_data_t)); |
| jobinfo_ptr->plugin_id = plugin_id; |
| jobinfo_ptr->data = (*(select_context[plugin_id].ops. |
| jobinfo_alloc))(); |
| return jobinfo_ptr; |
| } |
| |
| /* free storage previously allocated for a select job credential |
| * IN jobinfo - the select job credential to be freed |
| */ |
| extern int select_g_select_jobinfo_free(dynamic_plugin_data_t *jobinfo) |
| { |
| int rc = SLURM_SUCCESS; |
| |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| if(jobinfo) { |
| if(jobinfo->data) |
| rc = (*(select_context[jobinfo->plugin_id].ops. |
| jobinfo_free))(jobinfo->data); |
| xfree(jobinfo); |
| } |
| return rc; |
| } |
| |
| extern int select_g_select_jobinfo_set(dynamic_plugin_data_t *jobinfo, |
| enum select_jobdata_type data_type, |
| void *data) |
| { |
| void *jobdata = NULL; |
| uint32_t plugin_id; |
| |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| if(jobinfo) { |
| jobdata = jobinfo->data; |
| plugin_id = jobinfo->plugin_id; |
| } else |
| plugin_id = select_context_default; |
| |
| return (*(select_context[plugin_id].ops.jobinfo_set)) |
| (jobdata, data_type, data); |
| } |
| |
| /* get data from a select job credential |
| * IN jobinfo - updated select job credential |
| * IN data_type - type of data to enter into job credential |
| * IN/OUT data - the data to enter into job credential |
| */ |
| extern int select_g_select_jobinfo_get(dynamic_plugin_data_t *jobinfo, |
| enum select_jobdata_type data_type, |
| void *data) |
| { |
| void *jobdata = NULL; |
| uint32_t plugin_id; |
| |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| if(jobinfo) { |
| jobdata = jobinfo->data; |
| plugin_id = jobinfo->plugin_id; |
| } else |
| plugin_id = select_context_default; |
| |
| return (*(select_context[plugin_id].ops.jobinfo_get)) |
| (jobdata, data_type, data); |
| } |
| |
| /* copy a select job credential |
| * IN jobinfo - the select job credential to be copied |
| * RET - the copy or NULL on failure |
| * NOTE: returned value must be freed using select_g_free_jobinfo |
| */ |
| extern dynamic_plugin_data_t *select_g_select_jobinfo_copy( |
| dynamic_plugin_data_t *jobinfo) |
| { |
| dynamic_plugin_data_t *jobinfo_ptr = NULL; |
| if (slurm_select_init(0) < 0) |
| return NULL; |
| |
| jobinfo_ptr = xmalloc(sizeof(dynamic_plugin_data_t)); |
| if(jobinfo) { |
| jobinfo_ptr->plugin_id = jobinfo->plugin_id; |
| jobinfo_ptr->data = (*(select_context[jobinfo->plugin_id].ops. |
| jobinfo_copy))(jobinfo->data); |
| } else |
| jobinfo_ptr->plugin_id = select_context_default; |
| |
| return jobinfo_ptr; |
| } |
| |
| /* pack a select job credential into a buffer in machine independent form |
| * IN jobinfo - the select job credential to be saved |
| * OUT buffer - buffer with select credential appended |
| * RET - slurm error code |
| */ |
| extern int select_g_select_jobinfo_pack(dynamic_plugin_data_t *jobinfo, |
| Buf buffer, |
| uint16_t protocol_version) |
| { |
| void *data = NULL; |
| uint32_t plugin_id; |
| |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| if(jobinfo) { |
| data = jobinfo->data; |
| plugin_id = jobinfo->plugin_id; |
| } else |
| plugin_id = select_context_default; |
| |
| if(protocol_version >= SLURM_2_2_PROTOCOL_VERSION) |
| pack32(*(select_context[plugin_id].ops.plugin_id), |
| buffer); |
| return (*(select_context[plugin_id].ops. |
| jobinfo_pack))(data, buffer, protocol_version); |
| } |
| |
| /* unpack a select job credential from a buffer |
| * OUT jobinfo - the select job credential read |
| * IN buffer - buffer with select credential read from current pointer loc |
| * RET - slurm error code |
| * NOTE: returned value must be freed using select_g_free_jobinfo |
| */ |
| extern int select_g_select_jobinfo_unpack(dynamic_plugin_data_t **jobinfo, |
| Buf buffer, |
| uint16_t protocol_version) |
| { |
| dynamic_plugin_data_t *jobinfo_ptr = NULL; |
| |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| jobinfo_ptr = xmalloc(sizeof(dynamic_plugin_data_t)); |
| *jobinfo = jobinfo_ptr; |
| |
| if(protocol_version >= SLURM_2_2_PROTOCOL_VERSION) { |
| int i; |
| uint32_t plugin_id; |
| safe_unpack32(&plugin_id, buffer); |
| for (i=0; i<select_context_cnt; i++) |
| if(*(select_context[i].ops.plugin_id) == plugin_id) { |
| jobinfo_ptr->plugin_id = i; |
| break; |
| } |
| if (i >= select_context_cnt) { |
| error("we don't have this plugin type %u", plugin_id); |
| goto unpack_error; |
| } |
| } else |
| jobinfo_ptr->plugin_id = select_context_default; |
| |
| return (*(select_context[jobinfo_ptr->plugin_id].ops.jobinfo_unpack)) |
| ((select_jobinfo_t **)&jobinfo_ptr->data, buffer, |
| protocol_version); |
| unpack_error: |
| select_g_select_jobinfo_free(jobinfo_ptr); |
| *jobinfo = NULL; |
| error("select_g_select_jobinfo_unpack: unpack error"); |
| return SLURM_ERROR; |
| } |
| |
| /* write select job credential to a string |
| * IN jobinfo - a select job credential |
| * OUT buf - location to write job credential contents |
| * IN size - byte size of buf |
| * IN mode - print mode, see enum select_print_mode |
| * RET - the string, same as buf |
| */ |
| extern char *select_g_select_jobinfo_sprint(dynamic_plugin_data_t *jobinfo, |
| char *buf, size_t size, int mode) |
| { |
| void *data = NULL; |
| uint32_t plugin_id; |
| |
| if (slurm_select_init(0) < 0) |
| return NULL; |
| if(jobinfo) { |
| data = jobinfo->data; |
| plugin_id = jobinfo->plugin_id; |
| } else |
| plugin_id = select_context_default; |
| |
| return (*(select_context[plugin_id].ops. |
| jobinfo_sprint)) |
| (data, buf, size, mode); |
| } |
| /* write select job info to a string |
| * IN jobinfo - a select job credential |
| * IN mode - print mode, see enum select_print_mode |
| * RET - char * containing string of request |
| */ |
| extern char *select_g_select_jobinfo_xstrdup( |
| dynamic_plugin_data_t *jobinfo, int mode) |
| { |
| void *data = NULL; |
| uint32_t plugin_id; |
| |
| if (slurm_select_init(0) < 0) |
| return NULL; |
| |
| if(jobinfo) { |
| data = jobinfo->data; |
| plugin_id = jobinfo->plugin_id; |
| } else |
| plugin_id = select_context_default; |
| |
| return (*(select_context[plugin_id].ops. |
| jobinfo_xstrdup))(data, mode); |
| } |
| |
| /* |
| * Update specific block (usually something has gone wrong) |
| * IN block_desc_ptr - information about the block |
| */ |
| extern int select_g_update_block (update_block_msg_t *block_desc_ptr) |
| { |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| return (*(select_context[select_context_default].ops. |
| update_block))(block_desc_ptr); |
| } |
| |
| /* |
| * Update specific sub nodes (usually something has gone wrong) |
| * IN block_desc_ptr - information about the block |
| */ |
| extern int select_g_update_sub_node (update_block_msg_t *block_desc_ptr) |
| { |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| return (*(select_context[select_context_default].ops. |
| update_sub_node))(block_desc_ptr); |
| } |
| |
| /* |
| * Get select data from a plugin |
| * IN dinfo - type of data to get from the node record |
| * (see enum select_plugindata_info) |
| * IN/OUT data - the data to get from node record |
| */ |
| extern int select_g_get_info_from_plugin (enum select_plugindata_info dinfo, |
| struct job_record *job_ptr, |
| void *data) |
| { |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| return (*(select_context[select_context_default].ops. |
| get_info_from_plugin))(dinfo, job_ptr, data); |
| } |
| |
| /* |
| * Updated a node configuration. This happens when a node registers with |
| * more resources than originally configured (e.g. memory). |
| * IN index - index into the node record list |
| * RETURN SLURM_SUCCESS on success || SLURM_ERROR else wise |
| */ |
| extern int select_g_update_node_config (int index) |
| { |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| return (*(select_context[select_context_default].ops. |
| update_node_config))(index); |
| } |
| |
| /* |
| * Updated a node state in the plugin, this should happen when a node is |
| * drained or put into a down state then changed back. |
| * IN index - index into the node record list |
| * IN state - state to update to |
| * RETURN SLURM_SUCCESS on success || SLURM_ERROR else wise |
| */ |
| extern int select_g_update_node_state (int index, uint16_t state) |
| { |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| return (*(select_context[select_context_default].ops.update_node_state)) |
| (index, state); |
| } |
| |
| /* |
| * Alter the node count for a job given the type of system we are on |
| * IN/OUT job_desc - current job desc |
| */ |
| extern int select_g_alter_node_cnt (enum select_node_cnt type, void *data) |
| { |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| if (type == SELECT_GET_NODE_SCALING) { |
| /* default to one, so most plugins don't have to */ |
| uint32_t *nodes = (uint32_t *)data; |
| *nodes = 1; |
| } |
| return (*(select_context[select_context_default].ops.alter_node_cnt)) |
| (type, data); |
| } |
| |
| /* |
| * Note reconfiguration or change in partition configuration |
| */ |
| extern int select_g_reconfigure (void) |
| { |
| if (slurm_select_init(0) < 0) |
| return SLURM_ERROR; |
| |
| return (*(select_context[select_context_default].ops.reconfigure))(); |
| } |
| |
| |