Source code for elasticluster.gc3pie_config

#!/usr/bin/env python
# -*- coding: utf-8 -*-#
# @(#)
# Copyright (C) 2013-2014, 2019 University of Zurich. All rights reserved.
# This program is free software; you can redistribute it and/or modify it
# under the terms of the GNU General Public License as published by the
# Free Software Foundation; either version 2 of the License, or (at your
# option) any later version.
# This program is distributed in the hope that it will be useful, but
# WITHOUT ANY WARRANTY; without even the implied warranty of
# General Public License for more details.
# You should have received a copy of the GNU General Public License along
# with this program; if not, write to the Free Software Foundation, Inc.,
# 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA

__docformat__ = 'reStructuredText'
__author__ = 'Antonio Messina <>'

# compatibility imports
from future import standard_library

# stdlib imports
import re
import sys

if sys.version_info[0] == 2:
    from ConfigParser import RawConfigParser
    from cStringIO import StringIO
    from configparser import RawConfigParser
    from io import StringIO

# local imports
from elasticluster import log

# compute[001-002]      2    cloud*        idle    1    1:1:1    491     5026      1   (null) none
slurm_sinfo_regexp =  re.compile('^(?P<hostnames>[^ \t]+)\s+'
                                 '(?P<partition>[^ \t]+)\s+'
                                 '(?P<state>[^ \t]+)\s+'
                                 '(?P<features>[^ ]+)\s+'
                                 '(?P<reason>[^ \t]+)')

slurm_scontrol_maxtime_regexp = re.compile('.*\sMaxTime=(?P<MaxTime>[^ \t]+)\s+')

def inspect_slurm_cluster(ssh, node_information):
    (_in, _out, _err) = ssh.exec_command("sinfo -hNel")

    nodes = []
    for line in _out:
        match = slurm_sinfo_regexp.match(line)
        if match:
            num_nodes = int('num'))
            num_cores = int('cpus')) * num_nodes
            memory = int('memory')) * num_nodes
            memory_per_core = float('memory')) / num_cores
            nodes.append([num_nodes, num_cores, memory, memory_per_core])
            log.warning("Unable to parse output of sinfo: following line doesn't match node regexp: '%s'" % line.strip())
    node_information['num_nodes'] = sum(i[0] for i in nodes)
    node_information['max_cores'] = sum(i[1] for i in nodes)
    node_information['max_cores_per_job'] = max(i[1] for i in nodes)
    node_information['max_memory_per_core'] = max(i[2] for i in nodes)

    (_in, _out, _err) = ssh.exec_command("scontrol -o show part")
    # Assuming only one partition
    line =
    match = slurm_scontrol_maxtime_regexp.match(line)
    node_information['max_walltime'] = '672hours'
    if match:
        maxtime ='MaxTime')
        if maxtime != 'UNLIMITED':
            node_information['max_walltime'] = maxtime

    return node_information

def inspect_pbs_cluster(ssh):

sge_qhost_regexp = re.compile('(?P<hostname>[^\s]+)\s+'

# This function is took from GC3Pie,
# module gc3pie.gc3libs.utils
[docs]def to_bytes(s): """ Convert string `s` to an integer number of bytes. Suffixes like 'KB', 'MB', 'GB' (up to 'YB'), with or without the trailing 'B', are allowed and properly accounted for. Case is ignored in suffixes. Examples:: >>> to_bytes('12') 12 >>> to_bytes('12B') 12 >>> to_bytes('12KB') 12000 >>> to_bytes('1G') 1000000000 Binary units 'KiB', 'MiB' etc. are also accepted: >>> to_bytes('1KiB') 1024 >>> to_bytes('1MiB') 1048576 """ last = -1 unit = s[last].lower() if unit.isdigit(): # `s` is a integral number return int(s) if unit == 'b': # ignore the the 'b' or 'B' suffix last -= 1 unit = s[last].lower() if unit == 'i': k = 1024 last -= 1 unit = s[last].lower() else: k = 1000 # convert the substring of `s` that does not include the suffix if unit.isdigit(): return int(s[0:(last+1)]) if unit == 'k': return int(float(s[0:last])*k) if unit == 'm': return int(float(s[0:last])*k*k) if unit == 'g': return int(float(s[0:last])*k*k*k) if unit == 't': return int(float(s[0:last])*k*k*k*k) if unit == 'p': return int(float(s[0:last])*k*k*k*k*k) if unit == 'e': return int(float(s[0:last])*k*k*k*k*k*k) if unit == 'z': return int(float(s[0:last])*k*k*k*k*k*k*k) if unit == 'y': return int(float(s[0:last])*k*k*k*k*k*k*k*k)
def inspect_sge_cluster(ssh, node_information): (_in, _out, _err) = ssh.exec_command("qhost") nodes = [] for line in _out: match = sge_qhost_regexp.match(line) if match: nodes.append(('hostname'), int('ncpus')), to_bytes('memory')))) node_information['num_nodes'] = len(nodes) node_information['max_cores'] = sum(i[1] for i in nodes) node_information['max_cores_per_job'] = node_information['max_cores'] node_information['max_memory_per_core'] = max(i[2] for i in nodes) # No easy way to see the maximum walltime for a SGE cluster. We # should run qstat -g c to list the queues, and then run qconf -sq # <queue> and look for s_rt and h_rt node_information['max_walltime'] = '672hours'
[docs]def inspect_node(node): """ This function accept a `elasticluster.cluster.Node` class, connects to a node and tries to discover the kind of batch system installed, and some other information. """ node_information = {} ssh = node.connect() if not ssh: log.error("Unable to connect to node %s", return (_in, _out, _err) = ssh.exec_command("(type >& /dev/null -a srun && echo slurm) \ || (type >& /dev/null -a qconf && echo sge) \ || (type >& /dev/null -a pbsnodes && echo pbs) \ || echo UNKNOWN") node_information['type'] = (_in, _out, _err) = ssh.exec_command("arch") node_information['architecture'] = if node_information['type'] == 'slurm': inspect_slurm_cluster(ssh, node_information) elif node_information['type'] == 'sge': inspect_sge_cluster(ssh, node_information) ssh.close() return node_information
[docs]def create_gc3pie_config_snippet(cluster): """ Create a configuration file snippet to be used with GC3Pie. """ auth_section = 'auth/elasticluster_%s' % resource_section = 'resource/elasticluster_%s' % cfg = RawConfigParser() cfg.add_section(auth_section) frontend_node = cluster.get_ssh_to_node() cfg.set(auth_section, 'type', 'ssh') cfg.set(auth_section, 'username', frontend_node.image_user) cluster_info = inspect_node(frontend_node) cfg.add_section(resource_section) cfg.set(resource_section, 'enabled', 'yes') cfg.set(resource_section, 'transport', 'ssh') cfg.set(resource_section, 'frontend', frontend_node.preferred_ip) if not cluster_info: log.error("Unable to gather enough information from the cluster. " "Following informatino are only partial!") cluster_info = {'architecture': 'unknown', 'type': 'unknown', 'max_cores': -1, 'max_cores_per_job': -1, 'max_memory_per_core': -1, 'max_walltime': '672hours'} cfg.set(resource_section, 'type', cluster_info['type']) cfg.set(resource_section, 'architecture', cluster_info['architecture']) cfg.set(resource_section, 'max_cores', cluster_info.get('max_cores', 1)) cfg.set(resource_section, 'max_cores_per_job', cluster_info.get('max_cores_per_job', 1)) cfg.set(resource_section, 'max_memory_per_core', cluster_info.get('max_memory_per_core', '2GB')) cfg.set(resource_section, 'max_walltime', cluster_info.get('max_walltime', '672hours')) cfgstring = StringIO() cfg.write(cfgstring) return cfgstring.getvalue()