Elastic Cloud Computing Cluster (EC3) is a tool to create elastic virtual clusters on top of Infrastructure as a Service (IaaS) providers, either public (such as Amazon Web Services, Google Cloud or Microsoft Azure) or on-premises (such as OpenNebula and OpenStack). We offer recipes to deploy TORQUE (optionally with MAUI), SLURM, SGE, HTCondor, Mesos, Nomad and Kubernetes clusters that can be self-managed with CLUES: it starts with a single-node cluster and working nodes will be dynamically deployed and provisioned to fit increasing load (number of jobs at the LRMS). Working nodes will be undeployed when they are idle. This introduces a cost-efficient approach for Cluster-based computing.
The program ec3 requires Python 2.6+, PLY, PyYAML, Requests, jsonschema and an IM server, which is used to launch the virtual machines.
PyYAML is usually available in distribution repositories (python-yaml
in Debian;
PyYAML
in Red Hat; and PyYAML
in pip).
PLY is usually available in distribution repositories (python-ply
and ply
in pip).
Requests is usually available in distribution repositories (python-requests
and requests
in pip).
jsonschema is usually available in distribution repositories (python-jsonschema
and jsonschema
in pip).
By default ec3 uses our public IM server in appsgrycap.i3m.upv.es. Optionally you can deploy a local IM server following the instructions of the IM manual.
Also sshpass
command is required to provide the user with ssh access to the cluster.
First you need to install pip tool. To install them in Debian and Ubuntu based distributions, do:
sudo apt update sudo apt install python-pip
In Red Hat based distributions (RHEL, CentOS, Amazon Linux, Oracle Linux, Fedora, etc.), do:
sudo yum install epel-release sudo yum install which python-pip
Then you only have to call the install command of the pip tool with the ec3-cli package:
sudo pip install ec3-cli
You can also download the last ec3 version from this git repository:
git clone https://github.com/eubr-atmosphere/ec3.git
Then you can install it calling the pip tool with the current ec3 directory:
sudo pip install ./ec3
First create a file auth.dat
with valid credentials to access Fogbow and also create an account for IM:
id = fogbow ; type = FogBow ; host = https://services-atm-prod.lsd.ufcg.edu.br/fns; username = <<your user>> ; password = <<your password>>; domain = LSD type = InfrastructureManager; username = <<your user>>; password = <<your password>>
Replace <<your user>>
and <<your password>>
with the corresponding values
for the Fogbow account where the cluster will be deployed.
This file is the authorization file (see Authorization file), and can have more than one set of credentials.
The next command deploys a Kubernetes cluster based on an Ubuntu image:
$ ec3 launch mycluster kubernetes ubuntu-fbw -a auth.dat -y Creating infrastructure Infrastructure successfully created with ID: 14f6eb82-df9d-11e9-b45a-726c61f3e440 ▄▟▙▄¨ Front-end state: running, IP: 132.43.105.28
If you deployed a local IM server, use the next command instead:
$ ec3 launch mycluster kubernetes ubuntu-fbw -a auth.dat -u http://localhost:8899
This can take several minutes.
Bear in mind that you have to specify a resource manager (like kubernetes
in our example) in addition to the images that you want to deploy (e.g. ubuntu-fbw
). For more information about this check the templates documentation.
You can show basic information about the deployed clusters by executing:
$ ec3 list name state IP nodes --------------------------------------------- mycluster configured 132.43.105.28 0
Once the cluster has been deployed, open a ssh session to the front-end (you may need to install the sshpass
library):
$ ec3 ssh mycluster Welcome to Ubuntu 16.04.6 LTS (GNU/Linux 4.4.0-159-generic x86_64) Documentation: https://help.ubuntu.com/ fogbow@kubeserver:~$
You may use the cluster as usual, depending on the LRMS.
Notice that CLUES will intercept the jobs submited to the LRMS to deploy additional working nodes if needed. This might result in a customizable (180 seconds by default) blocking delay when submitting jobs when no additional working nodes are available. This guarantees that jobs will enter execution as soon as the working nodes are deployed and integrated in the cluster.
Working nodes will be provisioned and relinquished automatically to increase and decrease the cluster size according to the elasticity policies provided by CLUES.
Enjoy your virtual elastic cluster!
EC3 has an official Docker container image available in Docker Hub that can be used instead of installing the CLI. You can download it by typing:
$ sudo docker pull eubraatmosphere/ec3
You can exploit all the potential of EC3 as if you download the CLI and run it on your computer:
$ sudo docker run eubraatmosphere/ec3 list $ sudo docker run eubraatmosphere/ec3 templates
To launch a cluster, you can use the recipes that you have locally by mounting the folder as a volume. Also it is recommendable to mantain the data of active clusters locally, by mounting a volume as follows:
$ sudo docker run -v /home/user/:/tmp/ -v /home/user/ec3/templates/:/etc/ec3/templates -v /tmp/.ec3/clusters:/root/.ec3/clusters eubraatmosphere/ec3 launch mycluster torque ubuntu16 -a /tmp/auth.dat
Notice that you need to change the local paths to the paths where you store the auth file, the templates folder and the .ec3/clusters folder. So, once the front-end is deployed and configured you can connect to it by using:
$ sudo docker run -ti -v /tmp/.ec3/clusters:/root/.ec3/clusters eubraatmosphere/ec3 ssh mycluster
Later on, when you need to destroy the cluster, you can type:
$ sudo docker run -ti -v /tmp/.ec3/clusters:/root/.ec3/clusters eubraatmosphere/ec3 destroy mycluster
- EC3 Command-line Interface.
- Templates.
- Information about available templates:
ec3 templates [--search <topic>] [--full-description]
.