The fact is, Adelmo's death has caused much spiritual unease among my flock.
For a role based version check Flock2
Flock is an infrastructure prototype engine. You can use it to install and setup arbitrary large infrastructures from scratch with a laptop. All you need is some basic CLI tools, Ansible and Internet connection. It was tested in OS X but Linux should work as well.
Virtual environments are built in VirtualBox. The goal is to keep virtual (test) and production systems as close as possible. It is also easy to make CloudStack templates from VirtualBox images.
Anisble is chosen as configuration manager since it is dead easy, very intuitive and superior to other alternatives (aka Leave chefs in the kitchen alone!).
Currently, CloudStack is supported as a cloud backend. The Ansible part is cloud agnostic.
"With marrying it with something like cobbler you can in theory rack mount a machine, hit the power button and have a machine exactly how you want it to be in “x” about of time. This is great for big companies that always want machines to cookie cutter." jjasghar.github.io True story, but it is great for everyone not just big companies. that is exactly what Flock can do for you.
Install Homebrew and Ansible and the following packages. Do not use ansible development branch! Optionally, you should install Cloud Monkey to hack Cloudstack.
brew install nginx dnsmasq
Edit your .profile or .bash_profile and set PATH:
PATH=/usr/local/bin:$PATH
Install flock:
cd; git clone git://github.com/hornos/flock.git
Install VirtualBox with the extension pack.
Edit your .profile or .bash_profile and login again:
source $HOME/flock/flockrc
Mind that flock always works relative to the current directory so make your moves in the flock directory:
pushd flock
You need some initial step after install eg. generate SSH keys:
flock init
Keys and certificates are in the keys directory. By default all your operations are done by the sysop user on the remote machines.
Ansible host inventories and Cloud Monkey config files are kept under the inventory directory with names .ansible and .cmonkey, respectively. To change Ansbile or Cloud Monkey inventory:
flock on <HOSTS>
stack on <STACK>
or in one command:
flack on <HOSTS>@<STACK>
where <HOSTS> and <STACK> are the suffix truncated basename of the inventory file. Show actual inventory by inventory, list inventories by {flock,stack} ls [<INV>].
If you use RVM you can set prompt indicators, check profile.
flock - Ansible wrapper
cacert - Simple CA manager
vbox - VirtualBox wrapper
ovpn - OpenVPN wrapper
stack - Cloud Monkey wrapper
jockey - Bootp wrapper
cmonkey - Inventory aware Cloud Monkey CLI
Flock playbooks are never general. You should create customized playbook trees for production systems. Please keep the flock tree intact and create a new directory for your needs. Since flock commands are realtive you can use any directory, eg. name system directories according to the domain name.
Download syslinux 4.X and copy the following files to space/boot:
core/pxelinux.0
com32/mboot/mboot.c32
Download install images eg. for Debian (mind the trailing slash!):
pushd space/boot
rsync -avP ftp.us.debian.org::debian/dists/wheezy/main/installer-amd64/current/images/netboot/ ./wheezy
popd
Warning: Debian-based systems should be installed with NAT and not fully supported by all playbooks!
If you need professional stuff use eg. CentOS (mind the trailing slash!):
pushd space/boot
rsync -avP rsync.hrz.tu-chemnitz.de::ftp/pub/linux/centos/6.4/os/x86_64/isolinux/ ./centos64
popd
From the CoreOS PXE howto:
mkdir space/boot/coreos
pushd space/boot/coreos
curl http://storage.core-os.net/coreos/amd64-generic/72.0.0/coreos_production_pxe.vmlinuz > vmlinuz
curl http://storage.core-os.net/coreos/amd64-generic/72.0.0/coreos_production_pxe_image.cpio.gz > initrd.gz
Space Jockey (jockey) is a simple Cobbler replacement. You need an inventory (space/hosts) file with the bootp/dhcp parameters:
boot_server=10.1.1.254
dhcp_range="10.1.1.1,10.1.1.128,255.255.255.0,6h"
interface=eth1
The boot server listens on the boot_server IP address. Debian-based systems use the interface interface to reach the internet. DNSmasq DHCP allocates IPs from the dhcp_range.
Start bootp provision servers by (each in a separate terminal):
flock boot
(open a new terminal)
flock http
Terminate servers by pressing Ctrl-C.
Go to Install Core Server
Network can be different in the cloud. Usually, eth0 is connected to the Internet and eth1 is for internal connections:
Network | Inerface | IPv4 Addr | Mask | DHCP
-------------------------------------------------
external | eth0 | CS | CS | on
system | eth1 | off
You can create a template on Cloudstack or in your VirtualBox. Create a template machine (for CentOS with a 20GB disk):
vbox template centos-template
Create an Ansible inventory (.inventory/template.ansible):
[template]
centos-template ansible_ssh_host=10.1.1.10 ansible_connection=paramiko
Bootstrap the machine:
flock kick centos64-template @centos-template 10.1.1.10 centos-template
flock kick precise-template @ubuntu-template 10.1.1.11 ubuntu-template
Start the servers and the machine:
flock http
flock boot
vbox start centos-template
Switch off bootp and restart the machine:
vbox cycle centos-template with disk
Switch to the template inventory and start cheffing :)
flock on template
flock secure /template
vbox snap centos-template secure
Optionally, at this point you can prepare the template for upload:
flock play @@template minimal-template
Or reach the ground state:
flock play @@template ground
flock reboot @@template
vbox snap centos-template ground
Create a template. Mind that network, firewall resets! Also mind that SSH port is allowed without any FW restriction, and after VM create you should refine ipset/shorewall! You might have to change NTP settings as well. Fot the xenguest playbook you have to copy xen-guest-utilities*.rpm into the rpms directory. TBD multicast change
flock play @@template ground-template
flock shutdown @@template
vbox snap centos-template template
Clone the machine and start upload server:
vbox upload centos-template <ALLOW>
where <ALLOW> is an ngnix allow rule range or address (eg. 192.168.1.0/24).
Switch to your Cloud Monkey inventory by stack on <INVENTORY> and upload/register the YOLO:
set display table
list ostypes
list zones
register template displaytext=template-test format=VHD hypervisor=XenServer name=template-test ostypeid=<OSTYPEID> url=<URL TO VHD> zoneid=<ZONEID>
You might have to create an isolated network. You need the following ids of: zone, template, compute, disk offering, affinity and network:
list zones
list templates templatefilter=community
list serviceofferings
list diskofferings
list networkofferings
deploy virtualmachine name=test displayname=test zoneid=<ZONEID> templateid=<TEMPID> serviceofferingid=<SERVICEID> diskofferingid=<DISKID> networkids=<INTERNETID>,<INTERNALID>
The first card is the default and should be connected to the Internet. You can extend cmonkey inventory with default values and do the YOLO. Add a new section:
[defaultvm]
ostypeid = <OSTYPEID>
zoneid = <ZONEID>
templateid = <TEMPID>
serviceofferingid = <SERVICEID>
networkids = <INTERNETID>,<INTERNALID>
diskofferingid=<DISKID> is optional. Stack out the flock:
stack out 3 core
Create ansible inventory:
stack inventory <NAME>
flock on <NAME>
Profit and happy cheffing :)
flock ping @@core
flock setup @@core
flock command @@core hostname
flock ssh @@<HOST>
Fix ipset whitelist without system network:
flock play @@core roles/firewall/whitelist.yml --extra-vars="is_nosys=true"
TBD genders pssh TBD initial reset dyndns TBD ansible shell
The following network topology is used in the VirtualBox environment. You have to use vboxnet0 as eth0 since bootp works only on the first interface. For a production or a cloud environment you have to exchange the two interface (see above).
Network | VBox Net | IPv4 Addr | Mask | DHCP
-------------------------------------------------
system | vboxnet0 | 10.1.1.254 | 16 | off
external | NAT/Bridged |
Create 3 VMs (aka triangle):
flock out 3 core centos64
Start the boot servers on your OS X host:
flock http
flock boot
Start the machines in headless mode (/ - means group, @ - headless mode):
vbox start /@core
wait for the reboot signal and turn off the group. Mind that the installation is kickstart based, no user interventionis needed. The initial root password is auto-generated by the flock out command.
vbox off /core
Change the boot device to disk and start:
vbox boot /core disk
vbox start /@core
With the snap command you can snapshot the VM:
vbox snap /core init
Change the inventory for the Ansible steps:
flock on core
Bootstrap the flock. The subsequent steps are based on Ansible and you might have to delete old hostkey lines in $HOME/.ssh/known_hosts.
flock bootstrap /core
The bootstrap process installs the sysop administrator user. The SSH key generated by the flock init command is used for the sysop user. Host strings without a user default to sysop. Examples:
flock ping @core - ping cores by sysop
flock ping me@core - ping cores by me
flock ping me@@core - ping cores by me with sudo
flock ping @@@core - ping cores by sysop with sudo and ask for the sudo password
Verify with ping or setup:
flock ping @@core
flock ping @@setup
The networks.yml is a central network topology file, you should link in other playbook directories as well. It is possible to set interfaces in a delicate way. The interfaces structure defines pseudo-interfaces while the path structure defines the real ones (eg. good for bonding). Example:
paths:
eth0: eth0
ansible_paths:
eth0: 'ansible_eth0'
interfaces:
bmc: eth0
In playbooks or Jinja templates use the following statements:
{{paths[interfaces.bmc]}} -> eth0
{{ansible_paths[interfaces.bmc]}} -> ansible_eth0
The goal of the basic provisioning is to provide a good enough base for service roles. First you have to secure the group by basic hardening (TBD intermediate hardening: netlog, dresden, snoopy).
flock play @@core secure
flock reboot @@core
vbox snap /core secure
The ground state is the 2nd step of basic provisioning. It contains several useful stuff and setups basic clustering:
flock play @@core ground
flock reboot @@core
vbox snap /core ground
Mind that the system network is not protected! Due to performance reasons core servers can reach each other wihtout restriction or authentication. Cluster monitor is Ganglia in multicast mode. Remote logging is syslog-ng in multicast mode.
Check the cluster state at http://10.1.1.1/ganglia or by:
gstat -a
If you install PCP (pcp and pcp-gui) from ftp://oss.sgi.com/projects/pcp/download/mac/ you can get realtime statistics:
pmstat -h 10.1.1.1 -h 10.1.1.2 -h 10.1.1.3
/Applications/pmchart.app/Contents/MacOS/pmchart -h 10.1.1.1 -c Overview
Mind that the firewall is also more open for the so called operator hosts (your laptop) on the system network.
Service roles are additional fetures on top of the ground state.
Install the certificate utilities and Globus on your mac:
make globus_simple_ca globus_gsi_cert_utils
There is a hash mismatch between OpenSSL 0.9 and 1.X. Install newer OpenSSL on your mac. You can use the NCE module/package manager. Load the Globus and the new OpenSSL environment:
module load globus openssl
The Grid needs a PKI, which protects access and the communication. You can create as many CA as you like. It is advised to make many short-term flat CAs. Edit grid scripts as well as templates in share/globus_simple_ca if you want to change key parameters. Create a Core CA:
cacert create coreca 365 sysop@localhost
The new CA is created under the ca/coreca directory. The CA certificate is installed under ca/grid-security to make requests easy. If you compile Globus with the old OpenSSL (system default) you have to use old-style subject hash. Create old CA hash by:
cacert oldhash
Edit coreca/grid-ca-ssl.conf and add the following line under policy in CA_default section, this enables extension copy on sign and let alt names go.
copy_extensions = copy
Request & sign host certificates:
for i in 1 2 3 ; do cacert host coreca core-0$i; done
for i in 1 2 3 ; do cacert sign coreca core-0$i; done
Certs, private keys and requests are in ca/coreca/grid-security. There is also a ca/<CAHASH> directory link for each CA. You have to use the <CAHASH> in the playbooks. Get the <CAHASH>:
cacert cahash coreca
Edit roles/globus/vars/globus.yml and set the default CA hash.
Create and sign the sysop certificate:
cacert user coreca sysop "System Operator"
cacert sign coreca sysop
In order to use sysop as a default grid user you have to copy cert and key into the keys directory:
cacert keys coreca sysop
Create a pkcs12 version if you need for the browser (this command works in the keys directory):
cacert p12 sysop
Test your user certificate (you might have to create the old hash):
cacert verify coreca sysop
Install basic grid feature:
flock play @@core grid
Check ssl.conf for a strong PFS cipher setting.
If you want to enable the CA certificate system-wide run:
/root/bin/enable_grid_cert
In production add the ip parameter as well:
cacert host coreca <FQDN> -ip <IP>
cacert sign coreca <FQDN>
flock play @@<FQDN> grid
Install database:
flock play @@core database
Login to the master node and secure the installation:
mysql_secure_installation
flock play @@core roles/database/percona --extra-vars "master=core-01"
Login to the master node and bootstrap the cluster:
/etc/init.d/mysql start --wsrep-cluster-address="gcomm://"
mysql_secure_installation
Now start the whole cluster:
flock play @@core roles/adatabase/percona_start --extra-vars "master=core-01"
Verify:
echo "show status like 'wsrep%'" | mysql -u root -p
Install administrator interface:
flock play @@core roles/database/admin
Change to the latest mainline kernel:
flock play @@core roles/system/kernel
flock reboot @@core
Install Gluster and setup a 3-node FS cluster:
flock play @@core roles/hpc/gluster
Login to the master node and bootstrap the cluster:
/root/gluster_bootstrap
Finally, mount the common directory:
flock play @@core roles/hpc/glusterfs
flock play @@core roles/hpc/gtop
Monitor the cluster:
/root/bin/gtop
Generate the Munge auth key:
dd if=/dev/random bs=1 count=1024 > keys/munge.key
Install Slurm:
flock play @@core scheduler --extra-vars "master=core-01"
Login to the master node and test the queue:
srun -N 3 hostname
Download Easy RSA CA:
git clone git://github.com/OpenVPN/easy-rsa.git ca/easy-rsa
Create a VPN CA (vpnca):
ovpn create
Create server certificates:
for i in 1 2 3 ; do ovpn server vpnca core-0$i; done
Create sysop client certificate:
ovpn client vpnca sysop
Create DH and TA parameters:
ovpn param vpnca
You need the following files:
| Filename | Needed By | Purpose | Secret |
|---|---|---|---|
| ca.crt | server & clients | Root CA cert | NO |
| ca.key | sysop | Root CA key | YES |
| ta.key | server & clients | HMAC | YES |
| dh{n}.pem | server | DH parameters | NO |
| server.crt | server | Server Cert | NO |
| server.key | server | Server Key | YES |
| client.crt | client | Client Cert | NO |
| client.key | client | Client Key | YES |
Install OpenVPN servers:
flock play @@core roles/vpn/openvpn
Install Tunnelblick on your mac and link:
pushd $HOME
ln -s 'Library/Application Support/Tunnelblick/Configurations' .openvpn
popd
Install the sysop certificate for Tunnelblick:
ovpn blick vpnca sysop
Prepare VPN configuration for iPhone:
ovpn client vpnca iphone
ovpn iphone
Edit ca/vpnca/iphone/iphone.ovpn and connect iPhone go to DEVICES/Phone Apps tab and File Sharing section. Select OpenVPN and Add all the files in the iphone directory.
You can determine server IPs by Ting. Register a free pusher.com account. Generate a key:
openssl rand -base64 32 > keys/ting.key
Create keys/ting.yml with the following content:
ting:
key: ...
secret: ...
app_id: ...
Enable the role:
flock play @@core roles/vpn/ting
Install Ting on your mac:
pushd $HOME
git clone git://github.com/hornos/ting.git
Edit the ting.yml and fill in your app details.
The ting service will pong back the machines external IP ans SSH host fingerprints. On your local machine start the monitor:
pushd $HOME/ting
./ting -k ../flock/keys/ting.key -m client
Ping them
./ting -k ../flock/keys/ting.key ping
Hosts are collected in ting/hosts as json files. Create an Tunnelblick client configuration based on Ting pongs:
pushd $HOME/flock
flock-vpn ting core? core
Now connect with Tunnelblick.
Warewulf is an easy to use HPC kit comparable to proprietary HPC solutions. You need at least 3 manager nodes since the minimal size of a Percona cluster is 3. The scheduler (Slurm) works in a master-slave failover, so you can reserve the 3rd machine for other admin related tasks.
Create the controller tringle:
flock out 3 manager
Start the kickstart servers:
flock http
flock boot
and start the group in the background:
vbox start /@manager
wait for the reboot signal and turn off the group:
vbox off /manager
switch to disk boot make a snapshot and start (now in interactive mode):
vbox boot /manager disk
vbox start /manager
Swith to the manager nevironment:
flock on manager
Check the inventory by:
inventory
Delete SSH host keys in $HOME/.ssh/known_hosts if you have old ones. Bootstrap the machines (pre-generated password will be displayed):
flock bootstrap /manager
Verify by sysop:
flock ping @@manager
Check the network topology in networks.yml and secure the flock:
flock secure /manager
(wait for reboot)
vbox snap /manager secure
Your machines restarted and saved in a basic secure state.
Reach the common ground state:
flock play @@manager ground
flock reboot @@manager
Verify the ground state monitoring:
http://10.1.1.1/ganglia
http://10.1.1.1/phpsysinfo
You can also get live monitoring with PCP console or GUI:
pmstat -h 10.1.1.1
pmchart -h 10.1.1.1 -c Overview
Check the boot log:
flock bootlog @@manager
Check syslog in /var/log/loghost/manager-0{1,2,3}. Save:
vbox snap /manager ground
Change to the mainline kernel:
flock play @@manager kernel --extra-vars "clean=true"
flock reboot @@manager
vbox snap /manager kernel
Create a HPC CA:
module load openssl globus/5.2.0
cacert create hpctest
The new CA is created under the ca/hpctest directory. The CA certificate is installed under ca/grid-security. If you compile Globus with the old OpenSSL (system default) you have to use old-style subject hash. Create old CA hash by:
cacert oldhash
Edit ca/hpctest/grid-ca-ssl.conf and add the following line under policy in [CA_default] section, this enables extension copy on sign and let alt names go.
copy_extensions = copy
Request & sign host certificates for the manager nodes:
flock cert hpctest /manager
Certs, private keys and requests are in ca/hpctest/grid-security. There is also a ca/<CAHASH> directory link for each CA. You have to use the <CAHASH> in the playbooks. Get the <CAHASH>:
cacert cahash hpctest
Call the role with --extra-vars="defaultca=<CAHASH>".
Create and sign the sysop certificate:
cacert newuser hpctest sysop "System Operator"
In order to use sysop as a default grid user you have to copy cert and key into the keys directory:
cacert keys hpctest sysop
Test your user certificate (you might have to create the old hash):
cacert verify hpctest sysop
HERE Enable the Grid state: https://www.insecure.ws/2013/10/11/ssltls-configuration-for-apache-mod_ssl/
flock play @@manager roles/globus/globus --extra-vars="defaultca=<CAHASH>"
Check ssl.conf for a strong PFS cipher setting.
Verify https in your browser:
https://10.1.1.1/phpsysinfo
https://10.1.1.1/ganglia
Save:
vbox snap /manager globus
Name the master and backup node for a failover HA:
flock play @@manager roles/hpc/hosts --extra-vars=\"manager=manager-01 backup=manager-02 interface=eth0\"
vbox snap /manager hosts
SQL database is used for the scheduler backend. Install the SQL cluster:
flock play @@manager roles/database/percona --extra-vars "master=manager-01"
Login to the master node and bootstrap the cluster (as root):
/root/percona_bootstrap
Now start the whole cluster:
flock play @@manager roles/database/percona_start --extra-vars "master=manager-01"
Verify on the master node:
echo "show status like 'wsrep%'" | mysql -u root -p
Enable php admin interface:
flock play @@manager roles/database/admin
The mysql admin page is at http://10.1.1.1/phpmyadmin.
vbox snap /manager percona
Stop the machines and install a state disk (/dev/sdb):
flock shutdown @@manager
vbox statedisk /manager
vbox start /manager
vbox snap /manager statedisk
Install a common state directory with Gluster:
flock play @@manager roles/hpc/gluster --extra-vars="master=manager-01"
Login to the master node and bootstrap the cluster (as root):
/root/gluster_bootstrap
Finally, mount the common directory (check tune parameters):
flock play @@manager roles/hpc/glusterfs
flock play @@manager roles/hpc/gtop
Monitor the cluster (as root):
/root/bin/gtop
Save:
vbox snap /manager gluster
Generate a munge key for the compute cluster and setup the scheduler services:
dd if=/dev/random bs=1 count=4096 > keys/munge.key
flock play @@manager scheduler --extra-vars=\"master=manager-01 backup=manager-02 computes=manager-[01-3]\"
Scheduler authentication relies on NTP and the Munge key. Keep the key in secret! The basic Slurm setup contains only the controller machines. Restart slurmdbd can fail.
Login to the master node and check the queue:
sinfo -a
Verify the queue:
srun -N 3 hostname
Save:
vbox snap /manager slurm
Generate a cluster key. The cluster key is used to SSH to the compute nodes:
ssh-keygen -b4096 -N "" -f keys/cluster
Check networks.yml and vars/warewulf.yml for ip ranges of compute nodes. Install Warewulf:
flock play @@ww warewulf --extra-vars=\"master=ww-01 backup=ww-02\"
Verify HA so far. Shutdown the master node
flock shutdown @@ww-01
and check Slurm failover on the backup node (as root):
/root/bin/slurmlog
Verify NFS exports:
flock command @@ww "exportfs -v"
Save:
vbox snap /ww ww
Prepare the Warewulf OS node on the master node. Login to the master node and make a clone (as root):
pushd /ww/common/chroots
./cloneos centos-6
Install basic packages:
./clonepkg centos-6
NOTE: Ad-hoc installations done:
./cloneyum centos-6 install ptpd
NOTE: Install new Ansible scripts:
flock play @@ww roles/warewulf/ansible --extra-vars=\"master=ww-01 backup=ww-02\"
Configure the clone directory:
./clonesetup centos-6
On the compute nodes NTP is used for initial time sync and PTP is used for fine sync. The controller node is a PTP master running both NTP and PTP.
FIX (if applicable): Edit /usr/share/perl5/vendor_perl/Warewulf/Provision/Pxelinux.pm line 201 comment the if block
FIX (if applicable): Edit /usr/share/perl5/vendor_perl/Warewulf/Provision/Dhcp/Isc.pm line 273 comment if block
Edit /etc/warewulf/vnfs.conf to exclude unnecessary files and make the image (database tables are created automatically):
./cloneimage centos-6
Edit /etc/warewulf/bootstrap.conf to load kernel modules/drivers/firmwares and bootstrap the kernel:
./clonekernel list centos-6
./clonekernel centos-6 3.10.5-1.el6.elrepo.x86_64
FIX: Remove /etc/warewulf/database-root.conf which is used for DB creation
mv /etc/warewulf/database-root.conf /etc/warewulf/database-root.conf.old
Create compute nodes:
flock out 3 cn
Provision:
./clonescan centos-6/3.10.5-1.el6.elrepo.x86_64 compute/cn-0[1-2]
wwsh provision set cn-0[1-2] --fileadd dynamic_hosts
or do it manually:
wwsh node new cn-01 --netdev=eth0 --hwaddr=<CN-01 MAC> --ipaddr=10.1.1.11 --groups=compute
wwsh provision set --lookup groups compute --vnfs=centos-6 --bootstrap=3.10.5-1.el6.elrepo.x86_64
wwsh provision set --lookup groups compute --fileadd dynamic_hosts
Verify the connection and monitoring:
ssh -i playbooks/keys/cluster cn-01 hostname
gstat -a
TODO: master slave host names TODO: scripts + cgroup TODO: rsyslog pipi + inittab http://www.rsyslog.com/doc/ompipe.html
./slurmconf cn-0[1-2]
At first, you have to start slurm execute service as well:
./cloneservice compute slurm start
| Parameter | srun option | When Run | Run by | As User |
|---|---|---|---|---|
| PrologSlurmCtld | job start | slurmctld | SlurmUser | |
| Prolog | job start | slurmd | SlurmdUser | |
| TaskProlog (batch) | script start | slurmstepd | User | |
| SrunProlog | --prolog | step start | srun | User |
| TaskProlog | step start | slurmstepd | User |
| --task-prolog | step start | slurmstepd | User
| --task-epilog | step finish | slurmstepd | User
TaskEpilog | | step finish | slurmstepd | User SrunEpilog | --epilog | step finish | srun | User TaskEpilog (batch) | | script finish | slurmstepd | User Epilog | | job finish | slurmd | SlurmdUser EpilogSlurmCtld | | job finish | slurmctld | SlurmUser
Interactive test:
srun -N 2 -D /common/scratch hostname
Batch test:
cd /common/scratch
Create slurm.sh:
#!/bin/sh
#SBATCH -o StdOut
#SBATCH -e StdErr
srun hostname
sbatch -N 2 slurm.sh
MPI test on the host:
git clone git://github.com/NIIF/nce.git /common/software/nce
module use /common/software/nce/modulefiles/
export NCE_ROOT=/common/software/nce
module load nce/global
Compile the latest OpenMPI:
yum install gcc gcc-c++ make gcc-gfortran environment-modules
mkdir -p ${NCE_PACKAGES}/openmpi/1.6.4
./configure --prefix=${NCE_PACKAGES}/openmpi/1.6.4
make && make install
cd examples
make
cd ..
cp -R examples /common/scratch/
cd /common/scratch/examples
module load openmpi/1.6.4
Performance tests (TODO):
ftp://ftp.mcs.anl.gov/pub/mpi/tools/perftest.tar.gz
TODO: node firewalls
TODO: trusted login node
TODO: LDAP & storage
TODO: Soft RoCE
yum --enablerepo=elrepo-kernel install kernel-ml-devel kernel-ml-headers rpm-build
yum install libibverbs-devel libibverbs-utils libibverbs
yum install librdmacm-devel librdmacm-utils librdmacm
Downlod rxe package:
wget http://198.171.48.62/pub/OFED-1.5.2-rxe.tgz
Edit /etc/exports, export fs by exportfs -a. Edit Push out a new autofs.common config:
./autofsconf compute
Service restart:
./cloneservice compute gmond restart
TODO:
- Cgroup
- Wake-on-lan with compute suspend
- Kernel 3.10 full nohz cpuset scheduler numa shit ptp
- error: /usr/sbin/nhc: exited with status 0x0100
Download CoreOS:
TFTROOT=/ww/common/tftpboot/warewulf
mkdir -p ${TFTROOT}/coreos
cd ${TFTROOT}/coreos
curl http://storage.core-os.net/coreos/amd64-generic/72.0.0/coreos_production_pxe.vmlinuz > vmlinuz
curl http://storage.core-os.net/coreos/amd64-generic/72.0.0/coreos_production_pxe_image.cpio.gz > initrd.gz
Make the following PXE default config (${TFTROOT}/pxelinux.cfg/default):
default coreos
label coreos
menu default
kernel /coreos/vmlinuz
append initrd=/coreos/initrd.gz root=squashfs: state=tmpfs: sshkey="<SSHKEY>"
where <SSHKEY> is your SSH public key.
Enable elasticsearch:
flock play @@core roles/monitor/elasticsearch
Make 3 ground state gluster. Install storage disks:
flock shutdown @@gluster
vbox storage /gluster
vbox intnet /gluster
Change to the mainline kernel:
flock play @@gluster roles/system/kernel
flock reboot @@gluster
flock snap /gluster kernel
Start the storage network:
flock play @@gluster roles/hpc/storagenet
flock play @@gluster roles/hpc/storagefw
Start gluster server:
flock play @@gluster roles/hpc/gluster
Bootstrap the cluster from the master node:
/root/gluster_bootstrap
Start the client. (Be extremely careful with limits!):
flock play @@gluster roles/hpc/glusterfs
Install the Gluster top:
flock play @@gluster roles/hpc/gtop
Start to monitor with gtop:
/opt/gluster-monitor/gtop.py
Save:
flock snap /gluster gluster
Make a client (centos flock) with intnet and the new kernel and enable the storage network:
flock play @@centos roles/hpc/storagenet
flock play @@centos roles/hpc/storagefw --tag interface
Enable client access on servers:
flock play @@gluster roles/hpc/glusterfw
Try to ping and portmap Gluster ports from the client and mount the volume on the client:
flock play @@centos roles/hpc/glustermnt
Test the fs with FIO:
yum install fio
flock snap /centos test
flock snap /gluster test
HA tests OK. Tune the timeout parameters!
Check the network bandwidth:
[root@gluster-01 ~]# iperf -s -p 49999
[root@centos-01 ~]# iperf -c 10.2.0.1 -P3 -p 49999
Check local disk speed (dd zero many times):
[root@gluster-01 ~]# echo 3 > /proc/sys/vm/drop_caches
[root@gluster-01 ~]# dd if=/dev/zero of=/tmp/zero bs=1M count=1000
Check context switching (system cs column):
[root@gluster-01 ~]# pmstat
Make 3 ground state fhgfs. Install storage disks:
flock shutdown @@fhgfs
vbox storage /fhgfs
vbox intnet/fhgfs
CentOS kernel-devel build link is broken, fix (mind the actual version!):
ln -v -f -s ../../../usr/src/kernels/2.6.32-358.18.1.el6.x86_64 /lib/modules/2.6.32-358.el6.x86_64/build
flock snap /fhgfs storage
Install FhGFS (mind that InfiniBand is not configured):
flock play @@fhgfs roles/hpc/fhgfs --extra-vars="master=fhgfs-01"
Verify (ont the master node):
fhgfs-ctl --listnodes --nodetype=meta --details
fhgfs-ctl --listnodes --nodetype=storage --details
fhgfs-net
or check the GUI (use Java 6):
java -jar /opt/fhgfs/fhgfs-admon-gui/fhgfs-admon-gui.jar
mirror metadata:
mkdir /mnt/fhgfs/test
fhgfs-ctl --mirrormd /mnt/fhgfs/test
mirror data:
fhgfs-ctl --setpattern --chunksize=1m --numtargets=2 --raid10 /mnt/fhgfs/test
Save it for good:
vbox snap /fhgfs fhgfs
Mind that, currently there is no any HA in FhGFS (distribute-only).
For the sake of simplicity create a scheduler triangle (manager).
Deploy standalone HDFS cluster:
flock play @@hadoop roles/hadoop/deploy --extra-vars \"init=yes master=hadoop-01\"
Login to the master node and format the namenode:
/root/bin/hdfs_admin format
To remove old data if you have:
rm -Rf /data/hadoop/{1,2,3}/dfs/dn/*
ssh hadoop-02 rm -Rf /data/hadoop/{1,2,3}/dfs/dn/*
ssh hadoop-03 rm -Rf /data/hadoop/{1,2,3}/dfs/dn/*
Start the master name node and the data nodes:
flock play @@hadoop roles/hadoop/start_hdfs --extra-vars \"master=hadoop-01\"
Login to the master node and initialize Yarn:
/root/bin/hdfs_admin init
Verify:
sudo -u hdfs hadoop fs -ls -R /
Enable Yarn:
flock play @@hadoop roles/hadoop/start_yarn --extra-vars \"master=hadoop-01\"
vbox snap /hadoop hdfs
Verify mapreduce. Create a test file in /tmp (hdfs) and run:
hadoop jar /usr/lib/hadoop-0.20-mapreduce/hadoop-examples.jar grep /tmp/inputfile /tmp/outputfile 'hello'
hadoop fs -cat /tmp/hadoop.out/part-00000
Login to the master node and initialize Yarn:
/root/bin/hdfs_admin flume
Enable flume and syslog forwarder:
flock play @@core roles/hadoop/flume
Login to the master node and initilaize Zookeeper HA:
hdfs zkfc -formatZK
Install MariaDB master:
flock play @@com roles/database/mariadb --extra-vars "master=com"
flock play @@com roles/database/mariadb_master --extra-vars "master=com"
Log in and secure:
mysql_secure_installation
Install openfire:
flock play @@com roles/com/openfire --extra-vars \"schema=yes master=com\"
Setup the wizard:
jdbc:mysql://127.0.0.1:3306/openfire?rewriteBatchedStatements=true
Clustering:
SSL:
Remote kdump http://blog.kreyolys.com/2011/03/17/no-panic-its-just-a-kernel-panic/
Mount (link) the ISO under boot/xs62/repo into jockey's root directory.
vbox create xs62 RedHat_64 2 2048
jockey kick xs62 @xs62 10.1.1.30 xs62
Ground state controller:
export ANSIBLE_HOSTS=xen
flock play root@@xc bootstrap
flock play @@xc secure
flock reboot @@xc
vbox snap xc init
flock play @@xc ground
flock reboot @@xc
vbox snap xc ground
SQL and MQ:
flock play @@xc roles/database/mariadb --extra-vars "master=xc"
flock play @@xc roles/database/mariadb-nowsrep --extra-vars "master=xc"
Login to xc and secure SQL:
mysql_secure_installation
flock play @@xc roles/database/memcache.yml
flock play @@xc roles/mq/rabbitmq.yml
flock play @@xc roles/monitor/icinga.yml --extra-vars \"schema=yes master=xc\"
vbox snap xc prestack
Create admin token:
openssl rand -hex 10 > keys/admin_token
Install the keystone identity service:
flock play @@xc roles/openstack/identity --extra-vars "master=xc"
Login to the machine and init/start the service:
/root/init_openstack_identity
Install sample data:
/root/sample_data.sh
Verify:
keystone --os-username=admin --os-password=secrete --os-auth-url=http://localhost:35357/v2.0 token-get
Install glance service:
flock play @@xc roles/openstack/image --extra-vars "master=xc"
Login to the machine and init/start the service:
/root/init_openstack_image
TODO Verify:
Download the installer package:
mkdir -p space/boot/kali
pushd space/boot/kali
curl http://repo.kali.org/kali/dists/kali/main/installer-amd64/current/images/netboot/netboot.tar.gz | tar xvzf -
Create a machine and bootstrap:
vbox create kali Debian_64 1 2048
jockey kick kali @kali 10.1.1.42 kali
Create and ostest group with two machines:
vbox create raring Ubuntu_64 1 2048
vbox create centos64 RedHat_64 1 2048
jockey kick centos64 @centos64 10.1.1.1 centos64
jockey kick raring @raring 10.1.1.2 raring
Start the boot servers and install the systems and reboot. Use the following inventory (ostest):
[ostest]
centos64 ansible_ssh_host=10.1.1.1 ansible_connection=paramiko
raring ansible_ssh_host=10.1.1.2 ansible_connection=paramiko
Bootstrap and verify:
flock bootstrap /ostest
flock ping @@ostest
Secure and reboot:
flock play @@ostest secure
flock reboot @@ostest
Save and start to reach the ground state:
vbox snap /ostest secure
flock play @@ostest ground
flock reboot @@ostest
vbox snap /ostest ground
Make a manager triangle with Slurm on it. For the sake of simplicity manager nodes will carry the Docker VMs. This blueprint is based on nareshv's blog.
flock play @@manager roles/docker/docker
flock reboot @@manager
[sysop@manager-01 ~]$ sudo mkdir /glusterfs/common/docker
[sysop@manager-01 ~]$ sudo chown sysop /glusterfs/common/docker
aufs is there but Unable to load the AUFS module
