Table of contents
General information
The main use case of the VLBI HPC cluster is doing VLBI correlations. However some of the cluster resources can be used for other computationally intense tasks e.g. simulations. For an overview of the cluster capabilities see Cluster Layout below.
If you want to have access to the cluster please get in touch with H.Rottmann (see Section Contact)
Contact
Who | What | Phone |
---|---|---|
Helge Rottmann | cluster administrator | 123 |
Walter Alef | cluster administrator | 289 |
Hermann Sturm | operator | 220 |
Using the cluster
- Users need to get an account first (see H. Rottmann / W.Alef).
- Once your account has been set-up you can log-into the cluster via ssh to frontend.
- /home contains your home directory and is visible on all cluster nodes
- /cluster contains software installations and is visible on all cluster nodes.
- Printing to any CUPS printer of the MPIfR should be possible from frontend.
- If you need to install software please consult with the administrators first.
Rules of conduct
- Do not store dataset in your home directory. If you need storage capacity talk to the cluster administrators first.
- Use only those cluster nodes that have been assigned to you by the cluster administrators.
- When done with your project tell the administrators so that the nodes can be used for other tasks.
- VLBI correlation always has first priority. When asked by the administrators users have to interupt their jobs and return the assigned nodes back to the correlator (this happens rarely)
Preinstalled Software
The cluster has a number of software suites pre-installed, e.g. AIPS, ParselTogue etc. If you need additional software that might be of general interst talk to the administrators to have it installed on the cluster.
Special instructions exist for:
- AIPS
- ParselTongue
Activating software version
Some of the software, e.g. openMPI exists on the cluster in various versions. The available versions can be queried by
modules avail
Sample output:
backintime/1.0.40 knem/1.1.2 openmpi/1.10.0/gcc/4.8.2 openmpi/1.10.1/gcc/5.2.0 openmpi/3.0.0/gcc/4.8.2 gcc/5.2.0 munge/0.5.11 openmpi/1.10.1/gcc/4.8.2 openmpi/2.1.2/gcc/4.8.2 slurm/15.08.3
To activate a certain version, e.g.:
module load openmpi/1.10.1/gcc/4.8.2
Cluster Specs
- The cluster has 68 compute nodes, each equipped with 20 cores. These are named node01 to node68. The nodes can be accessed from frontend.
- All nodes are equipped with 64 GB of memory
- Each node has a scratch disk (~1TB)
- The cluster interconnect is realized with FDR Infininband @56 Gbps.
- User data storage is typically available under /data11/users/username (consult with the admins if you need storage capacity)
- Software available:
- gcc 4.2.1 (g++, gfortran)
- gcc 3.3.5 (g++, gfortran; under /cluster/gcc/3.3.5/bin/)
- OpenMPI (various versions)
- Intel Performance LIbrary (IPP)
- The SLURM batch system could be configured and enabled if need arises
Cluster Layout
Infiniband Switch | 4x Infiniband Switch | 4x Infiniband Switch | |
IO11 / 82 TB, IO12 / 82 TB | nodes 29 - 48 | Nodes 1 - 28 | |
IO10 / 46 TB | |||
IO05 - IO09 / 40 TB | Nodes 53 - 68 | 4 CASA nodes | IO01 / 164 TB, IO02 / 164 TB |
IO10 / 52 TB, IO01 /20 TB IO03&04 / 20 TB Nodes 41-60, IO01 / 20 TB, Nodes 1 - 40
IO05 - IO09 / 40 TB FXmanager Appliance, Frontend,
IO02 / 20 TB Infiniband switch