it center der rwth aachen university lab system environment paul kapinos 2014.10.07

10
IT Center der RWTH Aachen University Lab System Environment Paul Kapinos 2014.10.07

Upload: john-shepherd

Post on 31-Dec-2015

227 views

Category:

Documents


4 download

TRANSCRIPT

IT Center der RWTH Aachen University

Lab System Environment

Paul Kapinos

2014.10.07

aiXcelerate 2014 – Lab System EnvironmentPaul Kapinos | IT Center der RWTH Aachen University

2

Lab nodes: integrated to HPC Cluster

OS:

Scientific Linux 6.5 (RHEL6.5 compatible)

Batch system:

LSF 9.1

not for this lab

Storage:

NetApp filer ($HOME / $WORK) no backup on $WORK

Lustre ($HPCWORK) not available

aiXcelerate 2014 – Lab System EnvironmentPaul Kapinos | IT Center der RWTH Aachen University

3

Software Environment

Compiler:

Intel 15.0 (and older)

GCC 4.9 (and older)

Oracle Studio, PGI

MPI:

Open MPI, Intel MPI

No InfiniBand! 1GE only

Warnings and 1/20 of usual performance

Default:

intel/14.0 + openmpi/1.6.5

aiXcelerate 2014 – Lab System EnvironmentPaul Kapinos | IT Center der RWTH Aachen University

4

How to login

Frontends

login / SCP File transfer:

$ ssh [-Y] [email protected]

$ scp [[user@]host1:]file1 [...] [[user@]host2:]file2

then jump to the assigned lab node

$ ssh lab5[.rz.rwth-aachen.de]

cluster.rz.RWTH-Aachen.DE cluster2.rz.RWTH-Aachen.DE

cluster-x.rz.RWTH-Aachen.DE GUI cluster-x2.rz.RWTH-Aachen.DE GUI

cluster-linux.rz.RWTH-Aachen.DE cluster-linux-nehalem.rz.RWTH-Aachen.DE

cluster-linux-xeon.rz.RWTH-Aachen.DE cluster-linux-tuning.rz.RWTH-Aachen.DE

cluster-copy.rz.RWTH-Aachen.DE ‘scp’ cluster-copy2.rz.RWTH-Aachen.DE ‘scp’

aiXcelerate 2014 – Lab System EnvironmentPaul Kapinos | IT Center der RWTH Aachen University

6

Lab Node Assignment

Please use your’s allocated node only

or agree in advance with the node owner

aiXcelerate 2014 – Lab System EnvironmentPaul Kapinos | IT Center der RWTH Aachen University

7

Lab nodes

Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz

Packages(sockets) / Cores per package / Threads per core : 2/18/2

Cores / Processors(CPUs) : 36 / 72

AVX2: 256bit register

2x Fused Multiply Add (FMA) >> double peak performance cf. previous chips

64GB RAM

Stream: >100Gb/s (Triad)

No InfiniBand connection

MPI via 1GE network still possible

Warnings and 1/20 of usual performance

aiXcelerate 2014 – Lab System EnvironmentPaul Kapinos | IT Center der RWTH Aachen University

8

Module System

Many compilers, MPIs and ISV software The module system helps to manage all the packages

List loaded modules / available modules$ module list$ module avail

Load / unload a software$ module load <modulename>$ module unload <modulename>

Exchange a module (Some modules depend on each other)$ module switch <oldmodule> <newmodule>

$ module switch intel intel/15.0 Reload all modules (May fix your environment)

$ module reload Find out in which category a module is:

$ module apropos <modulename>

aiXcelerate 2014 – Lab System EnvironmentPaul Kapinos | IT Center der RWTH Aachen University

9

MPI

No InfiniBand connection

MPI via 1GE network, >> warnings and 1/20 of usual performance Default: Open MPI 1.6.5

e.g. switch to Intel MPI:

$ module switch openmpi intelmpi

Wrapper in $MPIEXEC redirects the processes to ‘back end nodes’

by default your processes run on (random) non-Haswell node

use the ‘-H’ option to start the processes on favoured node

$ $MPIEXEC -H lab5,lab6 -np 12 MPI_FastTest.exe other options of the interactive wrapper

$ $MPIEXEC -help | less

aiXcelerate 2014 – Lab System EnvironmentPaul Kapinos | IT Center der RWTH Aachen University

10

Documentation

RWTH Compute Cluster Environment

HPC Users‘s Guide (a bit outdated):

http://www.rz.rwth-aachen.de/hpc/primer

Online documentation (including example scripts):

https://doc.itc.rwth-aachen.de/

Man-Pages for all commands available

In case of errors / problems let us know:

[email protected]

aiXcelerate 2014 – Lab System EnvironmentPaul Kapinos | IT Center der RWTH Aachen University

11

Lab

We provide laptops

Log in to the laptops with the local „hpclab“ account (your own PC pool

accounts might also work)

Use X-Win32 to log in

to the cluster (use “hpclab0Z”

or your own account)

Log in to the labZ node (use

“hpclab0Z” account)

Feel free to ask questions Source: D. Both, Bull GmbH