Parallel computing¶
Video
Watch this in our courses: 2022 February, 2022 February real example with MPI, 2021 January
Parallel computing is what HPC is really all about: processing things on more than one processor at once. By now, you should have read all of the previous tutorials.
Abstract
You need to figure out what parallelization paradigm your program uses, otherwise you won’t know which options to use.
Embarrassingly parallel: use array jobs.
Multithreaded (OpenMP) or multiple tasks (like Python’s multiprocessing):
--cpus-per-task=N
,--mem-per-core=M
(if memory scales per CPU)MPI: compile to link with our Slurm and MPI libraries,
--ntasks=N
, always usesrun
to launch your job.module load
a MPI version for both compiling and running.
You must always monitor jobs to make sure they are using all the resources you request (
seff JOBID
).If you aren’t fully sure of how to scale up, contact us Research Software Engineers early.
Parallel programming models¶
Parallel programming is used to create programs that can execute instructions on multiple processors at a same time. Most of our users that run their programs in parallel utilize existing parallel execution features that are present in their programs and thus do not need to learn how to create parallel programs. But even when one is running programs in parallel, it is important to understand different models of parallel execution.
The two main models are:
(Embarrassingly parallel - array jobs.)
Shared memory (or multithreaded/multiprocess) programs run multiple independent workers on the same machine. As the name suggests, all of the computer’s memory has to be accessible to all of the processes. Thus programs that utilize this model should request one node, one task and multiple CPUs. Likewise, the maximum number of workers is usually the number of CPU cores available on the computational node. The code is easier to implement and the same code can still be run in a serial mode. Example applications that utilize this model: Matlab, R, Python multithreading/multiprocessing, OpenMP applications, BLAS libraries, FFTW libraries, typical multithreaded/multiprocess parallel desktop programs.
Message passing programming (e.g. MPI, message passing interface) can run on multiple nodes interconnected with the network via passing data through MPI software libraries. Almost all large-scale scientific programs utilize MPI. MPI can scale to thousands of CPU cores, but depending on the case it can be harder to implement from the programmer’s point of view. Programs that utilize this model should request single/multiple nodes with multiple tasks each. You should not request multiple CPUs per task. Example applications that utilize this model: CP2K, GPAW, LAMMPS, OpenFoam.
Both models, MPI and shared memory, can be combined in one application, in this case we are talking about hybrid parallel programming model. Programs that utilize this model can require both multiple tasks and multiple CPUs per task.
Most historical scientific code is MPI, but these days more and more people are using shared memory models.
Important
Normal serial code can’t just be run in parallel without modifications. As a user it is your responsibility to understand what parallel model implementation your code has, if any.
When deciding whether using parallel programming is worth the effort, one should be mindful of Amdahl’s law and Gustafson’s law. All programs have some parts that can only be executed in serial and thus the theoretical speedup that one can get from using parallel programming depends on two factors:
How much of programs’ execution could be done in parallel?
What would be the speedup for that parallel part?
Thus if your program runs mainly in serial but has a small parallel part, running it in parallel might not be worth it. Sometimes, doing data parallelism with e.g. array jobs is much more fruitful approach.
Another important note regarding parallelism is that all the applications scale good up to some upper limit which depends on application implementation, size and type of problem you solve and some other factors. The best practice is to benchmark your code on different number of CPU cores before you start actual production runs.
If you want to run some program in parallel, you have to know something about it - is it shared memory or MPI? A program doesn’t magically get faster when you ask more processors if it’s not designed to.
Embarrassingly parallel: array jobs¶
The array jobs we have already discussed. Don’t forget that this is one of the most common ways to parallelize! A large amount of work these days is “array jobs” + “shared memory for these jobs”
Message passing programs: MPI¶
For compiling/running an MPI job one has to pick up one of the MPI library
suites. There are various different MPI libraries that all implement the
MPI standard. We recommend that you use our OpenMPI installation
(openmpi/4.0.5
). For information on other installed versions, see the
MPI applications-page.
Some libraries/programs might have already existing requirement for a certain MPI version. If so, use that version or ask for administrators to create a version of the library with dependency on the MPI version you require.
Warning
Different versions of MPI are not compatible with each other. Each version of MPI will create code that will run correctly with only that version of MPI. Thus if you create code with a certain version, you will need to load the same version of the library when you are running the code.
Also, the MPI libraries are usually linked to slurm and network drivers. Thus, when slurm or driver versions are updated, some older versions of MPI might break. If you’re still using said versions, let us know. If you’re just starting a new project, it is recommended to use our recommended MPI libraries.
For basic use of MPI programs, you will need to use the
-n N
/--ntasks=N
-option to specify the number of MPI workers.
Compiling and running an MPI Hello world-program¶
The following example uses example codes stored in the hpc-examples-repository. You can get the repository with the following command:
git clone https://github.com/AaltoSciComp/hpc-examples/
Loading module:
module load gcc/8.4.0 # GCC
module load openmpi/4.0.5 # OpenMPI
Compiling the code:
C code is compiled with mpicc
:
cd hpc-examples/hello_mpi/
mpicc -O2 -g hello_mpi.c -o hello_mpi
Fortran code is compiled with mpifort
:
cd hpc-examples/hello_mpi_fortran/ # fortran
mpifort -O2 -g hello_mpi_fortran.f90 -o hello_mpi_fortran # Fortran code
For testing one might be interested in running the program with srun:
srun --time=00:05:00 --mem-per-cpu=200M --ntasks=4 ./hello_mpi
For actual jobs this is obviously not recommended as any problem with the login node can crash the whole MPI job. Thus we’ll want to run the program with a slurm script:
#!/bin/bash
#SBATCH --time=00:05:00 # takes 5 minutes all together
#SBATCH --mem-per-cpu=200M # 200MB per process
#SBATCH --ntasks=4 # 4 processes
module load openmpi/4.0.5 # NOTE: should be the same as you used to compile the code
srun ./hello_mpi
Important
It is important to use srun
when you launch your program.
This allows for the MPI libraries to obtain task placement information
(nodes, number of tasks per node etc.) from the slurm queue.
Spreading MPI workers evenly¶
In many cases you might require more than one node during your job’s runtime.
When this is the case, it is usually recommended to split the number of
workers evenly among the nodes. To do this, one can use
-N N
/--nodes=N
and --ntasks-per-node=n
. For example, you could
distribute the previously requested four tasks to two nodes with:
#!/bin/bash
#SBATCH --time=00:05:00 # takes 5 minutes all together
#SBATCH --mem-per-cpu=200M # 200MB per process
#SBATCH --nodes=2 # 2 nodes
#SBATCH --ntasks-per-node=2 # 2 processes per node * 2 nodes = 4 processes in total
#SBATCH --constraint=avx # set constraint for processor architecture
module load openmpi/4.0.5 # NOTE: should be the same as you used to compile the code
srun ./hello_mpi
This way the number of workers is distributed more evenly, which in turn
reduces communication overhead between workers. The total number of tasks is
--nodes
times the --ntasks-per-node
.
Setting a constraint for a specific CPU architecture¶
The number of CPUs/tasks one can specify for a single parallel job depends usually on the underlying algorithm. In many codes, such as many finite-difference codes, the workers are set in a grid-like structure. The user of said codes has then a choice of choosing the dimensions of the simulation grid aka. how many workers are in x-, y-, and z-dimensions.
For best perfomance one should reserve half or full nodes when possible. In heterogeneous clusters this can be a bit more complicated as different CPUs can have different numbers of cores.
In Triton CPU partitions there are machines with 24, 28 and 40 CPUs. See the list of available nodes for more information.
However, one can make the reservations easier by specifying a CPU architecture
with --constraint=ARCHITECTURE
. This tells Slurm to look for nodes that
satisfy a specific feature. To list available features, one can use
slurm features
.
For example, one could limit the code to the Haswell-architecture with the following script:
#!/bin/bash
#SBATCH --time=00:05:00 # takes 5 minutes all together
#SBATCH --mem-per-cpu=200M # 200MB per process
#SBATCH --nodes=1 # 1 node
#SBATCH --ntasks-per-node=24 # 24 processes as that is the number in the machine
#SBATCH --constraint=hsw # set constraint for processor architecture
module load openmpi/4.0.5 # NOTE: should be the same as you used to compile the code
srun ./hello_mpi
Monitoring performance¶
You can use seff JOBID
to see what percent of available CPUs and RAM was
utilized. Example output is given below:
$ seff 60985042
Job ID: 60985042
Cluster: triton
User/Group: tuomiss1/tuomiss1
State: COMPLETED (exit code 0)
Nodes: 1
Cores per node: 2
CPU Utilized: 00:00:29
CPU Efficiency: 90.62% of 00:00:32 core-walltime
Job Wall-clock time: 00:00:16
Memory Utilized: 1.59 MB
Memory Efficiency: 0.08% of 2.00 GB
If your processor usage is far below 100%, your code may not be working correctly. If your memory usage is far below 100% or above 100%, you might have a problem with your RAM requirements. You should set the RAM limit to be a bit above the RAM that you have utilized.
You can also monitor individual job steps by calling seff
with the syntax
seff JOBID.JOBSTEP
.
Important
When making job reservations it is important to distinguish
between requirements for the whole job (such as --mem
) and
requirements for each individual task/cpu (such as --mem-per-cpu
).
E.g. requesting --mem-per-cpu=2G
with --ntasks=2
and --cpus-per-task=4
will create a total memory reservation of
(2 tasks)*(4 cpus / task)*(2GB / cpu)=16GB.
Exercises¶
The scripts you need for the following exercises can be found in this git
repository: hpc-examples.
You can clone the repository by running
git clone https://github.com/AaltoSciComp/hpc-examples.git
. This repository
will be used for most of the tutorial exercises.
Parallel-2: OpenMP with Python
Find the files hpc-examples/openmp/hello_omp/hello_omp.c
and
hpc-examples/hello_omp/hello_omp.sh
that have a short example of OpenMP.
Compile and run it - a slurm script is included.
Parallel-3: OpenMP
Find the files in hpc-examples/python/python_openmp
. Try running the
example with a few different --constraint=X
and --cpus-per-task=C
.
In your opinion, what architecture / cpu number combination would provide the
best efficiency? Use seff
to verify.
Parallel-4: MPI
Find the files hpc-examples/mpi/hello_mpi/hello_mpi.c
and
hpc-examples/mpi/hello_mpi/hello_mpi.sh
that
have a short example of MPI.
Compile and run it - a slurm script is included.
See also¶
The Research Software Engineers can help in all aspects of parallel computing - we’d recommend anyone getting to this point set up a consultation to make sure your work is as efficient as it can be.
What’s next?¶
You have now seen the basics - but applying these in practice is still a difficult challenge! There is plenty to figure out while combining your own software, the Linux environment, and Slurm.
Your time is the most valuable thing you have. If you aren’t fully sure of how to use the tools, it is much better to ask that struggle forever. Contact us the Research Software Engineers early - for example in our daily garage, and we can help you get set up well. Then, you can continue your learning while your projects are progressing.