MPI

We offer tightly-integrated OpenMPI and MPICH2 on the HPCC.

MPI Selection (module)

The default MPI module is `mpi/openmpi-x86_64` (GNU CC/C++ & Fortran). If you wish to use the Intel compilers, or the MPICH2 MPI, you will need to edit your `~/.bashrc` file, find the ‘MPI SELECTION’ section, and uncomment ONLY the MPI module that you with to use, before any MPI commands are run on the HPCC. We recommend asking us about this first, if you believe that you should be doing this.

C MPI on the HPCC

More information: C/C++ on the HPCC (including MPI)

Fortran MPI on the HPCC

Submitting a Fortran MPI Job

Remember Job Management with job/slot limits for job queues.

Create Fortran MPI Commands File

Create a .f file with your commands, for example: Fortran/Sample MPI Commands File

Compile Fortran MPI Code

Compile the .f commands file with the appropriate compiler:

qrsh mpif90 -o your-commands-file your-commands-file.f

Create Fortran MPI Job Script

Create a .sh file with at least the following contents:

#!/bin/bash
mpiexec your-commands-file

Submit Fortran MPI Job

qsub -pe openmpi 4 your-job-script.sh

More information: Job Management