Difference between revisions of "OpenMPI"

From HPC users
Jump to navigationJump to search
 
Line 8: Line 8:
Before compiling [[User environment - The usage of module|load the actual module]] of OpenMPI, e.g.  
Before compiling [[User environment - The usage of module|load the actual module]] of OpenMPI, e.g.  


  module load openmpi/1.8.2/gcc
  module load openmpi/1.8.4/gcc


for using the GNU compiler.
for using the GNU compiler.
Line 36: Line 36:
These programs are only wrappers which means that the scripts sets additional flags for OpenMPI (e.g. include path, flags for linking OpenMPI libraries, ...). For using the Intel Compiler please use the module  
These programs are only wrappers which means that the scripts sets additional flags for OpenMPI (e.g. include path, flags for linking OpenMPI libraries, ...). For using the Intel Compiler please use the module  


  module load openmpi/1.8.2/intel
  module load openmpi/1.8.4/intel


Below there is a list of all environment variables for setting other compiler.
Below there is a list of all environment variables for setting other compiler.
Line 84: Line 84:
For using the JAVA bindings please load the OpenMPI module with the suffix ''java'', e.g.
For using the JAVA bindings please load the OpenMPI module with the suffix ''java'', e.g.


   module load openmpi/1.8.2/java
   module load openmpi/1.8.4/java


To compile a java program with MPI bindings please use ''mpijavac'' instead of ''javac''. For launching parallel JAVA programs please use
To compile a java program with MPI bindings please use ''mpijavac'' instead of ''javac''. For launching parallel JAVA programs please use


   mpirun -machinefile $TMPDIR/machines -np $NSLOTS -x MALLOC_ARENA_MAX java <options> <java_class>
   mpirun -machinefile $TMPDIR/machines -np $NSLOTS -x MALLOC_ARENA_MAX java -Xmx10m <options> <java_class>
 
The option ''-Xmx10m'' is needed to limit the memory usage of Java and can may increased or decreased in cases of problems.
Without this option the job will probably run out of memory.


== SGE script options ==
== SGE script options ==

Latest revision as of 15:35, 9 April 2015

The OpenMPI library is an open source implementation for the MPI protocol. Different versions compiled with different compilers are available. A list is given by typing

module avail openmpi

For performance reasons the latest release of OpenMPI should be used.

Compiling with OpenMPI

Before compiling load the actual module of OpenMPI, e.g.

module load openmpi/1.8.4/gcc

for using the GNU compiler.

The compilation can be done by following wrappers of OpenMPI:

Name Description
mpicc C compiler
mpic++, mpiCC or mpicxx C++ compiler
mpif77 Fortran 77 compiler
mpif90 Fortran 90 compiler

These programs are only wrappers which means that the scripts sets additional flags for OpenMPI (e.g. include path, flags for linking OpenMPI libraries, ...). For using the Intel Compiler please use the module

module load openmpi/1.8.4/intel

Below there is a list of all environment variables for setting other compiler.

Environment variable Description
OMPI_CC Set the C compiler
OMPI_CXX Sets the C++ compiler
OMPI_F77 Sets the Fortran 77 compiler
OMPI_FC Sets the Fortran 90 compiler

Run parallel programs

The typical call to launch a MPI program within an SGE script is

 mpirun -machinefile $TMPDIR/machines -np $NSLOTS <MPI_program> <MPI_program_options>

Please don't forget to load the correct OpenMPI module before (the same OpenMPI module which was used for compilation)!

On FLOW the communication will be done over InfiniBand (automatically). Due to new virtual nodes without InfiniBand the explicit setting of InfiniBand usage by setting the environment variable OMPI_MCA_btl by

 export OMPI_MCA_btl="openib,sm,self"

or by using the mpirun or mpiexec command line option

mpirun -mca btl "openib,sm,self" ...

is depricated and can causes problems on the vx* nodes!


Run parallel JAVA programs

For using the JAVA bindings please load the OpenMPI module with the suffix java, e.g.

 module load openmpi/1.8.4/java

To compile a java program with MPI bindings please use mpijavac instead of javac. For launching parallel JAVA programs please use

 mpirun -machinefile $TMPDIR/machines -np $NSLOTS -x MALLOC_ARENA_MAX java -Xmx10m <options> <java_class>

The option -Xmx10m is needed to limit the memory usage of Java and can may increased or decreased in cases of problems. Without this option the job will probably run out of memory.

SGE script options

To submit MPI programs by SGE you have to set a parallel environment. The the parallel environment must be specified by

 #$ -pe openmpi NUM_OF_CORES
 #$ -R y

Useful environment variables

During the execution of a program called by mpirun following useful environment variables are set

Environment variable Description
OMPI_COMM_WORLD_SIZE Total number of parallel processes.
OMPI_COMM_WORLD_RANK MPI rank of the current process.

External links