Difference between revisions of "Singularity 2016"

From HPC users
Jump to navigationJump to search
 
(34 intermediate revisions by 2 users not shown)
Line 1: Line 1:
== Introduction ==
== Introduction ==
Singularity is a portable application stack packaging and runtime utility. With Singularity, you can provide whole working environments and install the OS, and software you need to use within a container.  In the best case, "this means that you don’t have to ask your cluster admin to install anything for you - you can put it in a Singularity container and run."[¹ http://singularity.lbl.gov/]
Singularity is a portable application stack packaging and runtime utility. With Singularity, you can provide whole working environments and install the OS, and software you need to use within a container.  In the best case, "this means that you don’t have to ask your cluster admin to install anything for you - you can put it in a Singularity container and run." [https://sylabs.io/docs/]


== Installed version ==
== Installed version ==


The currently installed version is available on environment hpc-env/6.4
Singularity is installed as part of the Linux OS on the head and compute nodes. That means you do not need to load any modules (anymore) to use Singularity on the cluster. The exact version can be determined from the command
  '''Singularity/2.4.2-GCCcore-6.4.0'''
$ singularity version
'''Singularity/2.6.1-GCCcore-6.4.0''' (D)
  3.4.0-1.2.el7
Any containers you have created with Singularity 3.x should be compatible with this version, older 2.x containers should be rebuild.
 
'''Note:''' The earlier modules with Singularity 2.x should no longer be used and will be removed on February 29th, 2020.


== Using Singularity ==
== Using Singularity ==


Singularity can be a very powerful tool to create containers with your favoured OS environment, and all the tools you need - completely without help or permission of the HPC administrators. Since you need root permissions to build your own containers, you will need to prepare them on your local device. In the following chapters, we will show you exemplarily how to create your own containers, fill them with the software you need, and how to use them to start your computations. But first, Singularity must be installed on a local system.
Singularity can be a very powerful tool to create containers with your favoured OS environment, and all the tools you need - completely without help or permission of the HPC administrators. Since you need root permissions to build your own containers, you will need to prepare them on your local device. In the following chapters, we will show you exemplarily how to create your own containers, fill them with the software you need, and how to use them to start your computations. But first, Singularity must be installed on a local system where you have administration rights (e.g. <tt>sudo</tt>.


=== Installation ===
=== Installation on your Local System ===
In our example, we use CentOS 7. For other distributions, the procedure should be similar, but some steps might slightly differ, especially the needed dependencies for the installation.
The installation of singularity is [https://sylabs.io/guides/3.5/admin-guide/installation.html described in the documentation]. Below, the steps for Ubuntu are listed, which work similarly for other Linux distributions. The installation is done in three steps:


To install Singularity on your system, you need to install at least two dependencies , clone the GitHub repository and perform confingure/make/make install.
'''1. Install dependencies''': To do so, open a terminal and copy-paste the following commands:
To make this as simple as possible, open a terminal, and copy-paste this lines:
<pre>
cd ~ && mkdir singularity_install && cd singularity_install
$ sudo apt-get update && sudo apt-get install -y \
sudo yum -y install libarchive-devel squashfs-tools
    build-essential \
git clone https://github.com/singularityware/singularity.git
    libssl-dev \
cd singularity
    uuid-dev \
git fetch --all
    libgpgme11-dev \
git checkout 2.6.1
    squashfs-tools \
./autogen.sh
    libseccomp-dev \
./configure --prefix=/usr/local
    wget \
make
    pkg-config \
sudo make install
    git \
    cryptsetup
</pre>


Now, Singularity should be installed on your system. Just to be sure, check the version of the newly installed software:
'''2. Install Go''': Singularity 3.x is written primarily in Go, and you will need Go 1.13 or above installed to compile it from source. The following commands are one way to do so:
  singularity --version
<pre>
Singularity should display the installed version. <br/>
$ export VERSION=1.13.7 OS=linux ARCH=amd64 && \
Should this be the case, you can remove the folder containing the source files we originally needed for the installation (~/singularity_install).
    wget https://dl.google.com/go/go$VERSION.$OS-$ARCH.tar.gz && \
    sudo tar -C /usr/local -xzvf go$VERSION.$OS-$ARCH.tar.gz && \
    rm go$VERSION.$OS-$ARCH.tar.gz
</pre>
where you can change the version as needed. Then setup the environment to complete the Go installation:
<pre>
$ echo 'export GOPATH=${HOME}/go' >> ~/.bashrc && \
    echo 'export PATH=/usr/local/go/bin:${PATH}:${GOPATH}/bin' >> ~/.bashrc && \
    source ~/.bashrc
</pre>
You may want to repeat this step for other users on the same system if they need Go for other reasons.
 
'''3. Download, build and install Singularity''': The last commands finally install Singularity from source on your machine:
<pre>
$ export VERSION=3.5.2 && # adjust this as necessary \
    wget https://github.com/sylabs/singularity/releases/download/v${VERSION}/singularity-${VERSION}.tar.gz && \
    tar -xzf singularity-${VERSION}.tar.gz && \
    cd singularity
</pre>
where you can check the available releases on https://github.com/sylabs/singularity/releases. Finally, build and install with:
<pre>
$ ./mconfig && \
    make -C ./builddir && \
    sudo make -C ./builddir install
</pre>
which installs Singularity in <tt>/usr/local</tt>.
 
In order to test if Singularity was installed correctly, you can use the command:
  singularity version
which should display the installed version.
If everything is done, you can remove the folder containing the source files with
rm -r ~/singularity.


=== Creating Containers ===
=== Creating Containers ===
Line 44: Line 81:
'' Always remember: The following commands must be executed on your own local machine! '' <br/>
'' Always remember: The following commands must be executed on your own local machine! '' <br/>
To download an image only containing CentOS, type in the following command:
To download an image only containing CentOS, type in the following command:
  sudo singularity build --sandbox centos7.img docker://centos:7
  $ sudo singularity build --sandbox centos7.img/ docker://centos:7
This will create a directory <tt>centos7.img</tt> on your local computer which contains a basic Linux installation and can be modified to your needs.


As you can see, with ''docker://centos:'', we used a [https://hub.docker.com/search/?q=&type=image&image_filter=official&category=os docker image].
As you can see, with ''docker://centos:'', we used a [https://hub.docker.com/search/?q=&type=image&image_filter=official&category=os docker image] from the Docker hub. A lot of distributions, like Fedora, Ubuntu, Debian, etc.´, are available that way.
You can do this with a lot of distributions, like Fedora, Ubuntu, Debian, etc.) <br/>
 
Now, switch into the container environment and install the packets you need:
Now, you can start a shell within the container environment and install the packets you need:
  sudo singularity shell --writable centos7.img
  $ sudo singularity shell --writable centos7.img
  ''# Now, every shell command will be forwarded to the container centos.img''
  # ''# Now, every shell command will be executed within the container centos.img (idicated by the prompt changing from $ to #)''
  yum update -y
  # yum update -y
  yum install vim -y
  # yum install vim -y
  ''# To escape from the container shell, type in:''
  # ''# To escape from the container shell, type in:''
  exit
  # exit
In this example, we switched to the containers shell, did some updates, installed the text editor vim, and finally exited from the containers shell with ''exit''.
In this example, we switched to the containers shell, did some updates, installed the text editor vim, and finally exited from the containers shell with ''exit''. As you probably noticed, the commands within the container are executed as root because we ran Singularity with <tt>sudo</tt>.


When you finished modifying the image, create a read-only image file for the productive use (which additionally makes is very space efficient):
When you finished modifying the image, create a read-only image file for the productive use (which additionally makes is very space efficient):
  sudo singularity build centos7'''.simg''' centos7.img
  sudo singularity build centos7'''.sif''' centos7.img/
 
Singularity will automatically bind your $HOME directory so that you can enter your files from within the container.
If you need other paths to be connected to your container, you can can do this with the ''--bind '' command:
singularity shell --bind $WORK centos7.simg


==== B) Working with your own recipe (Bootstrapping) ====
==== B) Working with your own recipe (Bootstrapping) ====
Line 84: Line 118:
  %labels
  %labels
  AUTHOR Monika Testuser, monika.testuser@uol.de
  AUTHOR Monika Testuser, monika.testuser@uol.de
 
  ''# Run post-install commands. Use them just like you would do within the targeted operating system''  
  ''# Run post-install commands. Use them just like you would do within the targeted operating system''  
  %post
  %post
Line 121: Line 155:
   sudo singularity build my_container.simg my_container.img
   sudo singularity build my_container.simg my_container.img


==== Transferring your Container to CARL/ERLE and start working ====
==== Transferring your Container to CARL/EDDY and start working ====
If your container is ready for computing and finalized by building into a .simg file, you will probably want to transfer it to the cluster (CARL or EDDIE). For this task, we can use the tool scp, which should be pre installed on every common linux distribution. The usage is very straightforward:
If your container is ready for computing and finalized by building into a .simg file, you will probably want to transfer it to the cluster (CARL or EDDY). For this task, we can use the tool scp, which should be pre installed on every common linux distribution. The usage is very straightforward:
  scp <container_file> abcd1234>@carl.hpc.uni-oldenburg.de:/home/abcd1234/
  scp <container_file> abcd1234@carl.hpc.uni-oldenburg.de:/user/abcd1234/
  # Where abcd1234 is your domain user name.
  # Where abcd1234 is your domain user name.


Line 132: Line 166:


At [[#Putting in Practice|Putting in Practice]], we show exemplarily, how using a container could look like.
At [[#Putting in Practice|Putting in Practice]], we show exemplarily, how using a container could look like.
==== Troubleshooting ====
'''sudo: singularity: command not found''' <br/>
If this error occurs on your local machine, sudo may be missing a path to Singularitys installation directory.
With ''sudo visodu'' this can be changed by adding the path to a specific line:
Just add '':/usr/local/bin'' to the line containing ''secure_path'' so it looks like this:
Defaults    secure_path = /sbin:/bin:/usr/sbin:/usr/bin:/usr/local/bin


=== Putting in Practice ===
=== Putting in Practice ===
Now, we will show you how to use Singularity from the first step to the last command.
Now, we will show you how to use Singularity from the first step to the last command.<br/>
Let's say, we want to convert a video file into a different type: ''.flv'' to 'mp4'' because .flv files can't be handled by many video players.
Let's say, we want to convert a video file into a different type: ''flv'' to ''mp4'', because ''flv'' files can't be handled by many video players. <br/>
We already stored our beloved jellyfish clip in our HPCs Download directory:
We already stored our jellyfish clip in our HPCs Download directory:
  #on CARL/ERLE:
  #on CARL/EDDY:
  cd ~/Downloads
  cd ~/Downloads
  wget http://mirrors.standaloneinstaller.com/video-sample/jellyfish-25-mbps-hd-hevc.flv
  wget http://mirrors.standaloneinstaller.com/video-sample/jellyfish-25-mbps-hd-hevc.flv


We choose HandBrakeCLI as our converting tool. Since converting files can be very compute intensive, we decide to use the HPC Cluster. The only problem is, that HandBrakeCLI is not available as a loadable module. This is where Singularity comes in handy. <br/>
We choose HandBrakeCLI as our converting tool. Since converting files can be very compute intensive, we decide to use the HPC Cluster. The only problem is, that HandBrakeCLI is not available as a loadable module. This is where Singularity comes in handy. <br/>
After we [[#Installation|installed Singularity]], we ware ready to go: <br/>
After we [[#Installation|installed Singularity]], we are ready to go: <br/>
On our local machine, we create an image by writing a definition file with ''vim HandBrake.def''...
On our local machine, we create an image by writing a definition file with ''vim HandBrake.def''...


Line 175: Line 202:


Let's transfer the new image to our cluster:
Let's transfer the new image to our cluster:
  scp HandBrake.simg abcd1234>@carl.hpc.uni-oldenburg.de:/home/abcd1234/
  scp HandBrake.simg abcd1234@carl.hpc.uni-oldenburg.de:/user/abcd1234/


At last, we can actually ''work'' with our new image. <br/>
At last, we can actually ''work'' with our new image. <br/>
At this point, we will use the ''exec'' command:
On our local system, we would just call an ''exec'' command to ''HandBrakeCLI'' in the container:
  singularity exec HandBrake.simg HandBrakeCLI -i ~/Downloads/jellyfish-25-mbps-hd-hevc.flv -o ~/Downloads/jellyfish-25-mbps-hd-hevc.mp4
  singularity exec HandBrake.simg HandBrakeCLI -i ~/Downloads/jellyfish-25-mbps-hd-hevc.flv -o ~/Downloads/jellyfish-25-mbps-hd-hevc.mp4
#For better understanding: after ''singularity exec'', every command and argument is transferred
#into the container. If HandBrakeCLI had been installed on our system, it would have been ignored anyway.


And that's it! We used the exec command to call the containerized HandBrakeCLI and processed the file outside of the container.
But since we work on the cluster, we will create a job to use the SLURM queuing system.
First, we create a job file with vim...
vim HandBrake.job
... and then we will fill it with sbatch arguments and the actual commands we want it to get done:
 
#!/bin/bash
#SBATCH --nodes=1                   
#SBATCH --ntasks=1                 
#SBATCH --mem=2G                 
#SBATCH --time=0-1:00               
#SBATCH --output=slurm.%j.out       
#SBATCH --error=slurm.%j.err         
#SBATCH --mail-type=END,FAIL       
#SBATCH --mail-user=your.name@uol.de
# Load the modules we need:
module load hpc-env/6.4
module load Singularity/2.6.1-GCCcore-6.4.0
# Actually work with the singularity container:
singularity exec HandBrake.simg HandBrakeCLI -i ~/Downloads/jellyfish-25-mbps-hd-hevc.flv -o ~/Downloads/jellyfish-25-mbps-hd-hevc.mp4
 
Afterwards, we submit the job:
sbatch -p carl.p HandBrake.job
In this example, we reserved one hour on one node on the [[Partitions | carl.p partition]], which should be more than enough for this simple task. For more Information about our SLURM system and the corresponding options, visit our [[SLURM Job Management (Queueing) System | page about SLURM]] <br/> <br/>
But that's it! We used the exec command to call the containerized HandBrakeCLI and processed the file outside of the container.
This is possible, since Singularity automatically binds your $HOME path.
This is possible, since Singularity automatically binds your $HOME path.
And now, our beloved clip of jellyfish is watchable on nearly every video player (even if this should not be the main purpose of our HPC cluster (;  ).
=== Working outside of $HOME ===
Singularity will automatically bind your $HOME directory so that you can enter your files from within the container.
If you need other paths to be connected to your container, you can can do this with the ''--bind '' command:
singularity shell --bind $WORK centos7.simg
=== Troubleshooting ===
====sudo: singularity: command not found====
If this error occurs on your local machine, sudo may be missing a path to Singularitys installation directory.
With ''sudo visodu'' this can be changed by adding the path to a specific line:
Just add '':/usr/local/bin'' to the line containing ''secure_path'' so it looks like this:
Defaults    secure_path = /sbin:/bin:/usr/sbin:/usr/bin:/usr/local/bin
==== mkdir: cannot create directory ''<chosen_directory>'': Permission denied ====
Probably, the directory that you want wo work on was not made known to Singularity. <br>
You need to use the option ''--bind'' to set the directory, e.g.:
singularity run --bind $WORK /gss/work/abcd1234/my_image.simg -d /gss/work/abcd1234/project_folder/ -r
### With ''--bind $WORK'' Singularity gets the permission to write on /gss/work/abcd1234
== More Examples for Using Singularity ==


You can find more examples on how to use Singularity on the HPC cluster [[Singularity Examples | here]].


== Documentation ==
== Documentation ==


For further information developers offer a detailed [https://singularity.lbl.gov/ website] with further information, and a [https://singularity.lbl.gov/quickstart Quick Stard Guide].
For further information developers offer a detailed [https://sylabs.io/docs/ website] with further information, and a [https://sylabs.io/guides/3.5/user-guide/quick_start.html Quick Stard Guide].

Latest revision as of 13:07, 11 February 2020

Introduction

Singularity is a portable application stack packaging and runtime utility. With Singularity, you can provide whole working environments and install the OS, and software you need to use within a container. In the best case, "this means that you don’t have to ask your cluster admin to install anything for you - you can put it in a Singularity container and run." [1]

Installed version

Singularity is installed as part of the Linux OS on the head and compute nodes. That means you do not need to load any modules (anymore) to use Singularity on the cluster. The exact version can be determined from the command

$ singularity version
3.4.0-1.2.el7

Any containers you have created with Singularity 3.x should be compatible with this version, older 2.x containers should be rebuild.

Note: The earlier modules with Singularity 2.x should no longer be used and will be removed on February 29th, 2020.

Using Singularity

Singularity can be a very powerful tool to create containers with your favoured OS environment, and all the tools you need - completely without help or permission of the HPC administrators. Since you need root permissions to build your own containers, you will need to prepare them on your local device. In the following chapters, we will show you exemplarily how to create your own containers, fill them with the software you need, and how to use them to start your computations. But first, Singularity must be installed on a local system where you have administration rights (e.g. sudo.

Installation on your Local System

The installation of singularity is described in the documentation. Below, the steps for Ubuntu are listed, which work similarly for other Linux distributions. The installation is done in three steps:

1. Install dependencies: To do so, open a terminal and copy-paste the following commands:

$ sudo apt-get update && sudo apt-get install -y \
    build-essential \
    libssl-dev \
    uuid-dev \
    libgpgme11-dev \
    squashfs-tools \
    libseccomp-dev \
    wget \
    pkg-config \
    git \
    cryptsetup

2. Install Go: Singularity 3.x is written primarily in Go, and you will need Go 1.13 or above installed to compile it from source. The following commands are one way to do so:

$ export VERSION=1.13.7 OS=linux ARCH=amd64 && \
    wget https://dl.google.com/go/go$VERSION.$OS-$ARCH.tar.gz && \
    sudo tar -C /usr/local -xzvf go$VERSION.$OS-$ARCH.tar.gz && \
    rm go$VERSION.$OS-$ARCH.tar.gz

where you can change the version as needed. Then setup the environment to complete the Go installation:

$ echo 'export GOPATH=${HOME}/go' >> ~/.bashrc && \
    echo 'export PATH=/usr/local/go/bin:${PATH}:${GOPATH}/bin' >> ~/.bashrc && \
    source ~/.bashrc

You may want to repeat this step for other users on the same system if they need Go for other reasons.

3. Download, build and install Singularity: The last commands finally install Singularity from source on your machine:

$ export VERSION=3.5.2 && # adjust this as necessary \
    wget https://github.com/sylabs/singularity/releases/download/v${VERSION}/singularity-${VERSION}.tar.gz && \
    tar -xzf singularity-${VERSION}.tar.gz && \
    cd singularity

where you can check the available releases on https://github.com/sylabs/singularity/releases. Finally, build and install with:

$ ./mconfig && \
    make -C ./builddir && \
    sudo make -C ./builddir install

which installs Singularity in /usr/local.

In order to test if Singularity was installed correctly, you can use the command:

singularity version

which should display the installed version.

If everything is done, you can remove the folder containing the source files with

rm -r ~/singularity.

Creating Containers

There are three ways to create containers which will be discussed:

  • A) Downloading solely OS containers and manually adding software packets
  • B) Creating installation scripts, including OS and post-install commands
  • C) Searching for 'off-the-shelf scripts' from Singularitys repository

Below we will briefly cover each of the methods mentioned.

A) Download and modify

Always remember: The following commands must be executed on your own local machine!
To download an image only containing CentOS, type in the following command:

$ sudo singularity build --sandbox centos7.img/ docker://centos:7

This will create a directory centos7.img on your local computer which contains a basic Linux installation and can be modified to your needs.

As you can see, with docker://centos:, we used a docker image from the Docker hub. A lot of distributions, like Fedora, Ubuntu, Debian, etc.´, are available that way.

Now, you can start a shell within the container environment and install the packets you need:

$ sudo singularity shell --writable centos7.img
# # Now, every shell command will be executed within the container centos.img (idicated by the prompt changing from $ to #)
# yum update -y
# yum install vim -y
# # To escape from the container shell, type in:
# exit

In this example, we switched to the containers shell, did some updates, installed the text editor vim, and finally exited from the containers shell with exit. As you probably noticed, the commands within the container are executed as root because we ran Singularity with sudo.

When you finished modifying the image, create a read-only image file for the productive use (which additionally makes is very space efficient):

sudo singularity build centos7.sif centos7.img/

B) Working with your own recipe (Bootstrapping)

By creating a personal recipe (also called definition file = .def), you have the advantages of creating your personal environment, choosing the OS, and installing the software that fits your needs perfectly. When the recipe is created, a new container can be bootstrapped with one single command at any time.

Start by creating a script file, e.g. with the vim editor...

vim my_container.def

... and fill it with the settings you prefer. You can comment out with #, so that you can remember the purpose of every line, if you want to store the recipe file for some time.

# Chose an OS, preferably from the official docker images
Bootstrap: docker
From: centos:latest

# Set environment variables
%environment
PATH=new/directory:$PATH

# Set labels 
%labels
AUTHOR Monika Testuser, monika.testuser@uol.de

# Run post-install commands. Use them just like you would do within the targeted operating system 
%post
yum -y update && yum -y install wget vim

# Copy files into the container.
%files
source/path/to/file.txt target/path/to/file.txt

After saving and closing the file, you start the installation with:

sudo singularity build --sandbox my_container.img my_container.def

Since we used the argument --sandbox, we can modify the newly built image afterwards like shown on A)

sudo singularity shell --writable my_container.img
touch testfile.txt
yum -y install vim 
exit

Like in A), create a read-only image file when you finished modifying the image:

sudo singularity build my_container.simg my_container.img

C) Use an 'off-the-shelf' image

On the Singularity Hub, you have access to a lot of pre-puilt images and the corresponting recipes.

Here, we have two options: A) Downloading the complete image files or B) downloading the corresponding recipe. For downloading the ready-to-use image, Singularity has a pull function.

# Pulling a tensorflow image from the singularity hub:
singularity pull shub://marcc-hpc/tensorflow

This command will download a read-only image file. You use it to start calculations but to change or modify it, you would have to change the image structure. Should you need to do any changes, consider using the recipe. On the project site of tensorflow (for example), you can download the recipe for the container image. When downloaded, the recipe can be extended by any desired argument and be build as --writable or --sandbox if any further changes are desired.

https://singularity-hub.org/containers/2725/download/recipe

Like in A) and B, finalize your image into a .simg file when you are done modifying.

 sudo singularity build my_container.simg my_container.img

Transferring your Container to CARL/EDDY and start working

If your container is ready for computing and finalized by building into a .simg file, you will probably want to transfer it to the cluster (CARL or EDDY). For this task, we can use the tool scp, which should be pre installed on every common linux distribution. The usage is very straightforward:

scp <container_file> abcd1234@carl.hpc.uni-oldenburg.de:/user/abcd1234/
# Where abcd1234 is your domain user name.

Now you can login to CARL, load the Singularity module, and start using your container:

ssh abcd1234 carl.hpc.uni-oldenburg.de
module load hpc-env/6.4
module load Singularity

At Putting in Practice, we show exemplarily, how using a container could look like.

Putting in Practice

Now, we will show you how to use Singularity from the first step to the last command.
Let's say, we want to convert a video file into a different type: flv to mp4, because flv files can't be handled by many video players.
We already stored our jellyfish clip in our HPCs Download directory:

#on CARL/EDDY:
cd ~/Downloads
wget http://mirrors.standaloneinstaller.com/video-sample/jellyfish-25-mbps-hd-hevc.flv

We choose HandBrakeCLI as our converting tool. Since converting files can be very compute intensive, we decide to use the HPC Cluster. The only problem is, that HandBrakeCLI is not available as a loadable module. This is where Singularity comes in handy.
After we installed Singularity, we are ready to go:
On our local machine, we create an image by writing a definition file with vim HandBrake.def...

Bootstrap: docker
From: ubuntu:latest

%post
# install dependencies 
apt-get -y update
apt-get -y install autoconf automake build-essential cmake git libass-dev libbz2-dev libfontconfig1-dev libfreetype6-dev libfribidi-dev libharfbuzz-dev libjansson-dev liblzma-dev libmp3lame-dev libogg-dev libopus-dev libsamplerate-dev libspeex-dev libtheora-dev libtool libtool-bin libvorbis-dev libx264-dev libxml2-dev libvpx-dev m4 make nasm patch pkg-config python tar yasm zlib1g-dev libdrm-dev

# pull source files, make & install HandBrakeCLI & delete the install-dir
mkdir /handbrake-install
cd /handbrake-install
git clone https://github.com/HandBrake/HandBrake.git && cd HandBrake
./configure --launch-jobs=$(nproc) --launch --disable-gtk
make --directory=build install
cd /
rm -rf /handbrake-install

... and build the image with this recipe:

sudo singularity build HandBrake.simg HandBrake.def 

--> Since we did not need to modify anything after the building process, we skipped --sandbox and directly created a .simg file!

Let's transfer the new image to our cluster:

scp HandBrake.simg abcd1234@carl.hpc.uni-oldenburg.de:/user/abcd1234/

At last, we can actually work with our new image.
On our local system, we would just call an exec command to HandBrakeCLI in the container:

singularity exec HandBrake.simg HandBrakeCLI -i ~/Downloads/jellyfish-25-mbps-hd-hevc.flv -o ~/Downloads/jellyfish-25-mbps-hd-hevc.mp4

#For better understanding: after singularity exec, every command and argument is transferred
#into the container. If HandBrakeCLI had been installed on our system, it would have been ignored anyway.

But since we work on the cluster, we will create a job to use the SLURM queuing system. First, we create a job file with vim...

vim HandBrake.job

... and then we will fill it with sbatch arguments and the actual commands we want it to get done:

#!/bin/bash

#SBATCH --nodes=1                    
#SBATCH --ntasks=1                  
#SBATCH --mem=2G                  
#SBATCH --time=0-1:00                
#SBATCH --output=slurm.%j.out        
#SBATCH --error=slurm.%j.err          
#SBATCH --mail-type=END,FAIL         
#SBATCH --mail-user=your.name@uol.de 

# Load the modules we need:
module load hpc-env/6.4
module load Singularity/2.6.1-GCCcore-6.4.0

# Actually work with the singularity container:
singularity exec HandBrake.simg HandBrakeCLI -i ~/Downloads/jellyfish-25-mbps-hd-hevc.flv -o ~/Downloads/jellyfish-25-mbps-hd-hevc.mp4

Afterwards, we submit the job:

sbatch -p carl.p HandBrake.job

In this example, we reserved one hour on one node on the carl.p partition, which should be more than enough for this simple task. For more Information about our SLURM system and the corresponding options, visit our page about SLURM

But that's it! We used the exec command to call the containerized HandBrakeCLI and processed the file outside of the container. This is possible, since Singularity automatically binds your $HOME path. And now, our beloved clip of jellyfish is watchable on nearly every video player (even if this should not be the main purpose of our HPC cluster (; ).

Working outside of $HOME

Singularity will automatically bind your $HOME directory so that you can enter your files from within the container. If you need other paths to be connected to your container, you can can do this with the --bind command:

singularity shell --bind $WORK centos7.simg


Troubleshooting

sudo: singularity: command not found

If this error occurs on your local machine, sudo may be missing a path to Singularitys installation directory. With sudo visodu this can be changed by adding the path to a specific line:

Just add :/usr/local/bin to the line containing secure_path so it looks like this:
Defaults    secure_path = /sbin:/bin:/usr/sbin:/usr/bin:/usr/local/bin

mkdir: cannot create directory <chosen_directory>: Permission denied

Probably, the directory that you want wo work on was not made known to Singularity.
You need to use the option --bind to set the directory, e.g.:

singularity run --bind $WORK /gss/work/abcd1234/my_image.simg -d /gss/work/abcd1234/project_folder/ -r
### With --bind $WORK Singularity gets the permission to write on /gss/work/abcd1234

More Examples for Using Singularity

You can find more examples on how to use Singularity on the HPC cluster here.

Documentation

For further information developers offer a detailed website with further information, and a Quick Stard Guide.