Storage areas

This page describes the directories / file systems that are available to each user for storing files and data sets. To see how much space you have available at the different file systems, the hbquota tool can be used. Please refer to the Quota page for more information.

Each user of the system has its own home directory located at one of the home directory servers. This is the directory you will be in after logging in to the system. Since we do make a tape backup of this area the amount of space for each user is limited using quotas. Currently the limit is 50GB by default.

The data on the home directories is available on all nodes in the system. The user home directories will be in /home[n], where n is a number. The range of numbers depends on the number of storage servers deployed. Your home directory may for example be in /home3/username, where you have to replace your username by your p-, s- or f-number.

On the command line the shortcuts $HOME or ~ can be used to reach your home directory.

Each user of the system also has a directory in /scratch, which is meant for storing large amounts of data that needs to be processed. Please be aware that backups of this data are not made, and that /scratch is not meant for long term data storage. This means that you have to copy important data to other storage yourself. This storage can for example be the /projects or RDMS storage system.

Also on /scratch quotas are applied to prevent the system from running out of space. Currently the limit is 250GB by default. If this limit is too low for your research purposes, you can request us to change this limit. The limit can be increased to a “fair use” value without issues. When more space is required it is expected that /scratch is still only used as a staging area for data that will immediately be processed, and that a suitable storage system is available elsewhere for storing the full data collection. These storage systems can again be the /projects or RDMS systems described below.

The data on /scratch is available on all nodes in the system.

There is also a limit on the number of files that can be stored. This to reduce the load on the file system metadata server, which keeps track of the data about files (time of access, change, size, location, etc.). Handling a huge number of files is a challenge for most shared file systems and accessing a huge amount of files will lead to performance bottlenecks.

The best way of handling data sets with many (> 10,000) files is to not store them on /scratch as is, but as (compressed) archive files. These files can then be extracted to the fast local storage on the compute nodes at the beginning of a job.

When the processing is performed on the fast local storage the job performance will be much better.

Each node of the cluster has an amount of fast internal disk space. Most of this space is mounted under a job-specific path that can be reached using the $TMPDIR environment variable. This space is only available for running jobs. For each job a temporary directory is created on this disk space. To prevent people from storing data permanently these directories are removed automatically after the jobs is finished. This means that you have to copy away important data from this location at the end of your job script.

Note that the disk space in $TMPDIR is not shared between the nodes. You cannot access the files on one machine from another, without copying them explicitly over the network.

You can use the name of directory in your job scripts using $TMPDIR. The $ sign denotes that you are referring to the environment variable TMPDIR. Examples:

Change directory:

cd $TMPDIR

Copy files to the temporary directory:

cp mydirectory/* $TMPDIR

So a full fictitious jobscript could look like:

module purge
module load MYPROGRAM/1.0
cp -r mydataset/* $TMPDIR
cd $TMPDIR
myprogram -i inputdata -o outputdata
cp -r outputdata /scratch/$USER/mydataset/

Relevant external data stores

On the login/interactive nodes each user will have storage space in /projects/$USER. This storage area is meant for storing data for a longer period. It cannot be reached by the compute nodes as it is not optimized for processing data.

By default 250GB is allocated. This space can be increase upon request, based on a “fair use” principle. Above a certain threshold payment for the space will be required.

The Research Data Management system (RDMS) is also available on the Hábrók nodes. More details about this system can be found in the dedicated wiki: https://wiki.hpc.rug.nl/rdms/start