# Research Storage

(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)

Research Storage is a scalable storage fabric designed to grow with your research operations. Data set sizes are growing dramatically. Keeping enough storage on hand to manage Big Data is a challenge for everyone. The Research Storage service provides nearly unlimited capacity to hold the data important to your research. The service is built using flexible technologies that can support any research data requirement.

## Introduction

The following description of research storage locations and hardware assumes that all files and data will be placed on Cheaha with the intent of being used for computation related to a legitimate research need. Compute storage is not intended for archival or backup purposes, and is not intended to store personal, non-research and non-educational data. Educational data should be limited to coursework requiring computational research. Research Computing does not own any data placed by users on Cheaha. Backup services are not provided by the Research Computing department and must be maintained by the data owner.

The following terms may be useful:

• GB - gigabyte
• TB - terabyte (1024 GB)
• PB - petabyte (1024 PB)
• $- anything following a$ symbol is a shell variable. All of the variables used here should be predefined in your shell environment on Cheaha
• $USER - a shell variable containing your username/blazerid or XIAS account username All users are provided with a 5 TB personal allocation under /data/user/$USER, also known as $USER_DATA. Users with collaboration needs such as PI labs, large-scale research projects or external collaborators on projects may request space under /data/project, also known as $SHARE_PROJECT. The default shared space is 50 TB.

Users may also make use of global and node-local scratch space. Global scratch space is /scratch/$USER also known as $USER_SCRATCH and has 1 PB total. Node-local scratch space is /scratch/local or $LOCAL_SCRATCH and has 1 TB per node. Files in these locations should be cleaned up by users regularly when used, and as soon as jobs are completed for $LOCAL_SCRATCH.

## Description

The following description of research storage locations and hardware assumes that all files and data will be placed on Cheaha with the intent of being used for computation related to a legitimate research need. Compute storage is not intended for archival or backup purposes, and is not intended to store personal, non-research related data. Research Computing does not own any data placed by users on Cheaha. Backup services are not provided by the Research Computing department and must be maintained by the data owner.

The following terms may be useful for the description of research storage:

• GB - gigabyte
• TB - terabyte (1024 GB)
• PB - petabyte (1024 PB)
• $- anything following a$ symbol is a shell variable. All of the variables used here should be predefined in your shell environment on Cheaha
• $USER - a shell variable containing your username/blazerid or XIAS account username ##### Home Directory aka $HOME

Points to /home/$USER. Your home directory, where the operating system and most software stores personal configuration files. Historically on a separate filesystem from other storage locations, with a 20 GB quota. The previous hardware was retired and its data was placed on the primary $USER_DATA hardware and merged into that quota.

Points to /scratch/$USER. The preferred location for storing any temporary files related to research computing. Total space is 1 PB (petabyte) or 1024 TB shared among all users of the cluster. Users should delete files placed here on a regular basis or on job completion. ##### $LOCAL_SCRATCH

Points to /scratch/local. The preferred location for storing small quantities of temporary files for currently running jobs. Local to each compute node and not shared between nodes, but shared between all users. Typically this space is about 1 TB. Highest possible IO performance for a single node job. Users must delete files placed here on job completion.

To review project quota please use the following command at the terminal, replacing <PROJECT-SLUG> with the name of your project directory. Not the full directory path, just the part immediately after /data/project/.

/usr/lpp/mmfs/bin/mmlsquota --block-size=auto -v -j <PROJECT-SLUG> data


You can also add a function to your .bashrc file as follows. Use it as project_quota <PROJECT-SLUG>.

function project_quota() {