Humus

From ACENET
Jump to: navigation, search

Humus is a pair of computers located at Dalhousie University managed by ACENET on behalf of researchers who purchased the equipment. Humus is integrated with the Nefelibata cluster, but Humus and the balance of Nefelibata are owned by two different research groups.

Access

You may only access Humus with the permission of the contributing Principal Investigator. If you don't have access to Humus and believe you should, please write to support@ace-net.ca and say so. Copy the PI, and mention in the email your Digital Research Alliance of Canada username.

The two machines are named humus216 and humus217. (The numbers are in relation to the node numbering of Nefelibata.) Choose one to log into, like so:

 ssh -J username@nefelibata.ace-net.ca username@humus216

or

 ssh -J username@nefelibata.ace-net.ca username@humus217

Supply your Alliance password when prompted. The -J parameter above stands for "jump"; you are "jumping" through nefelibata.ace-net.ca to get to humus21[6,7].

We strongly recommend creating and installing an SSH key pair and using that for access after your first login.

Storage system

When you log in to Humus your default working directory will be /home/username on the shared Nefelibata filesystem. But there is very little space on it for Humus users.

For primary storage please use /scratch/username on each Humus node. Each node has approximately 58 TB of (unshared, node-local) storage space in /scratch.

Data transfer

sftp can be used for interactive command-line data transfer, following the pattern shown above for ssh:

sftp -J username@nefelibata.ace-net.ca username@humus216
> cd /scratch/username
> put somefile.txt
> get some-other-file
> exit

scp can be used similarly for wild-card or script-driven data transfers:

scp -J username@nefelibata.ace-net.ca somefiles.* username@humus216:/scratch/username/

Graphical (drag-and-drop) file transfer applications can be used if they implement the scp or sftp protocols. WinSCP has been tested successfully; use username@humus216.nefelibata.ace-net.ca etc.

Software

Humus uses the same modules system as Alliance clusters, providing access to the same list of available software.

For software which requires a graphical user interface (e.g. RStudio), you can connect to Humus using VNC. Here is a step-by-step guide to creating a VNC session.

Job scheduling

We recommend using batch jobs for your work whenever feasible, using the Slurm job scheduler. This allows you to queue up several tasks which will execute one after another without constant human supervision. It also allows group members to share the resources more effectively, 'e.g.' the scheduler will not try to run two different jobs that both need a GPU on the same node at the same time; it will either run them on different nodes or at different times.

The Humus nodes carry NVidia A30 GPUs, so the following command will run `jobscript` on one entire node:

sbatch --time=6:0:0 --gres=gpu:a30:1 --cpus-per-task=64 --mem=0 jobscript

If the data resides on (for example) humus217 but not on humus216, you should specify which node the job should execute on with `--nodelist`. Don't omit to specify the CPU, GPU and memory you need though! For example:

sbatch --nodelist=humus217 --time=6:0:0 --gres=gpu:a30:1 --cpus-per-task=64 --mem=0 jobscript

If you do not need the GPU for a certain calculation, or you need less than an entire node's CPUs and memory, please specify only what you need. The scheduler can run more than one such job on the node at a time. For example:

sbatch --time=24:0:0 --cpus-per-task=1 --mem=4G jobscript