At a minimum, the scheduler needs to know:
- How many CPU cores you need, and whether they must be on the same node
- If you don't know, the answer is probably "1 core". If the program supports parallel, it probably supports multithreading - multiple cores on a single node. To use multiple nodes, a program generally needs MPI, so you should only request multiple nodes if you are sure the program can use them.
- How much memory you need
- NYU has nodes with 60GB, 124GB and 252GB available to jobs. (The remaining memory is needed by the operating system)
- How long the job is expected to take
- NYU HPC users can request up to 168 hours (1 week) for a single job. But priority is given to jobs requesting less time
CPUs - nodes and cores
HPC is associated with parallel processes - but it's not magic! To use multiple CPUs, the program must have been written with either threading (eg OpenMP) or message passing (MPI).
If in doubt:
- try with 1 core first
- check the documentation of the software, and next try with multiple cores on 1 node
- when using multiple nodes, check whether the job is actually running on all of the nodes. Contact us for help with this.
How much do I need?
The HPC cluster is not magic, its CPUs are only as fast as any other contemporary CPU. In fact, some nodes on Mercer are a few years old, and may be slower than your desktop (see Clusters July 2017 for a table of node types and when we installed them).
The performance of the HPC cluster comes from its scale. Most nodes in the cluster have 12 or 20 cores, 48GB up to 192GB of RAM, access to a large fast parallel filesystem and there is a 40Gb/s dedicated network link between any two nodes in each of the main groups. And there are thousands of nodes.
So although the resources your job needs depend very much on your job, and there is no simple rule for estimating requirements, you can make some initial guesses based on why you need the HPC cluster:
- My desktop has not enough RAM
You should request at least as much RAM as your desktop possesses. The time required will probably be similar to the time required on your desktop. Be aware though that many problems scale with O(n^2) or more, so doubling the number of data points might require 4x the RAM and 8x the compute time.
- Each run takes 4 hours on my 32GB desktop, and I have 1000 experiments to run
Each experiment will probably take 32GB of memory and 4 hours on the HPC cluster too - but you can submit 1000 of them at once and a few hundred might run simultaneously
For a few one-off jobs, you can safely request much more than you need. When jobs are done, run 'sacct' to check actual usages (run 'sacct -j <jobID> -l ' would show much more information.):
From which your can deduce that the job took 42 minutes of wall clock time and about 647MB of memory. So a sensible resource request for the next job is (see RESOURCES for more about the options):
Slurm schedules jobs based on the resources they expect to need. The default at NYU is 1 CPU core and 2GB of memory for 1 hour. To specify what you job will actually need, use these #SBATCH option:
Options to request compute resources:
--nodes, --ntasks-per-node, --cpus-per-task
Number of nodes, number of tasks per node, and number of CPU core per task required. Default is, one node, one task per node, one CPU core per task, so is 1 CPU core on 1 node
--mem, memory required per node in MegaBytes
--mem-per-cpu, memory required per CPU in MegaBytes
Maximum run time of a job. Time formats include "minutes", "minutes:seconds", "hours:minutes:seconds", "days-hours", "days-hours:minutes" and "days-hours:minutes:seconds".
A job submitted without requesting any specific resources will get the default resource limits for the default partition.
Requesting the resources you need, as accurately as possible, allows your job to be started at the earliest opportunity as well as helping the system to schedule work efficiently to everyone's benefit.
A small amount of memory on each node is needed by the operating system, so for example on a 128GB node, only about 125GB is available to jobs. A job requesting 128GB of memory will therefore be too big for a 128GB node, and Slurm will schedule it on other nodes with larger memory instead. We have fewer nodes with so much memory, so the job is likely to spend longer waiting in the queue. Tip: try requesting 125GB instead.
Nodes on Prince have memory of 62 GB, 125 GB and 250 GB. Requesting a large portion of the memory on a node will cause Slurm to reserve an entire node for your job even if you only request 1 CPU, since there will be insufficient remaining memory to run other jobs.
The serial queues on NYU HPC clusters are limited to a single node, but allow multiple processors on that node to be used. Therefore, parallel jobs using only one node, such as OpenMP or multithreaded jobs can be submitted to a serial queue.
When using more than one node, the job script is executed only on the first node. To make use of the other nodes you must use MPI.
To request GPU nodes:
1 node with 1 core and 1 GPU card
1 node with 2 cores and 2 GPU cards
- --gres=gpu:k80:3 -c3
1 node with 3 cores and 3 GPU cards, specifically the type of Tesla K80 cards. Note that It is always best to request at least as many CPU cores are GPUs
The available GPU node configurations are shown here.
When you request GPUs, the system will set two environment variables - we strongly recommend you do not change these:
To your application, it will look like you have GPU 0,1,.. (up to as many GPUs as you requested). So if for example, there are two jobs from different users: the first one requesting 1 GPU card, the second 3 GPU cards, and they happen landing on the same node gpu-08:
The above boxes shown that there is an overlapped value of '0' between the two jobs for the two environment variables. But the GPU cards assigned are actually different for the jobs, as we should trust UUIDs the nvidia-smi command display to us.
Try submitting a simple job. Here's an example script: