Storrs HPC Cluster
The Storrs HPC cluster currently comprises over 23,000 cores, spread among 400 nodes, with two CPUs per node. The nodes include four generations of Intel CPUs: SandyBridge, IvyBridge, Haswell, and Skylake, and the latest AMD Epyc CPU. We also have a small number of Intel Broadwell nodes and Intel Phi nodes (see table below).
Hi-speed parallel storage is provided; 220TB of scratch storage and 3.65PB of persistent storage, including archive storage.
We have four types GPU nodes available, with a total of 135 general-purpose GPUs, listed in a table below.
The Red Hat RHEL8 operating system runs on the Skylake nodes and newer. The remaining nodes will be upgraded to RHEL8 soon.
The Slurm scheduler manages jobs. Network traffic travels over Ethernet at 10 or 25Gb per second between nodes, and file data travels over Infiniband at 56Gb or 100Gb per second, depending on the node. The nodes are each connected via our Infiniband network to over 3.65PB of parallel storage, managed by WekaIO.
The Storrs HPC cluster is supported by three full-time staff, and two or more student workers. Scientific applications are installed as needed; to date over 200 have been made available.
GPU Node Details
|GPU Type||CPU Type||Number of Nodes||Cores per Node||Total Cores||GPUs/node||Total GPUS|
|NVidia Tesla V100||Skylake||2||36||72||1 to 3||18|
|NVidia Tesla GTX 1080 Ti||Skylake||11||20||220||1 to 3||37|
|NVidia Tesla RTX 2080 TI||Skylake||10||32||320||8||40|
|NVidia Tesla A100||AMD Epyc||16||64||1024||1 to 3||28|
The Condo Model
In the Condo Model, researchers who fund nodes get priority access. However, if their priority job queue becomes idle, unprivileged jobs may run instead. Once started, unprivileged jobs can run for up to twelve hours before they stop. So although priority jobs could wait twelve hours to start, typically most priority jobs wait less than an hour. Futhermore, if priority users keep their job queue full, their jobs will not wait at all.
You can read more about the “Condo Model” on the HPC Knowledge Base.
Last updated December 2, 2022