Hardware Information¶
The following hardware summaries may be useful for selecting partitions for workflows and for grant proposal writing. If any information is missing that would be helpful to you, please be sure to contact us or create an issue on our tracker.
Tip
The tables in this section are wide and can be scrolled horizontally to display more information.
Cheaha HPC Cluster¶
Summary¶
The table below contains a summary of the computational resources available on Cheaha and relevant Quality of Service (QoS) Limits. QoS limits allow us to balance usage and ensure fairness for all researchers using the cluster. QoS limits are not a guarantee of resource availability.
In the table, Slurm partitions are grouped by shared QoS limits on cores, memory, and GPUs. Node limits are applied to partitions independently. All limits are applied to researchers independently.
Examples of how to make use of the table:
- Suppose you submit 30 jobs to the "express" partition, and suppose each job needs 10 cores each. Hypothetically, in order for all of the jobs to start at once, 300 cores would be required. The QoS limit on cores is 264 on the "express" partition, so at most 26 jobs (260 cores) can start at once. The remaining 4 jobs will be held in queue, because starting one more would go beyond the QoS limit (270 > 264).
- Suppose you submit 5 jobs to the "medium" partition and 5 to the "long" partition, each requiring 1 node. Then, 10 total nodes would be needed. In this case, it is possible that all 10 jobs can start at once because partition node limits are separate. If all 5 jobs start, jobs on the "medium" partition.
- Suppose you submit 5 jobs to the "amperenodes" partition and 5 to "amperenodes-medium", for a total of 10 A100 GPUs. Additionally, you also submit 4 jobs to the "pascalnodes" partition totaling 8 P100 GPUs. Then 4 of the "gpu: ampere" group jobs can start at once, because the QoS limit is 4 GPUs there. Additionally, all 4 of the "gpu: pascal" group jobs, because the QoS limit is 8 GPUs there. In this case, the QoS for each group is separate.
Partition | Time Limit in Hours | Nodes (Limit/Partition) | Cores/Node (Limit/Person) | Mem GB/Node (Limit/Person) | GPU/Node (Limit/Person) |
---|---|---|---|---|---|
cpu: amd | |||||
amd-hdr100 | 150 | 34 (5) | 128 (264) | 504 (3072) | |
cpu: intel | |||||
express | 2 | 51 (~) | 48 (264) | 754 (3072) | |
short | 12 | 51 (44) | 48 (264) | 754 (3072) | |
medium | 50 | 51 (44) | 48 (264) | 754 (3072) | |
long | 150 | 51 (5) | 48 (264) | 754 (3072) | |
gpu: ampere | |||||
amperenodes | 12 | 20 (TBD) | 32 (64) | 189 (384) | 2 (4) |
amperenodes-medium | 48 | 20 (TBD) | 32 (64) | 189 (384) | 2 (4) |
gpu: pascal | |||||
pascalnodes | 12 | 18 (~) | 28 (56) | 252 (500) | 4 (8) |
pascalnodes-medium | 48 | 7 (~) | 28 (56) | 252 (500) | 4 (8) |
mem: large | |||||
largemem | 50 | 13 (10) | 24 (290) | 755 (7168) | |
largemem-long | 150 | 5 (10) | 24 (290) | 755 (7168) |
The full table can be downloaded here.
Details¶
Detailed hardware information, including processor and GPU makes and models, core clock frequencies, and other information for current hardware are in the table below.
Generation | Compute Type | Total Cores | Total Memory Gb | Total Gpus | Cores Per Node | Cores Per Die | Dies Per Node | Die Brand | Die Name | Die Frequency Ghz | Memory Per Node Gb | Gpu Per Node | Gpu Brand | Gpu Name | Gpu Memory Gb | Nodes |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1 | cpu: amd | 128 | 1024 | 2 | 1 | 2 | AMD | Opteron 242 | 1.6 | 16 | 64 | |||||
2 | cpu: intel | 192 | 1152 | 8 | 4 | 2 | Intel | Xeon E5450 | 3 | 48 | 24 | |||||
3 | cpu: intel | 384 | 1536 | 12 | 6 | 2 | Intel | Xeon X5650 | 2.66 | 48 | 32 | |||||
3 | cpu: intel | 192 | 1536 | 12 | 6 | 2 | Intel | Xeon X5650 | 2.66 | 96 | 16 | |||||
4 | cpu: intel | 48 | 1152 | 16 | 8 | 2 | Intel | Xeon X5650 | 2.7 | 384 | 3 | |||||
5 | cpu: intel | 192 | 1152 | 16 | 8 | 2 | Intel | Xeon E2650 | 2 | 96 | 12 | |||||
6 | cpu: intel | 336 | 5376 | 24 | 12 | 2 | Intel | Xeon E5-2680 v3 | 2.5 | 384 | 14 | |||||
6 | cpu: intel | 912 | 9728 | 24 | 12 | 2 | Intel | Xeon E5-2680 v3 | 2.5 | 256 | 38 | |||||
6 | cpu: intel | 1056 | 5632 | 24 | 12 | 2 | Intel | Xeon E5-2680 v3 | 2.5 | 128 | 44 | |||||
7 | gpu: pascal | 504 | 4608 | 72 | 28 | 14 | 2 | Intel | Xeon E5-2680 v4 | 2.4 | 256 | 4 | NVIDIA | Tesla P100 | 16 | 18 |
8 | cpu: intel | 504 | 4032 | 24 | 12 | 2 | Intel | Xeon E5-2680 v4 | 2.5 | 192 | 21 | |||||
8 | mem: large | 240 | 7680 | 24 | 12 | 2 | Intel | Xeon E5-2680 v4 | 2.5 | 768 | 10 | |||||
8 | mem: large | 96 | 6144 | 24 | 12 | 2 | Intel | Xeon E5-2680 v4 | 2.5 | 1536 | 4 | |||||
9 | cpu: intel | 2496 | 39936 | 48 | 24 | 2 | Intel | Xeon Gold 6248R | 3 | 768 | 52 | |||||
10 | cpu: amd | 4352 | 17408 | 128 | 64 | 2 | AMD | Epyc 7713 Milan | 2 | 512 | 34 | |||||
11 | gpu: ampere | 2560 | 10240 | 40 | 128 | 64 | 2 | AMD | Epyc 7763 Milan | 2.45 | 512 | 2 | NVIDIA | A100 | 80 | 20 |
1 | cpu: intel | 240 | 960 | 48 | 12 | 4 | Intel | Xeon Gold 6248R | 3 | 192 | 5 | |||||
1 | gpu: ampere | 512 | 4096 | 32 | 128 | 64 | 2 | AMD | Epyc 7742 Rome | 2.25 | 1024 | 8 | NVIDIA | A100 | 40 | 4 |
1 | cpu: intel | 144 | 576 | 48 | 12 | 4 | Intel | Xeon Gold 6248R | 3 | 192 | 3 | |||||
1 | gpu: ampere | 512 | 4096 | 32 | 128 | 64 | 2 | AMD | Epyc 7742 Rome | 2.25 | 1024 | 8 | NVIDIA | A100 | 40 | 4 |
The full table can be downloaded here.
The table below is a theoretical analysis of FLOPS (floating point operations per second) based on processor instructions and core counts, and is not a reflection of efficiency in practice.
Generation | Cpu Tflops Per Node | Gpu Tflops Per Node | Tflops Per Node | Nodes | Tflops |
---|---|---|---|---|---|
7 | 1.08 | 17.06 | 18.14 | 18 | 326.43 |
8 | 0.96 | 0.96 | 21 | 20.16 | |
8 | 0.96 | 0.96 | 10 | 9.60 | |
8 | 0.96 | 0.96 | 4 | 3.84 | |
9 | 2.30 | 2.30 | 52 | 119.81 | |
10 | 4.10 | 4.10 | 34 | 139.26 | |
11 | 5.02 | 15.14 | 20.15 | 20 | 403.10 |
Total | 1,022.20 |
The full table can be downloaded here.
For information on using Cheaha, see our dedicated section.
Cloud Service at cloud.rc¶
The Cloud service hardware consists of 5 Intel nodes and 4 DGX-A100 nodes. A description of the available hardware are summarized in the following table.
Fabric | Generation | Compute Type | Partition | Total Cores | Total Memory Gb | Total Gpus | Cores Per Node | Memory Per Node Gb | Nodes | Cpu Info | Gpu Info |
---|---|---|---|---|---|---|---|---|---|---|---|
cloud | 1 | cpu | 240 | 960 | 48 | 192 | 5 | Intel Xeon Gold 6248R 3.00 GHz | |||
cloud | 1 | gpu | 512 | 4096 | 32 | 128 | 1024 | 4 | AMD Epyc 7742 Rome 2.25 GHz | NVIDIA A100 40 GB | |
Total | 752 | 5056 | 32 | 9 |
The full table can be downloaded here.
The table below is a theoretical analysis of FLOPS (floating point operations per second) based on processor instructions and core counts, and is not a reflection of efficiency in practice.
Generation | Cpu Tflops Per Node | Gpu Tflops Per Node | Tflops Per Node | Nodes | Tflops |
---|---|---|---|---|---|
1 | 2.30 | 2.30 | 5 | 11.52 | |
1 | 4.61 | 77.97 | 82.58 | 4 | 330.3 |
Total | 341.82 |
The full table can be downloaded here.
For information on using our Cloud service at cloud.rc, see our dedicated section.
Kubernetes Container Service¶
Important
The Kubernetes fabric is still in deployment and not ready for researcher use. We will be sure to inform you when the service is ready. The following information is planned hardware.
The Kubernetes container service hardware consists of 5 Intel nodes and 4 DGX-A100 nodes. A description of the available hardware are summarized in the following table.
Fabric | Generation | Compute Type | Partition | Total Cores | Total Memory Gb | Total Gpus | Cores Per Node | Memory Per Node Gb | Nodes | Cpu Info | Gpu Info |
---|---|---|---|---|---|---|---|---|---|---|---|
container | 1 | cpu | 144 | 576 | 48 | 192 | 3 | Intel Xeon Gold 6248R 3.00 GHz | |||
container | 1 | gpu | 512 | 4096 | 32 | 128 | 1024 | 4 | AMD Epyc 7742 Rome 2.25 GHz | NVIDIA A100 40 GB | |
Total | 656 | 4672 | 32 | 7 |
The full table can be downloaded here.
The table below is a theoretical analysis of FLOPS (floating point operations per second) based on processor instructions and core counts, and is not a reflection of efficiency in practice.
Generation | Cpu Tflops Per Node | Gpu Tflops Per Node | Tflops Per Node | Nodes | Tflops |
---|---|---|---|---|---|
1 | 2.30 | 2.30 | 3 | 6.91 | |
1 | 4.61 | 77.97 | 82.58 | 4 | 330.3 |
Total | 337.21 |
The full table can be downloaded here.