Pegasus HPC Cluster

Image
Pegasus banner

   Get Started

   Features

  • GW's flagship High Performance Computing (HPC) cluster is Pegasus, a shared system supporting a broad array of supercomputing and data intensive workloads
  • Pegasus consists of 200+  nodes accessible through 4 high available login nodes. The robust Dell built cluster utilizes R740s and C4140 and can be broken down by compute, GPU (small and large), high memory and high throughput nodes. All nodes are loaded with CentOS 8 and utilize the SLURM job scheduler. With the cluster capable of a total of 2.14 PFLOPs Single Precision.
  • State-of-the-Art GPU Systems: Incorporates the Nvidia Grace Hopper “superchip” and L40S GPU, providing cutting-edge computational power.
  • 10x Higher Performance: Achieves significantly higher performance for advanced deep learning, AI research, computational fluid dynamics and more.
  • Enhanced Memory Capabilities: Supports larger datasets.
  • High-Speed, Low-Latency Components: Optimizes computing power with advanced networking, memory, storage, and file systems.
  • Improved Processing Speed: Enables faster processing of complex simulations and computations.
  • Simultaneous High-Memory Applications: The Grace Hopper Superchip's memory bandwidth, combined with the AI and graphics capabilities of the L40S GPU, allows multiple high-memory applications to run simultaneously without performance degradation.
     
Core Counts

 

8,112
Total CPU Cores
Not including Login Nodes or Head Nodes

76,800
Total NVIDIA Tensor Cores
Not including Login Nodes or Head Nodes

614,400
Total NVIDIA CUDA Cores
Not including Login Nodes or Head Nodes

  • Compute Nodes (164 Total) Dual 20-Core 3.70GHz Intel Xeon Gold 6148 processors
    • 6,560 cpu cores
  • GPU Nodes :Small GPU (16 Total) Dual 20-Core 3.70GHz Intel Xeon Gold 6148 processors & Two (2) NVIDIA Tesla V100 GPU
    • 640 cpu cores
    • 20,480 NVIDIA Tensor Cores
    • 163,840 NVIDIA CUDA® Cores
  • Large GPU (22 Total) Dual 18-Core 3.70GHz Intel Xeon Gold 6140 processors & Four (4) Nvidia Tesla V100 SXM2 16GB GPU
    • 792 cpu cores
    • 56,320 NVIDIA Tensor Cores
    • 450,560 NVIDIA CUDA® Cores
  • High Throughput Node (6 Total) Dual 4-Core 3.70GHz Intel Xeon Gold 5122 processors
    • 48 cpu cores
  • High Memory Node (2 Total) Dual 18-Core 3.70GHz Intel Xeon Gold 6140M processors
    • 72 cpu cores
GPU Nodes

18 One GPU nodes

There are 16 Small GPU nodes in Pegasus. Each of these is a:

  • Dell PowerEdge R740 server
  • (2) NVIDIA Tesla V100 GPU
  • Dual 20-Core 3.70GHz Intel Xeon Gold 6148 processors
  • 192GB of 2666MHz DDR4 ECC Register DRAM
  • 800 GB SSD onboard storage (used for boot and local scratch space)
  • Mellanox EDR Infiniband controller to 100GB fabric

21 Four GPU nodes

There are 21 Large GPU nodes in Pegasus. Each of these is a:

  • Dell C4140 server
  • 6TB NVMe card
  • Four (4) Nvidia Tesla V100 SXM2 16GB GPUs with NVLink enabled
  • Dual 18-Core 3.70GHz Intel Xeon Gold 6140 processors
  • 384GB of 2666MHz DDR4 ECC Register DRAM
  • 800 GB SSD onboard storage (used for boot and local scratch space)
  • Mellanox EDR Infiniband controller to 100GB fabric

2 Eight A100 GPU nodes

There are 2 Eight GPU nodes in Pegasus. Each of these is a:

  • Lenovo ThinkSystem SR670 V2
  • (8) NVIDIA A100 80GB PCIe Gen4 Passive GPU
  • Dual 26 Cores Intel Xeon Gold 5320 26C
  • 512GB TruDDR4 3200 MHz (2Rx4 1.2V) RDIMM
  • 800 GB SSD onboard storage (used for boot and local scratch space)
  • Mellanox EDR Infiniband controller to 100GB fabric

2 GH200 Grace Hopper Superchip nodes

There are 2 Gracehopper superchip nodes in Pegasus. Each of these is a:

  • Quanta S74G-2U Grace Hopper
  • (8) NVIDIA A100 80GB PCIe Gen4 Passive GPU
  • Grace CPU with 72 Arm Neoverse V2 cores, up to 480GB LPDDRX memory
  • Hopper H100 GPU 96GB HBM3 memory
  • 7.68TB E1.S SSD
High Throughput Node

There are 6 High Throughput nodes in Pegasus. Each of these is a:

  • Dell PowerEdge R740 server
  • Dual 4-Core 3.70GHz Intel Xeon Gold 5122 processors
  • 384GB of 2666MHz DDR4 ECC Register DRAM
  • 800 GB SSD onboard storage (used for boot and local scratch space)
  • Mellanox EDR Infiniband controller to 100GB fabric
Memory Nodes

Medium Memory Node

There are 54 Medium Memory Nodes in Pegasus. Each of these is a:

  • Dell PowerEdge R740 server with
  • Dual 20-Core 3.70GHz Intel Xeon Gold 6148 processors
  • 384GB of 2666MHz DDR4 ECC Register DRAM
  • 800 GB SSD onboard storage (used for boot and local scratch space)
  • Mellanox EDR Infiniband controller to 100GB fabric

High Memory Node

There are 2 High Memory Nodes in Pegasus. Each of these is a:

  • Dell PowerEdge R740 server
  • Dual 18-Core 3.70GHz Intel Xeon Gold 6140M processors
  • 3TB of 2666MHz DDR4 ECC Register DRAM
  • 800 GB SSD onboard storage (used for boot and local scratch space)
  • Mellanox EDR Infiniband controller to 100GB fabric
Login & Head Nodes

Login Nodes

There are 4 High Available Login Nodes in Pegasus. Each of these is a:

  • Dell PowerEdge R740 server
  • Dual 16-Core 3.70GHz Intel Xeon Gold 6130 processors
  • 192GB of 2666MHz DDR4 ECC Register DRAM
  • 2TB RAID I HDD onboard storage (used for boot and local scratch space)
  • 40Gb/s Ethernet for the external network
  • Mellanox EDR Infiniband controller to 100GB fabric

Head Nodes

There are 2 High Available Head Nodes to control Pegasus. Each of these is a:

  • Dell PowerEdge R740 server
  • Dual 16-Core 3.70GHz Intel Xeon Gold 6130 processors
  • 192GB of 2666MHz DDR4 ECC Register DRAM
  • 5TB RAID5 HDD onboard storage
  • 40Gb/s Ethernet for the external network
  • Mellanox EDR Infiniband controller to 100GB fabric
Filesystem

For NFS, the cluster utilized Qumulo Cluster with 2PB replicated across

For scratch Lenovo DSS Solution with 2PB Storage.

Note: Neither filesystem is meant for longterm storage and is subject to requests of removal by support staff.



All GW HPC clusters are managed by professional staff in GW Information Technology, with additional support from Columbian College of Arts and Sciences, School of Engineering and Applied Science, School of Public Health, School of Medicine and Health Sciences, and the Offices of the Provost and Vice Provost for Research.

   Support

 

   Available to

  • Researchers

   Related Service Offerings

 

Phone

202-994-4948
24 hours / 7 days a week

Knowledge Base

Explore our knowledge base for how-to articles and guides.

IT Help