What is SCRP?

SCRP is a high-performance computing cluster managed by the Department of Economics of The Chinese University of Hong Kong, designed to serve both research and teaching roles. Although miniscule in size when compared to most HPC clusters, it utilizes many of the same technology such as fast Infiniband interconnect, parallel storage and multi-node workload management.

Design

SCRP’s design is similar to most HPC cluster and consists of four types of computer servers:

  • Login nodes handle user login and light computation.
  • Compute nodes handle heavy computation.
  • Storage nodes handle file storage.
  • Management node handles background tasks.

SCRP Struture

As a user, you will spend most of your time interacting with the login nodes—tasks such as file transfer and coding should all be done on them. As all nodes share a common file system, your files will be available everywhere even though you are only uploading them to a login node. See Account and Access for details.

You can do light computations on the login nodes, but since they are shared by all users they are not suitable for heavy use. For the latter you will want to run your code on a compute node. SCRP uses the Slurm Workload Manager to allocate computational resources to users. The guides to available software have more details on how to use Slurm.

Hardware Specifications

SCRP consists of 13 nodes in total:

  • scrp-login-1
    • AMD EPYC 7542 (32 cores, 64 threads)
    • 256GB DDR4-2666 ECC RAM
    • 500GB SATA SSD x 2 + 10TB SATA HDD x 12
    • Mellanox Connect-IB Single 56Gb/s NIC
  • scrp-login-2
    • AMD EPYC 7302 (16 cores, 32 threads)
    • 128GB DDR4-3200 ECC RAM
    • 500GB SATA SSD x 2
    • Mellanox ConnectX-3 Single 56Gb/s NIC
  • scrp-node-1
    • AMD EPYC 7742 (64 cores, 64 threads)
    • 256GB DDR4-3200 ECC RAM
    • NVIDIA GeForce RTX 3060 x 3
    • 500GB SATA SSD + 2TB NVMe SSD
    • Mellanox ConnectX-3 Single 56Gb/s NIC
  • scrp-node-[2,6]
    • AMD EPYC 7742 (64 cores, 64 threads)
    • 512GB DDR4-3200 ECC RAM
    • NVIDIA GeForce RTX 3060 x 4
    • 500GB SATA SSD + 2TB NVMe SSD
    • Mellanox ConnectX-3 Single 56Gb/s NIC
  • scrp-node-[3,9]
    • AMD EPYC 7742 (64 cores, 64 threads)
    • 256GB DDR4-3200 ECC RAM
    • NVIDIA GeForce RTX 3090 x 4 with NVLink
    • 500GB SATA SSD + 2TB NVMe SSD
    • Network Interface Cards:
      • Mellanox ConnectX-6 Single 200Gb/s NIC (direct connection between node-3 and node-9)
      • Mellanox Connect-IB Single 56Gb/s NIC
  • scrp-node-[4,5,8]
    • AMD Ryzen 5950X (16 cores, 16 threads)
    • 64GB DDR4-3200 ECC RAM
    • 1TB NVMe SSD
    • Mellanox Connect-IB Single 56Gb/s NIC
  • scrp-node-7
    • AMD Ryzen Threadripper PRO 3995WX (64 cores, 128 threads)
    • 1TB DDR4-3200 ECC RAM
    • 250GB SATA SSD + 2TB NVMe SSD x 2
    • Mellanox Connect-IB Single 56Gb/s NIC
  • scrp-node-10
    • AMD EPYC 7542 (32 cores, 32 threads)
    • 512GB DDR4-3200 ECC RAM
    • NVIDIA A100 x 2 with NVLink
    • 500GB SATA SSD + 2TB NVMe SSD
    • Mellanox Connect-IB Single 56Gb/s NIC
  • scrp-control is the cluster’s management node.
    • AMD EPYC 7302 (16 cores, 32 threads)
    • 128GB DDR4-2666 ECC RAM
    • Storage Disks:
      • 4TB Intel P4510 NVMe SSD x 2
      • 8TB Intel P4510 NVMe SSD x 2
    • Network Interface Cards:
      • Mellanox Connect-IB Dual 56Gb/s NIC x 2
      • Mellanox ConnectX-3 Single 56Gb/s NIC x 1
  • scrp-data is the cluster’s distrubted storage node.
    • AMD EPYC 7302 (16 cores, 32 threads)
    • 256GB DDR4-3200 ECC RAM
    • Storage Disks:
      • 1.6TB Intel P4600 NVMe SSD x 2
      • 8TB Intel P4510 NVMe SSD x 6
    • Network Interface Cards:
      • Mellanox Connect-IB Dual 56Gb/s NIC x 2
      • Mellanox ConnectX-3 Single 56Gb/s NIC x 2

Staff

Director of Economic Computing Services: Dr. Vinci Chow (vincichow@cuhk.edu.hk)

Assistant Computer Officer: Gary Yeung

Computer Technician: Paul Chan