About the Institutional Cluster (IC) at the SDCC.

Prerequisites:

  • Have a valid account with the SDCC
  • Have a valid account in slurm
    • Your liaison should contact us with your name or user id via the ticketing system

Cluster Information:

The cluster consists of:

  • 216 worker nodes
  • 2 submit nodes
  • 2 master nodes

The worker nodes detail:

  • HPE ProLiant XL190r Gen9
  • 2 CPUs Intel(R) Xeon(R) CPU E5-2695 v4 @ 2.10GHz
  • NUMA node0 CPU(s): 0-8,18-26
  • NUMA node1 CPU(s): 9-17,27-35
  • Thread(s) per core: 1
  • Core(s) per socket: 18
  • Socket(s): 2
  • NUMA node(s): 2
  • 2x Nvidia K80 (108 worker nodes) or 2x P100 (108 worker nodes) per node (4 K80  or 2 P100 devices  per node)
  • 256 GB Memory
  • InfiniBand EDR connectivity

Storage:

  • 1.9 TB of local disk storage per node
  • 1 PB of GPFS distributed storage
  • Cluster Storage

Partitions:

partition time limit allowed qos default time preempt mode user availability
debug 30 minutes normal 5 minutes off ic
long 24 hours normal 5 minutes off ic
scavenger 6 hours scavenger 5 minutes cancel ic
sky 24 hours sky 5 minutes off lqcd only
gen3 72 hours gen3 2 hours off cfn only
gen4 72 hours gen4 2 hours off cfn only


Debug partition:

The debug partition consists of 2 x HPE ProLiant XL190r Gen9 with Nvidia K80

Limits:

  • Each user can submit a maximum of 50 jobs
  • The maximum number of nodes running jobs for an account varies with the size of the allocation.

Licensed Software

The institutional cluster offers only 2 licensed software:

  • Intel Parallel Studio XE
  • PGI Compiler

Any other licensed software that you may see available via modules is installed and maintained by various groups and SDCC/CSI does not offer support for it.