CU Research Computing’s Alpine cluster, the third generation CURC High Performance Computing Cluster, launched in May 2022 and as of October 2022 has resources at approximate parity with Summit. The warranty of Summit’s scratch storage ended in September of 2022, so data on
/scratch/summit will be lost if the filesystem fails. CU Boulder users should migrate workflows to Alpine as soon as possible. Non-CU users will soon be able to migrate to Alpine as well.
Summit is the second-generation HPC cluster at University of Colorado Research Computing, following Janus. Summit is an NSF-Funded cluster shared between th University of Colorado, Colorado State University, and institutions in the Rocky Mountain Advanced Computing Consortuium (RMACC).
- From a login node, run “module load slurm/summit” to access the Slurm job scheduler instance for Summit.
- Consult the Table and the Examples section below to learn how to direct your jobs to the appropriate compute nodes.
- If needed, compile your application on a compile node (Compiling and Linking).
All jobs are run through a batch/queue system. Interactive jobs on compute nodes are allowed but these must be initiated through the scheduler. High-priority jobs move to the top of the queue and are thus guaranteed to start running within a few minutes, unless other high-priority jobs are already queued or running ahead of them. High-priority jobs can run for a maximum wall time of 7 days. Low-priority jobs have a maximum wall time of 24 hours.
More details about how to use Slurm can be found here.
Job resources are specified via slurm job scripts, more information and script examples can be found at our batch job page. Summit resources are described in detail at our slurm flags, partitions, and qos page
Couldn’t find what you need? Provide feedback on these docs!