Difference between revisions of "Allocation"

From CAC Wiki
Jump to: navigation, search
(Comparison: Allocation on SW/CAC (SGE) vs Frontenac (SLURM))
(Comparison: Allocation on SW/CAC (SGE) vs Frontenac (SLURM))
Line 18: Line 18:
 
|'''Frontenac (SLURM)'''
 
|'''Frontenac (SLURM)'''
 
|-
 
|-
| Default allocation  
+
| Default allocation (compute)
 
| 48 core limit
 
| 48 core limit
 
| 50 core years over 1 year
 
| 50 core years over 1 year
 
|-
 
|-
| December 1, 2017
+
| RAC allocation (compute)
 
|  
 
|  
* User notification by email
+
* fixed core limit
* '''All users receive access to new systems'''
+
* according to allocation from Compute Canada
| Frontenac ("new system")
+
* associated with dedicated node
 +
* scheduled to specific node by user
 +
|
 +
* fixed core-years allocation over one year
 +
* allocation from Compute Canada (less than ask!)
 +
* scheduled to "reserved pool" at higher priority
 +
* no dedicated resources
 +
* compete with other users of same priority
 
|-
 
|-
 
| January 3, 2017
 
| January 3, 2017

Revision as of 16:15, 8 February 2018

Resource Allocations on the Frontenac Cluster

This Wiki entry is meant to explain how resources are shared on the CAC Frontenac cluster. This includes default allocations in terms of Compute time as well as extended resources that were allocated by Compute Canada or that come from contributed systems. We also point out differences between the current Frontenac allocation scheme and the older scheme that was used on the now decommissioned SW/CAC clusters.

Fair share vs core/job restrictions

Our job scheduler on Frontenac is [SLURM https://slurm.schedmd.com/]. All resource allocations and limitations are applied through this scheduler. For a basic intro on how to use it, please see our scheduler help file.

Every user on our systems has at least one SLURM account, the default account. Users with access to extended resources have additional accounts corresponding to these allocations. These SLURM accounts have intrinsic restrictions and allow scheduling of jobs up to these limits.


Comparison: Allocation on SW/CAC (SGE) vs Frontenac (SLURM)

Allocation Feature SW/CAC (SGE) Frontenac (SLURM)
Default allocation (compute) 48 core limit 50 core years over 1 year
RAC allocation (compute)
  • fixed core limit
  • according to allocation from Compute Canada
  • associated with dedicated node
  • scheduled to specific node by user
  • fixed core-years allocation over one year
  • allocation from Compute Canada (less than ask!)
  • scheduled to "reserved pool" at higher priority
  • no dedicated resources
  • compete with other users of same priority
January 3, 2017
  • Data synchronization stops
  • User data that differ after this date must be transferred by users
  • Grid Engine scheduling disabled (nodes "draining")
SW ("old system")
January 19, 2018
  • All running jobs are terminated
  • Remaining hardware is transferred to new system
SW ("old system")
January 26, 2018
  • User access to sflogin0/swlogin1 closed
  • SNOlab (SX) cluster jobs terminated
  • SNOlab (SX) login nodes closed
SW ("old system")