Difference between revisions of "Frontenac:Fees"
(→"Dedicated" and "A la carte" Compute access) |
(→Project and Nearline storage) |
||
Line 63: | Line 63: | ||
== Project and Nearline storage == | == Project and Nearline storage == | ||
+ | |||
+ | There are two standard types of storage on the Frontenac file system, both part of the "Hierarchical Storage Management" system. "Project" storage refers to storage immediately accessible on a disk through the GPFS file system. "Nearline" storage refers to data that reside mostly on tape, but are accessible through disk when needed, albeit with a delay in the case of larger amounts of data being accessed. Here is a more detailed explanation: | ||
+ | |||
+ | {| class="wikitable" | '''Difference between "old" SW (Linux) and "new" CAC (Frontenac) clusters''' | ||
+ | |- | ||
+ | |'''Type''' | ||
+ | |'''Explanation''' | ||
+ | |- | ||
+ | | Project | ||
+ | | | ||
+ | * Used for frequently used, "active" data | ||
+ | * Data reside on disk | ||
+ | * standard areas are : /global/home, /global/project | ||
+ | * access is immediate (at the speed of the GPFS system) | ||
+ | * home and project are backed up, scratch is not | ||
+ | |- | ||
+ | | Nearline | ||
+ | | | ||
+ | * Used for infrequently used, "passive" data | ||
+ | * Data reside mostly on tape, with "stubs" on disk | ||
+ | * standard areas are : /global/home, /global/project | ||
+ | * access requires (automatic) retrieval to disk and entails delays depending on data size | ||
+ | * backup policy the same as for Project data | ||
+ | |- | ||
+ | | Intermediate data | ||
+ | | | ||
+ | * Data reside on global or local disk | ||
+ | * subject to frequent purges | ||
+ | * standard areas are : /global/scratch, /lscratch, /tmp | ||
+ | * used for data transactions free of charge (for registered users) | ||
+ | |} |
Revision as of 16:09, 16 January 2019
Contents
Fee Structure @ Frontenac
Frontenac serves as our main compute cluster and is operated through the SLURM scheduler. Until March 31, allocations from the 2018 Resource Allocation Competition of Compute Canada, are running on this cluster. The cluster is not among the allocatable systems for the 2019 Compute Canada allocation round ("RAC2019"). Therefore, the operation of Frontenac will be on a cost-recovery basis from April 1, 2019. This page provides details about the fee structure.
Compute and Storage
The new fee structure for the Frontenac Compute cluster applies both to the usage of CPU's and for storage on disk and tape. The fees are raised per annum basis, but can be pro-rated to a shorter duration without penalty. The standard units are :
Type | Unit | Explanation |
CPU usage | core-year (cyr) |
|
Storage | terabyte-year (tbyr) |
|
"High-priority" and "Metered" Compute access
There are two standard types of access to the Frontenac cluster": "High-Priority" access which provides scheduled access to the cluster which will in most cases be "rapid" for smaller jobs, and "Metered" access which uses a standard priority that may entail longer waiting times but will only be charged according to actual usage. In addition we offer special arrangements. Here is a more detailed explanation:
Type | Explanation |
High-Priority |
|
Metered |
|
Special arrangements" |
|
Project and Nearline storage
There are two standard types of storage on the Frontenac file system, both part of the "Hierarchical Storage Management" system. "Project" storage refers to storage immediately accessible on a disk through the GPFS file system. "Nearline" storage refers to data that reside mostly on tape, but are accessible through disk when needed, albeit with a delay in the case of larger amounts of data being accessed. Here is a more detailed explanation:
Type | Explanation |
Project |
|
Nearline |
|
Intermediate data |
|