User Tools

Site Tools


cluster_billing

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
cluster_billing [2015/12/02 17:33]
mgstauff [Pricing]
cluster_billing [2017/01/05 19:18] (current)
mgstauff [Storage]
Line 5: Line 5:
  
   - Storage - disk space usage   - Storage - disk space usage
-  - //Slot Groups// - computational slot quotas +  - //Project Slot Quotas// - computational slot quotas 
  
-Users of the CfN cluster are assigned to one or more //Slot Groups//. Each //Slot Group// is assigned to a single //Billing Entity// who is responsible for payment of usage fees.+Users of the CfN cluster are assigned to one or more //Project Slot Quotas//. Each //Project Slot Quota// is assigned to a single //Billing Entity// who is responsible for payment of usage fees.
  
-** There is no charge per user assigned to a //Billing Entity//.** Charges are based just on cumulative disk usage and //Slot Group// quotas shared among any number of users.+** There is no charge per user assigned to a //Billing Entity//.** Charges are based just on cumulative disk usage and //Project Slot Quota// quotas shared among any number of users.
 ==== Billing Entity ==== ==== Billing Entity ====
-A //Billing Entity// is typically a PI or research center that can be billed for cluster usage. A //Billing Entity// is responsible for all disk space and //Slot Groups// assigned to it. Via the //Slot Groups//, any number of cluster users (i.e. people with cluster login accounts) can be included within a //Billing Entity//. //There is no charge for user accounts themselves//. //Slot Groups// can include users that are not part of the //Billing Entity's// lab per se, but rather are collaborating with the lab on one of its projects. Individual users can belong to any number of //Slot Groups// under any number of //Billing Entities//.+A //Billing Entity// is typically a PI or research center that can be billed for cluster usage. A //Billing Entity// is responsible for all disk space and //Project Slot Quotas// assigned to it. Via the //Project Slot Quotas//, any number of cluster users (i.e. people with cluster login accounts) can be included within a //Billing Entity//. //There is no charge for user accounts themselves//. //Project Slot Quotas// can include users that are not part of the //Billing Entity's// lab per se, but rather are collaborating with the lab on one of its projects. Individual users can belong to any number of //Project Slot Quotas// under any number of //Billing Entities//.
  
 ==== Storage ==== ==== Storage ====
 Disk space is organized into project trees, consisting of a directory and all its sub-directories (e.g. /data/jet/mgstauff). Each project tree is assigned to a single //Billing Entity//. A billing entity is responsible for all project trees assigned to it, regardless of file ownership within the tree. Disk space is organized into project trees, consisting of a directory and all its sub-directories (e.g. /data/jet/mgstauff). Each project tree is assigned to a single //Billing Entity//. A billing entity is responsible for all project trees assigned to it, regardless of file ownership within the tree.
  
-Project trees can be setup as a shared directory for all users within a lab and/or within a project. Or, project trees can be setup for each user within a lab or project. However, the shared tree approach is preferred for easier management, maintenance and to facillitate file sharing within labs and projects.+Project trees can be setup as a shared directory for all users within a lab and/or within a project. Or, project trees can be setup for each user within a lab or project. However, the shared tree approach is preferred for easier management, maintenance and to facilitate file sharing within a lab and project.
  
 All storage is on dedicated enterprise-grade hardware RAIDs utilizing RAID-6 for redundancy. All storage is on dedicated enterprise-grade hardware RAIDs utilizing RAID-6 for redundancy.
 +
 +You may want to archive your data off of the cluster to save costs. If so, we have [[dataarchving|some suggestions here]]
  
 === Backup === === Backup ===
  
-All storage is backed up to tape on a quarterly basis and tapes are stored off-site. Users are responsible for maintaining their own copies of original data in the event of catastrophic failure of the system.+All storage is backed up to tape on a near-quarterly basis (at least 3 times per year) and tapes are stored off-site. __Users are responsible for maintaining their own copies of original data in the event of catastrophic failure of the system.__
  
 === Non-cluster Archiving & Cold Storage === === Non-cluster Archiving & Cold Storage ===
 [[dataarchving|See here]] for suggestions on archiving your data off of the CfN cluster. [[dataarchving|See here]] for suggestions on archiving your data off of the CfN cluster.
-==== Slot Groups === +==== Project Slot Quotas === 
-A //Slot Group// is an assignment of SGE slot quotas ([[using_ogs_sge#resources_memory_slots_cpu_cores|see here for details]]) to one or more users. Each //Slot Group// is assigned to exactly one //Billing Entity//, and a //Billing Entity// must have at least one //Slot Group//, but can have any number more as needed.+A //Project Slot Quota// is an assignment of SGE slot quotas ([[using_ogs_sge#resources_memory_slots_cpu_cores|see here for details]]) to one or more users. Each //Project Slot Quota// is assigned to exactly one //Billing Entity//, and a //Billing Entity// must have at least one //Project Slot Quota//
 + 
 +One slot is a quota of 1 cpu core and 6GB RAM (3GB on the 'basic' compute nodes). 
 + 
 +Each //Project Slot Quota// sets an aggregate slot quota for all users assigned to the //Project Slot Quota//. This quota limits the aggregate number of concurrent slots used by all the users in the //Project Slot Quota//. Currently there is a maximum of 200 slots per __Billing Entity__.
  
-Each //Slot Group// sets an aggregate slot quota for all users assigned to the //Slot Group//. This quota limits the aggregate number of concurrent slots used by all the users in the //Slot Group//There is an individual limit of 50 slots for each user with a //Slot Group//. In addition, a smaller quota on sub-groups of individual users can be assigned. Different individual quotas can be assigned for each user if needed, e.g. so a power user can be guaranteed more slots than the casual users in a group.+There is an individual limit of 40 slots for each user with a //Project Slot Quota//. Different individual quotas can be assigned for each user if needed, e.g. so a power user can be guaranteed more slots than the casual users in a group.
  
 For example: For example:
  
-^Slot Group Name | picsl|+^Project Slot Quota Name | picsl|
 ^Users | mgstauff, pcook, jtduda| ^Users | mgstauff, pcook, jtduda|
-^Sub-group1|mgstauff| 
-^Sub-group2|pcook, jtduda| 
 ^Aggregate quota|80| ^Aggregate quota|80|
-^Sub-group1 quota|20| +^mgstauff quota|20| 
-^Sub-group2 quota|50|+^pcook, jtduda quota|40|
  
-In this example, the three users can together never use more than 80 slots at the same time. In addition, user mgstauff can never use more than 20, and users pcook and jtduda can each never user more than 50. Note that even though pcook and jtduda have 50-slot quotas, if pcook were already using 50 slots, then jtduda could only use 30 slots with any jobs he submitted at that time.+In this example, the three users can together never use more than 80 slots at the same time. In addition, user mgstauff can never use more than 20, and users pcook and jtduda can each never user more than 40. Note that even though pcook and jtduda have 40-slot quotas, if pcook were already using 40 slots, and mgstauff were using 15, then jtduda could only use 25 slots with any jobs he submitted at that time.
  
-A //Billing Entity// may decide to create individual //Slot Groups// for power users, so they are not limited by an overall group slot quota.+A //Billing Entity// may decide to create individual //Project Slot Quotas// for power users, so they are not limited by an overall group slot quota.
  
 ==== Users Working for Multiple Labs/PI's === ==== Users Working for Multiple Labs/PI's ===
Line 52: Line 56:
 The data for the different labs/PI's should be in different high-level directory trees. If you currently have all your data for both labs in, e.g., /data/jet/mydir, we'd create a new dir /data/jet/myotherdir for all data belonging to the second lab/PI. Then each dir is tracked separately on our side for usage. Or you can move your data to a shared lab dir if your lab has one (or we can set one up). The data for the different labs/PI's should be in different high-level directory trees. If you currently have all your data for both labs in, e.g., /data/jet/mydir, we'd create a new dir /data/jet/myotherdir for all data belonging to the second lab/PI. Then each dir is tracked separately on our side for usage. Or you can move your data to a shared lab dir if your lab has one (or we can set one up).
  
-=== Slot Groups === +=== Project Slot Quotas === 
-A single user can belong to multiple slot groups assigned to different //Billing Entities// (and also multiple Slot Groups belonging to a single //Billing Entity//). In that case you'll need to run your cluster jobs with an additional parameter to specify which slot group you want to have the particular jobs count against. I'll have details on that in the coming weeks.+A single user can belong to multiple Project Slot Quotas assigned to different //Billing Entities// (and also multiple Project Slot Quotas belonging to a single //Billing Entity//). In that case you'll need to run your cluster jobs with an additional parameter to specify which Project Slot Quota you want to have the particular jobs count against. I'll have details on that in the coming weeks.
  
 ---- ----
Line 67: Line 71:
 It provides: It provides:
   * 150 GB Disk Space   * 150 GB Disk Space
-  * One //Slot Group// with a 16-core quota on the //Secondary Cluster Queue//. This queue consists of about 240 cpu cores (slots) on older compute nodes (from the PICSL cluster) with less memory (32GB / machine) and a slower connection to the cluster frontend and data storage (1Gbit vs 10Gbit).+  * One //Basic Project Slot Quota// with a 16-core quota on the //Basic Cluster Queue//. This queue consists of about 240 cpu cores (slots) on older compute nodes (from the PICSL cluster) with less memory (16GB / machine) and a slower connection to the cluster frontend and data storage (1Gbit vs 10Gbit).
   * Unlimited number of assigned users   * Unlimited number of assigned users
   * Additional disk space is priced as described below.   * Additional disk space is priced as described below.
Line 73: Line 77:
 It does not provide: It does not provide:
   * Flexible per-user sub-quotas   * Flexible per-user sub-quotas
-  * Dyanmic quotas+  * Dynamic quotas
  
 === Storage - $700 / TB / year ( $400 / TB / year above 8TB ) === === Storage - $700 / TB / year ( $400 / TB / year above 8TB ) ===
Line 80: Line 84:
   * Charges are based on average weekly usage over the billing cycle. There's no need to request more space.   * Charges are based on average weekly usage over the billing cycle. There's no need to request more space.
   * Multiple data directories can be assigned to a single //Billing Entity//   * Multiple data directories can be assigned to a single //Billing Entity//
-  * Available to both Basic Accounts and to //Billing Entities// that opt for High-Speed Slot Groups.+  * Available to both Basic Accounts and to //Billing Entities// that opt for High-Speed Project Slot Quotas. 
 +  * You may want to archive your data off of the cluster to save costs. If so, we have [[dataarchving|some suggestions here]]
  
-=== High-speed Slot Groups - $30 / slot / year ===+=== High-speed Project Slot Quotas - $30 / slot / year ===
  
-  * Additional //Slot Groups// give access to the //High-speed Cluster Queue// which utilizes newer high-speed, high-memory compute nodes (64GB), with fast (10Gbit) connection to storage. There is a total of 336 slots in this queue (cpu cores). +  * Additional //High-Speed Project Slot Quotas// give access to the //High-speed Cluster Queue// which utilizes newer high-speed, high-memory compute nodes (64GB), with fast (10Gbit) connection to storage. There is a total of 336 slots in this queue (cpu cores). 
-  * //Slot Groups// are sized in 16-slot increments.+  * High-speed //Project Slot Quotas// are sized in 16-slot increments.
   * Unlimited number of assigned users   * Unlimited number of assigned users
   * Flexible sub-quotas for any number of assigned users (see example above).   * Flexible sub-quotas for any number of assigned users (see example above).
-  * The maximum individual per-user base quota in a slot group is 50 (to help avoid individual users from dominating the cluster)+  * The maximum individual per-user base quota in a High-speed Project Slot Quota is 50 (to help avoid individual users from dominating the cluster)
   * Dynamic quotas -  every user's slot quota is increased when the cluster has idle computing power. Dynamic quota increases are proportional to each user's base quota.   * Dynamic quotas -  every user's slot quota is increased when the cluster has idle computing power. Dynamic quota increases are proportional to each user's base quota.
-  * Quarterly billing for Slot Groups is calculated using the largest number of slots allotted during the quarter.+  * Quarterly billing for Project Slot Quotas is calculated using the largest number of slots allotted during the quarter
 +  * [[slot_usage_reports|Slot Usage Reports]] are provided to help decide how many slots to request for a group.
  
-  * NOTE: We have tried to design the Slot Group system to be simple, flexible, affordable and to help researchers avoid the scenario in which they ask "Should I run this analysis again? How much is it going to cost me!?". We expect most users to change their Slot Group assignments rarely, perhaps once or twice a year at most. If we end up receiving many requests from a user to increase and decrease Slot Group assignments, we will discuss the circumstances with the user to understand why it is necessary. That is, please don't try to game the system.+  * NOTE: We have tried to design the Project Slot Quota system to be simple, flexible, affordable and to help researchers avoid the scenario in which they ask "Should I run this analysis again? How much is it going to cost me!?". We expect most labs and users to rarely change their Project Slot Quota assignments and quotas, perhaps once or twice a year at most. If we end up receiving many requests from a user to increase and decrease Project Slot Quota assignments, we will discuss the circumstances with the user to understand why it is necessary. That is, please don't try to game the system.
 ==== Billing Cycle ==== ==== Billing Cycle ====
  
-Billing will be conducted quarterly and published as a [[disk usage billing report]] and [[slot usage billing report]].+Billing will be conducted quarterly and published as a [[disk usage billing report]] and [[slot usage billing report]]. Slot usage details are reported under [[slot_usage_reports|Slot Usage Reports]].
  
 **NOTE** __The first billing cycle will begin 7/1/2015, with charges applied 10/1/2015. For the first cycle, storage amounts won't be tracked weekly until 9/1/2015, to give labs time to clean up disk space.__ **NOTE** __The first billing cycle will begin 7/1/2015, with charges applied 10/1/2015. For the first cycle, storage amounts won't be tracked weekly until 9/1/2015, to give labs time to clean up disk space.__
Line 103: Line 109:
   * 1-2 regular users, few casual users   * 1-2 regular users, few casual users
   * $100   Mandatory Basic Account   * $100   Mandatory Basic Account
-  * $1440 48-slot //Slot Group//+  * $1440 48-slot High-speed //Project Slot Quota//
   * $1400 2.15 TB total storage   * $1400 2.15 TB total storage
   * ---------   * ---------
Line 111: Line 117:
   * 4 regular users, several casual users   * 4 regular users, several casual users
   * $100  Mandatory Basic Account   * $100  Mandatory Basic Account
-  * $2880 96-slot //Slot Group//+  * $2880 96-slot High-speed //Project Slot Quota//
   * $2800 4.15 TB total storage   * $2800 4.15 TB total storage
   * ---------   * ---------
Line 119: Line 125:
  
 For comparison, you may want to check [[http://www.med.upenn.edu/hpc/billing-pricing-service-descriptions.html|the PMACS HPC services]]. For comparison, you may want to check [[http://www.med.upenn.edu/hpc/billing-pricing-service-descriptions.html|the PMACS HPC services]].
 +
 +
cluster_billing.1449077594.txt.gz · Last modified: 2015/12/02 17:33 by mgstauff