hpc:hpc_clusters
Differences
This shows you the differences between two versions of the page.
| Both sides previous revisionPrevious revisionNext revision | Previous revision | ||
| hpc:hpc_clusters [2025/08/21 15:46] – [Cost model] Yann Sagon | hpc:hpc_clusters [2025/12/16 17:04] (current) – [Cost model] Yann Sagon | ||
|---|---|---|---|
| Line 82: | Line 82: | ||
| * **Purchase or rent** compute nodes for more intensive workloads. | * **Purchase or rent** compute nodes for more intensive workloads. | ||
| - | You can as well find a summary of how this model is implemented yet: https:// | + | |
| + | **Summary: | ||
| + | |||
| + | * Starting this year, you receive a **CPU hours credit** based on the hardware you own (if any) in the cluster (private partition). | ||
| + | * You can find instructions on how to check your annual credit here: [[accounting# | ||
| + | * The credit calculation in the provided script assumes a **5-year hardware ownership period**. However, **if** this policy was introduced after your compute nodes were purchased, we have extended the production duration by two years. | ||
| + | * To ensure **flexibility and simplicity**, | ||
| + | * You can use your credit across all three clusters (**Baobab, Yggdrasil, and Bamboo**), not just on your private compute nodes. However, when using your own compute nodes, you will receive a **higher priority**. | ||
| + | * To check your group' | ||
| ==== Price per hour ==== | ==== Price per hour ==== | ||
| <WRAP center round important 60%> | <WRAP center round important 60%> | ||
| Line 98: | Line 106: | ||
| + | === Progressive Pricing for HPC Compute Hours === | ||
| + | A tiered pricing model applies to compute hour billing. Discounts increase as usage grows: once you reach 200K, 500K, and 1,000K compute hours, an additional 10% reduction is applied at each threshold. This ensures cost efficiency for large-scale workloads. | ||
| + | ^ Usage (Compute Hours) ^ Discount Applied ^ | ||
| + | | 0 – 199, | ||
| + | | 200,000 – 499, | ||
| + | | 500,000 – 999, | ||
| + | | 1, | ||
| ===== Purchasing or Renting Private Compute Nodes ===== | ===== Purchasing or Renting Private Compute Nodes ===== | ||
| Line 106: | Line 121: | ||
| * **Shared Integration**: | * **Shared Integration**: | ||
| - | * **Maximum | + | * **Usage |
| - | * **Cost**: In addition to the base cost of the compute node, a **15% surcharge** is applied to cover operational expenses such as cables, racks, switches, and storage. | + | * **Cost**: In addition to the base cost of the compute node, a **15% surcharge** is applied to cover operational expenses such as cables, racks, switches, and storage |
| * **Ownership Period**: The compute node remains the property of the research group for **5 years**. After this period, the node may remain in production but will only be accessible via public and shared partitions. | * **Ownership Period**: The compute node remains the property of the research group for **5 years**. After this period, the node may remain in production but will only be accessible via public and shared partitions. | ||
| * **Warranty and Repairs**: Nodes come with a **3-year warranty**. If the node fails after this period, the research group is responsible for **100% of repair costs**. Repairing the node involves sending it to the vendor for diagnostics and a quote, with a maximum diagnostic fee of **420 CHF**, even if the node is irreparable. | * **Warranty and Repairs**: Nodes come with a **3-year warranty**. If the node fails after this period, the research group is responsible for **100% of repair costs**. Repairing the node involves sending it to the vendor for diagnostics and a quote, with a maximum diagnostic fee of **420 CHF**, even if the node is irreparable. | ||
| Line 119: | Line 134: | ||
| For example, consider a CPU compute node with a vendor price of **14,361 CHF**. Adding **15% for extra costs** brings the total to **16,515.15 CHF**. Dividing this by 60 months (5 years) results in a monthly rental cost of approximately **275.25 CHF**. | For example, consider a CPU compute node with a vendor price of **14,361 CHF**. Adding **15% for extra costs** brings the total to **16,515.15 CHF**. Dividing this by 60 months (5 years) results in a monthly rental cost of approximately **275.25 CHF**. | ||
| + | |||
| + | Currently, we only rent standard AMD compute nodes, which have two 64-CPU cores and 512 GB of RAM. You will receive 1.34 million billing credits per year with this model. | ||
| + | |||
| + | The minimum rental period for a compute node is six months. Any unused allocated resources will be lost at the end of the year. | ||
| For more details or to request a specific quote, please contact the HPC support team. | For more details or to request a specific quote, please contact the HPC support team. | ||
| Line 189: | Line 208: | ||
| We usually install and order the nodes twice per year. | We usually install and order the nodes twice per year. | ||
| - | If you want to ask a financial contribution from UNIGE you must complete a COINF application : https:// | + | If you want to ask a financial contribution from UNIGE you must complete |
| + | |COINF]]. | ||
| ====== Use Baobab for teaching ====== | ====== Use Baobab for teaching ====== | ||
| Line 229: | Line 248: | ||
| Both clusters contain a mix of " | Both clusters contain a mix of " | ||
| - | general | + | general |
| + | |COINF]] | ||
| request compute resources on any node (public and private), but a research group who owns " | request compute resources on any node (public and private), but a research group who owns " | ||
| a higher priority on its " | a higher priority on its " | ||
| Line 283: | Line 303: | ||
| === CPUs on Bamboo === | === CPUs on Bamboo === | ||
| - | ^ Generation ^ Model ^ Freq ^ Nb cores ^ Architecture | + | ^ Generation ^ Model ^ Freq ^ Nb cores ^ Architecture |
| - | | V8 | EPYC-7742 | 2.25GHz | 128 cores| " | + | | V8 | EPYC-7742 | 2.25GHz | 128 cores | " |
| - | | V8 | EPYC-7742 | 2.25GHz | 128 cores| " | + | | V8 | EPYC-7742 | 2.25GHz | 128 cores | " |
| - | | V8 | EPYC-7302P| 3.0GHz | + | | V8 | EPYC-7302P| 3.0GHz |
| - | | V10 | EPYC-72F3 | 3.7GHz | + | | V10 | EPYC-72F3 | 3.7GHz |
| - | | V10 | EPYC-7763 | 2.45GHz | 128 cores| " | + | | V10 | EPYC-7763 | 2.45GHz | 128 cores | " |
| - | | V11 | EPYC-9554 | 3.10GHz | 64 cores | " | + | | V11 | EPYC-9554 | 3.10GHz | 64 cores | " |
| + | | V11 | EPYC-9554 | 3.10GHz | 128 cores | " | ||
| + | | V12 | EPYC-9654 | 3.70GHz | 96 cores | " | ||
| + | | V13 | EPYC-9754 | 3.70GHz | 128 cores | " | ||
| === GPUs on Bamboo === | === GPUs on Bamboo === | ||
| - | ^ GPU model | + | ^ GPU model ^ Architecture ^ Mem |
| - | | RTX 3090 | Ampere | + | | RTX 3090 |
| - | | A100 | Ampere | + | | A100 |
| - | | H100 | Hopper | + | | H100 |
| - | | H200 | Hopper | + | | H200 |
| + | | H200 | Hopper | ||
| + | | RTX Pro 6000 Blackwell | Blackwell | ||
| ==== Baobab ==== | ==== Baobab ==== | ||
| Line 304: | Line 329: | ||
| Since our clusters are regularly expanded, the nodes are not all from the same generation. You can see the details in the following table. | Since our clusters are regularly expanded, the nodes are not all from the same generation. You can see the details in the following table. | ||
| - | ^ Generation ^ Model | + | ^ Generation ^ Model ^ Freq ^ Nb cores ^ Architecture |
| - | | V2 | X5650 | 2.67GHz | 12 cores | " | + | | V5 | E5-2643V3 |
| - | | V3 | E5-2660V0 | 2.20GHz | 16 cores | "Sandy Bridge-EP" | + | | V6 | E5-2630V4 |
| - | | V3 | E5-2660V0 | 2.20GHz | 16 cores | "Sandy Bridge-EP" | + | | V6 | E5-2637V4 |
| - | | V3 | E5-2660V0 | 2.20GHz | 16 cores | "Sandy Bridge-EP" | + | | V6 | E5-2643V4 |
| - | | V3 | E5-2670V0 | 2.60GHz | 16 cores | "Sandy Bridge-EP" | + | | V6 | E5-2680V4 |
| - | | V3 | E5-4640V0 | 2.40GHz | 32 cores | "Sandy Bridge-EP" | + | | V7 | EPYC-7601 |
| - | | V4 | E5-2650V2 | 2.60GHz | 16 cores | "Ivy Bridge-EP" | + | | V8 | EPYC-7742 |
| - | | V5 | E5-2643V3 | 3.40GHz | 12 cores | " | + | | V9 | SILVER-4210R | 2.60GHz | 36 cores | " |
| - | | V6 | E5-2630V4 | 2.20GHz | 20 cores | " | + | | V9 | GOLD-6240 |
| - | | V6 | E5-2637V4 | 3.50GHz | 8 cores | " | + | | V9 | GOLD-6244 |
| - | | V6 | E5-2643V4 | 3.40GHz | 12 cores | " | + | | V10 | EPYC-7763 |
| - | | V6 | E5-2680V4 | 2.40GHz | 28 cores | " | + | | V11 | EPYC-9554 |
| - | | V7 | EPYC-7601 | 2.20GHz | 64 cores | " | + | | V12 | EPYC-9654 |
| - | | V8 | EPYC-7742 | 2.25GHz | 128 cores| " | + | | V12 | EPYC-9654 |
| - | | V9 | GOLD-6240 | 2.60GHz | 36 cores | " | + | |
| - | | V9 | GOLD-6244 | 3.60GHz | 16 cores | “Intel Xeon Gold 6244 CPU” | cpu[351] | + | |
| - | | V10 | EPYC-7763 | 2.45GHz | 128 cores| " | + | |
| - | | V11 | EPYC-9554 | 3.10GHz | 128 cores| " | + | |
| - | | V11 | EPYC-9654 | 3.70GHz | 96 cores | " | + | |
| The " | The " | ||
| Line 352: | Line 372: | ||
| | Titan X | Pascal | | Titan X | Pascal | ||
| | RTX 2080 Ti | Turing | | RTX 2080 Ti | Turing | ||
| - | | RTX 2080 Ti | Turing | + | | RTX 2080 Ti | Turing |
| - | | RTX 2080 Ti | Turing | + | | RTX 2080 Ti | Turing |
| | RTX 2080 Ti | Turing | | RTX 2080 Ti | Turing | ||
| | RTX 3090 | Ampere | | RTX 3090 | Ampere | ||
| Line 382: | Line 402: | ||
| Since our clusters are regularly expanded, the nodes are not all from the same generation. You can see the details in the following table. | Since our clusters are regularly expanded, the nodes are not all from the same generation. You can see the details in the following table. | ||
| - | ^ Generation ^ Model | + | ^ Generation ^ Model |
| - | | V9 | [[https:// | + | | V9 |
| - | | V9 | [[https:// | + | | V9 |
| - | | V8 | EPYC-7742 | + | | V8 |
| - | | V9 | [[https:// | + | | V9 |
| - | | V9 | [[https:// | + | | V9 |
| + | | V12 | EPYC-9654 | ||
| The " | The " | ||
hpc/hpc_clusters.1755791192.txt.gz · Last modified: by Yann Sagon