hpc:hpc_clusters
Differences
This shows you the differences between two versions of the page.
Both sides previous revisionPrevious revisionNext revision | Previous revision | ||
hpc:hpc_clusters [2025/01/08 19:50] – [CPUs on Baobab] Gaël Rossignol | hpc:hpc_clusters [2025/03/14 14:17] (current) – [CPUs on Baobab] Gaël Rossignol | ||
---|---|---|---|
Line 89: | Line 89: | ||
You can find the whole table that you can send to the FNS {{: | You can find the whole table that you can send to the FNS {{: | ||
- | ==== Private nodes ==== | + | Users of a given PI are entitled to 100k CPU hours per year free of charge (per PI, not per user). See [[hpc: |
+ | ==== Cost of Renting a Compute Node ==== | ||
- | Research groups can buy " | + | The cost of renting a compute node is calculated based on the vendor price of the node, adjusted |
- | Rules: | + | For example, consider a CPU compute node with a vendor price of **14,361 CHF**. Adding **15% for extra costs** brings the total to **16,515.15 CHF**. Dividing this by 60 months (5 years) results in a monthly rental cost of approximately **275.25 CHF**. |
- | * The compute node is added to the corresponding shared partition, which means that other users can use it when it is not being used by its owner. See [[hpc/ | + | |
- | * In addition to the regular | + | For more details or to request a specific quote, please contact the HPC support team. |
- | * The compute node remains the property of the research group for a period of 5 years. After this time, the node can remain in production but will only be available | + | |
- | * There is a three year warranty | + | ==== Usage Limits ==== |
- | * The research group doesn' | + | |
- | * The compute node is installed and maintained by the HPC team in the same way as the other compute nodes. | + | Users are entitled to utilize up to 60% of the computational resources they own or rent within the cluster. For example, if you rent a compute node with 128 CPU cores for one year, you will receive a total credit of **128 (cores) × 24 (hours) × 365 (days) × 0.6 (max usage rate) = 672,768 core-hours**. This credit can be used across any of our three clusters -- Bamboo, Baobab, and Yggdrasil -- regardless of where the compute node was rented or purchased. |
- | * The HPC team can decide to decommission the node when it is too old, but the hardware | + | |
+ | The main advantage is that you are not restricted to using your private nodes, but can access the three clusters and even the GPUs. | ||
+ | |||
+ | We are developing scripts to allow to check the usage and the amount of hours you have the right to use regarding the hardware your group owns. | ||
+ | |||
+ | The key distinction when using your own resources is that you benefit from a higher scheduling priority, ensuring quicker access to computational resources. As well, you are not restricted to using your private nodes, but can access the three clusters and even the GPUs. | ||
+ | |||
+ | For more details, please contact the HPC support team. | ||
+ | |||
+ | |||
+ | ===== Purchasing or Renting Private Compute Nodes ===== | ||
+ | |||
+ | Research groups have the option to purchase or rent " | ||
+ | |||
+ | ==== Key Rules and Details ==== | ||
+ | |||
+ | * **Shared Integration**: | ||
+ | * **Maximum Usage**: Research groups can utilize up to **60% of the node's maximum theoretical computational capacity**. This ensures fair access to shared resources. See [[hpc: | ||
+ | * **Cost**: | ||
+ | * **Ownership Period**: | ||
+ | * **Warranty and Repairs**: Nodes come with a **3-year warranty**. If the node fails after this period, the research group is responsible for **100% of repair costs**. Repairing | ||
+ | * **Administrative Access**: | ||
+ | * **Maintenance**: | ||
+ | * **Decommissioning**: | ||
+ | |||
+ | ==== CPU and GPU server example pricing ==== | ||
- | Please note that you may as well rent private nodes for a minimal duration of 6 months instead of buying it. | ||
See below the current price of a compute node (without the extra 15% and without VAT) | See below the current price of a compute node (without the extra 15% and without VAT) | ||
Line 113: | Line 138: | ||
* ~ 14' | * ~ 14' | ||
+ | |||
+ | * 2 x 96 Core AMD EPYC 9754 2.4GHz Processor | ||
+ | * 768GB DDR45 4800MHz Memory (24x32GB) | ||
+ | * 100G IB EDR card | ||
+ | * 960GB SSD | ||
+ | * ~ 16'464 CHF TTC | ||
+ | |||
+ | Key differences: | ||
+ | * + 9754 has higher memory performance of up to 460.8 GB/s vs 7763 which has 190.73 GB/s | ||
+ | * + 9754 has a bigger cache | ||
+ | * - 9754 is more expensive | ||
+ | * - power consumption is 400W for 9754 vs 240W for 7763 | ||
+ | * - 9754 is more difficult to cool as the inlet temperature for air colling must be 22° max | ||
=== GPU H100 with AMD=== | === GPU H100 with AMD=== | ||
Line 133: | Line 171: | ||
If you want to ask a financial contribution from UNIGE you must complete a COINF application : https:// | If you want to ask a financial contribution from UNIGE you must complete a COINF application : https:// | ||
+ | |||
+ | ====== Use Baobab for teaching ====== | ||
+ | |||
+ | Baobab, our HPC infrastructure, | ||
+ | |||
+ | Teachers can request access via [dw.unige.ch](final link to be added later, use hpc@unige.ch in the meantime), and once the request is fulfilled, a special account named < | ||
+ | |||
+ | A shared storage space can also be created optionally, accessible at ''/ | ||
+ | |||
+ | **All student usage is free of charge if they submit their job to the correct account**. | ||
+ | |||
+ | We strongly recommend that teachers use and promote our user-friendly web portal at [[hpc: | ||
+ | |||
====== How do I use your clusters ? ====== | ====== How do I use your clusters ? ====== | ||
Line 216: | Line 267: | ||
| V8 | EPYC-7742 | 2.25GHz | 128 cores| " | | V8 | EPYC-7742 | 2.25GHz | 128 cores| " | ||
| V10 | EPYC-72F3 | 3.7GHz | | V10 | EPYC-72F3 | 3.7GHz | ||
+ | | V10 | EPYC-7763 | 2.45GHz | 128 cores| " | ||
| V8 | EPYC-7302P| 3.0GHz | | V8 | EPYC-7302P| 3.0GHz | ||
=== GPUs on Bamboo === | === GPUs on Bamboo === | ||
Line 234: | Line 286: | ||
| V3 | E5-2660V0 | 2.20GHz | 16 cores | "Sandy Bridge-EP" | | V3 | E5-2660V0 | 2.20GHz | 16 cores | "Sandy Bridge-EP" | ||
| V3 | E5-2660V0 | 2.20GHz | 16 cores | "Sandy Bridge-EP" | | V3 | E5-2660V0 | 2.20GHz | 16 cores | "Sandy Bridge-EP" | ||
- | | V3 | E5-2660V0 | 2.20GHz | 16 cores | "Sandy Bridge-EP" | + | | V3 | E5-2660V0 | 2.20GHz | 16 cores | "Sandy Bridge-EP" |
- | | V3 | E5-2670V0 | 2.60GHz | 16 cores | "Sandy Bridge-EP" | + | | V3 | E5-2670V0 | 2.60GHz | 16 cores | "Sandy Bridge-EP" |
- | | V3 | E5-4640V0 | 2.40GHz | 32 cores | "Sandy Bridge-EP" | + | | V3 | E5-4640V0 | 2.40GHz | 32 cores | "Sandy Bridge-EP" |
- | | V4 | E5-2650V2 | 2.60GHz | 16 cores | "Ivy Bridge-EP" | + | | V4 | E5-2650V2 | 2.60GHz | 16 cores | "Ivy Bridge-EP" |
- | | V5 | E5-2643V3 | 3.40GHz | 12 cores | " | + | | V5 | E5-2643V3 | 3.40GHz | 12 cores | " |
- | | V6 | E5-2630V4 | 2.20GHz | 20 cores | " | + | | V6 | E5-2630V4 | 2.20GHz | 20 cores | " |
| V6 | E5-2637V4 | 3.50GHz | 8 cores | " | | V6 | E5-2637V4 | 3.50GHz | 8 cores | " | ||
| V6 | E5-2643V4 | 3.40GHz | 12 cores | " | | V6 | E5-2643V4 | 3.40GHz | 12 cores | " | ||
Line 245: | Line 297: | ||
| V7 | EPYC-7601 | 2.20GHz | 64 cores | " | | V7 | EPYC-7601 | 2.20GHz | 64 cores | " | ||
| V8 | EPYC-7742 | 2.25GHz | 128 cores| " | | V8 | EPYC-7742 | 2.25GHz | 128 cores| " | ||
- | | V9 | GOLD-6240 | 2.60GHz | 36 cores | " | + | | V9 | GOLD-6240 | 2.60GHz | 36 cores | " |
- | | V10 | + | | V10 | EPYC-7763 | 2.45GHz | 128 cores| " |
- | | V11 | + | | V11 | EPYC-9554 | 3.10GHz | 128 cores| " |
hpc/hpc_clusters.1736365839.txt.gz · Last modified: 2025/01/08 19:50 by Gaël Rossignol