hpc:storage_on_hpc
Differences
This shows you the differences between two versions of the page.
Both sides previous revisionPrevious revisionNext revision | Previous revision | ||
hpc:storage_on_hpc [2024/12/09 11:10] – [NASAC] Adrien Albert | hpc:storage_on_hpc [2025/03/14 10:45] (current) – [NASAC] Gaël Rossignol | ||
---|---|---|---|
Line 51: | Line 51: | ||
To resume the situation, you should clean up some data in your home directory and or migrate some data to your scratch directory. | To resume the situation, you should clean up some data in your home directory and or migrate some data to your scratch directory. | ||
- | ===== Scratch | + | ===== Scratch |
- | Your '' | + | **Location and Accessibility: |
- | \\ It is available on the login node and on each compute nodes of the cluster. It has more space than '' | + | |
+ | Your scratch | ||
- | **N.B.** : '' | + | |
+ | | ||
+ | * It is not backed up | ||
- | The scratch directory allows you to store any data that is not unique or that can be regenerated. Please use this to store any file that doesn' | ||
- | It is also acceptable to store for instance | + | **N.B.: '' |
- | The content of this folder is persistent and accessible from any node, but there is no backup. | ||
- | Your '' | ||
+ | **Purpose of the Scratch Directory: | ||
- | Also, the scratch directory is not a permanent storage solution, we strongly advise you to move/clean useless/ | + | The scratch directory is intended for storing non-unique or regenerable data. You should use it for: |
+ | |||
+ | * Temporary storage of application-generated files. | ||
+ | * Large datasets used as input during computations. | ||
+ | * However, this directory is not backed up. Please avoid storing critical data here. | ||
+ | |||
+ | **Permissions and Access Control:** | ||
+ | |||
+ | * Your '' | ||
+ | * Permission modifications are not allowed and will be **automatically** reset. | ||
+ | * If you need to share files, refer to: [[hpc: | ||
+ | |||
+ | **Best Practices: | ||
+ | |||
+ | The scratch directory is **not a permanent** storage solution. To ensure efficient use: | ||
+ | |||
+ | * Regularly | ||
+ | * Move important | ||
==== Quota ==== | ==== Quota ==== | ||
+ | Since the scratch storage is shared among all users, a file count quota is enforced to ensure fair usage: | ||
+ | * Maximum file count: 10 million (10M) | ||
- | As the storage is shared by everyone, | + | If you exceed |
+ | |||
+ | |||
+ | ==== Data Retention Policy ==== | ||
+ | < | ||
+ | Important: The data retention policy will be implemented | ||
+ | [[https:// | ||
+ | </ | ||
- | **The maximum file count is currently set to 10M.** | + | **Automatic Deletion Rules:** |
- | What does it mean for you: if the number of files in your scratch space is higher | + | * Files **older |
+ | * Deletion is based on the last access (read or writte) date of each file. | ||
- | Error message: | + | **What This Means for You:** |
- | Disk quota exceeded | + | * Any file not accessed within the last 3 months will be considered inactive and deleted. |
+ | * Frequently used files will remain unaffected. | ||
- | To resume the situation, you should clean up some data in your scratch | + | By following these guidelines, you can ensure efficient use of the scratch |
===== Fast directory ===== | ===== Fast directory ===== | ||
Line 140: | Line 169: | ||
* in " | * in " | ||
- | If you need one, please | + | If you need one, please |
- | * send the email to [[hpc@unige.ch]] with your PI in cc | + | |
- | * provide the following : | + | If you are an Outisder user and you don't have access |
- | * do you need a " | + | |
- | * the list of people (email | + | |
- | * the desired name of the shared folder | + | |
<note important> | <note important> | ||
Line 204: | Line 230: | ||
<code console> | <code console> | ||
- | (baobab)-[sagon@login2 | + | (baobab)-[sagon@login1 |
home dir: /home/sagon | home dir: /home/sagon | ||
scratch dir: / | scratch dir: / | ||
Line 277: | Line 303: | ||
===== NASAC ===== | ===== NASAC ===== | ||
- | <WRAP round alert 50%> | ||
- | GIFS is not working due to a dummy patch integrated by the fast network supplier (Infiniband by Mellanox/ | ||
- | 2024-12-05 Update: Since Rocky9 was deployed on Bamboo during the last maintenance, | ||
- | |||
- | Rocky9 will also be deployed on Baobab and Yggdrasil during the next maintenances. | ||
- | |||
- | For more information: | ||
- | </ | ||
If you need to mount an external share (NAS for example) on Baobab from command line, you can proceed as | If you need to mount an external share (NAS for example) on Baobab from command line, you can proceed as | ||
Line 293: | Line 311: | ||
<code console> | <code console> | ||
- | [sagon@login2 | + | [sagon@login1 |
</ | </ | ||
Line 299: | Line 317: | ||
<code console> | <code console> | ||
- | [sagon@login2 | + | [sagon@login1 |
</ | </ | ||
Line 315: | Line 333: | ||
<code console> | <code console> | ||
- | [sagon@login2 | + | [sagon@login1 |
</ | </ | ||
- | <note important> | + | <note important> |
- | If you need to access the data on the nodes, you need to mount them there as well in your sbatch script.</ | + | If you need to access the on other nodes, you need to mount them there as well in your sbatch script.</ |
If you need to script this, you can put your credentials in a file in your home directory. | If you need to script this, you can put your credentials in a file in your home directory. | ||
Line 332: | Line 350: | ||
Mount example using credentials in a script: | Mount example using credentials in a script: | ||
<code console> | <code console> | ||
- | [sagon@login2 | + | [sagon@login1 |
</ | </ | ||
Line 346: | Line 364: | ||
<code console> | <code console> | ||
- | [sagon@login2 | + | [sagon@login1 |
sagon 196919 | sagon 196919 | ||
</ | </ | ||
Line 355: | Line 373: | ||
<code console> | <code console> | ||
- | [sagon@login2 | + | [sagon@login1 |
196761 / | 196761 / | ||
224317 / | 224317 / | ||
Line 399: | Line 417: | ||
====== Robinhood ====== | ====== Robinhood ====== | ||
Robinhood Policy Engine is a versatile tool to manage contents of large file systems. It daily scans the scratch beegfs filesystems. It makes it possible to schedule mass action on filesystem entries by defining attribute-based policies. | Robinhood Policy Engine is a versatile tool to manage contents of large file systems. It daily scans the scratch beegfs filesystems. It makes it possible to schedule mass action on filesystem entries by defining attribute-based policies. | ||
+ | <WRAP center round important 60%> | ||
+ | We are working on the newer functionality needed to enforce our scratch data retention policy, the report are out of date until further notice | ||
+ | </ | ||
==== Policies ==== | ==== Policies ==== |
hpc/storage_on_hpc.1733742600.txt.gz · Last modified: 2024/12/09 11:10 by Adrien Albert