User Tools

Site Tools


hpc:storage_on_hpc

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revisionPrevious revision
Next revision
Previous revision
hpc:storage_on_hpc [2025/02/06 09:37] Adrien Alberthpc:storage_on_hpc [2025/03/14 10:45] (current) – [NASAC] Gaël Rossignol
Line 53: Line 53:
 ===== Scratch Directory ===== ===== Scratch Directory =====
  
-==== Location and Accessibility ====+**Location and Accessibility:** 
 + 
 Your scratch directory is located at: ''$HOME/scratch''. Your scratch directory is located at: ''$HOME/scratch''.
  
   * It is available on both the login node and all compute nodes of the cluster.   * It is available on both the login node and all compute nodes of the cluster.
   * It offers more storage space than ''$HOME''.   * It offers more storage space than ''$HOME''.
 +  * It is not backed up
  
  
Line 64: Line 66:
  
  
-==== Purpose of the Scratch Directory ====+**Purpose of the Scratch Directory:** 
 The scratch directory is intended for storing non-unique or regenerable data. You should use it for: The scratch directory is intended for storing non-unique or regenerable data. You should use it for:
  
Line 71: Line 74:
   * However, this directory is not backed up. Please avoid storing critical data here.   * However, this directory is not backed up. Please avoid storing critical data here.
  
-==== Permissions and Access Control ====+**Permissions and Access Control:** 
  
-Your ''$HOME/scratch'' is **only** accessible by you (permission ''0700'').+  * Your ''$HOME/scratch'' is **only** accessible by you (permission ''0700'')
 +  * Permission modifications are not allowed and will be **automatically** reset. 
 +  * If you need to share files, refer to: [[hpc:storage_on_hpc#sharing_files_with_other_users|Sharing files with other users]].
  
-Permission modifications are not allowed and will be **automatically** reset.+**Best Practices:**
  
-If you need to share files, refer to: [[hpc:storage_on_hpc#sharing_files_with_other_users|Sharing files with other users]]. 
- 
-==== Best Practices ==== 
 The scratch directory is **not a permanent** storage solution. To ensure efficient use: The scratch directory is **not a permanent** storage solution. To ensure efficient use:
  
Line 85: Line 87:
   *  Move important data elsewhere once your project is completed. Unige provides [[https://www.unige.ch/researchdata/en/store/storage-unige/|Storage solution]] which can be mounted on Clusters    *  Move important data elsewhere once your project is completed. Unige provides [[https://www.unige.ch/researchdata/en/store/storage-unige/|Storage solution]] which can be mounted on Clusters 
  
-===== Quota =====+==== Quota ====
 Since the scratch storage is shared among all users, a file count quota is enforced to ensure fair usage: Since the scratch storage is shared among all users, a file count quota is enforced to ensure fair usage:
  
Line 93: Line 95:
  
  
-===== Data Retention Policy =====+==== Data Retention Policy ====
 <note> <note>
 Important: The data retention policy will be implemented on Baobab during the next maintenance from February 18 to 21, 2025. Important: The data retention policy will be implemented on Baobab during the next maintenance from February 18 to 21, 2025.
 [[https://hpc-community.unige.ch/t/important-boabab-new-data-retention-policy-for-scratch-filesystem/3813|More Info]] [[https://hpc-community.unige.ch/t/important-boabab-new-data-retention-policy-for-scratch-filesystem/3813|More Info]]
 </note> </note>
-==== Automatic Deletion Rules ==== 
  
-  Files older than 3 months will be automatically deleted. +**Automatic Deletion Rules:**
-  * Deletion is based on the last access date of each file.+
  
-==== What This Means for You ====+  * Files **older than 3 months** will be automatically deleted. 
 +  * Deletion is based on the last access (read or writte) date of each file.
  
-Any file not accessed within the last 3 months will be considered inactive and deleted.+**What This Means for You:**
  
-Frequently used files will remain unaffected.+  * Any file not accessed within the last 3 months will be considered inactive and deleted. 
 +  * Frequently used files will remain unaffected.
  
 By following these guidelines, you can ensure efficient use of the scratch storage while avoiding data loss. By following these guidelines, you can ensure efficient use of the scratch storage while avoiding data loss.
Line 167: Line 169:
   * in "scratch" (''/srv/beegfs/scratch/shares/'') - to share datasets for instance    * in "scratch" (''/srv/beegfs/scratch/shares/'') - to share datasets for instance 
  
-If you need one, please contact us by email : +If you need one, please fill the form on DW: https://dw.unige.ch/openentry.html?tid=hpc  
-  * send the email to [[hpc@unige.ch]] with your PI in cc + 
-  * provide the following : +If you are an Outisder user and you don't have access to DW,  please request to your PI to fill the form.
-    * do you need a "home" or "scratch" shared folder (or both) ? +
-    * the list of people (email and username) who needs to access the shared folder +
-    * the desired name of the shared folder+
  
 <note important> <note important>
Line 231: Line 230:
  
 <code console> <code console>
-(baobab)-[sagon@login2 ~]$ beegfs-get-quota-home-scratch.sh+(baobab)-[sagon@login1 ~]$ beegfs-get-quota-home-scratch.sh
 home dir: /home/sagon home dir: /home/sagon
 scratch dir: /srv/beegfs/scratch/users/s/sagon scratch dir: /srv/beegfs/scratch/users/s/sagon
Line 304: Line 303:
 ===== NASAC  ===== ===== NASAC  =====
  
-<WRAP round alert 50%> 
-CIFS is not working due to a dummy patch integrated by the fast network supplier (Infiniband by Mellanox/Nvidia). 
  
-2024-12-05 Update: Since Rocky9 was deployed on Bamboo during the last maintenance, the module now seems to be available again. You should be able to mount your NASAC share on both the login nodes and compute nodes. 
- 
-Rocky9 will also be deployed on Baobab and Yggdrasil during the next maintenances. 
- 
-For more information: https://hpc-community.unige.ch/t/2024-current-issues-on-hpc-cluster/3245/15 
-</WRAP> 
  
 If you need to mount an external share (NAS for example) on Baobab from command line, you can proceed as  If you need to mount an external share (NAS for example) on Baobab from command line, you can proceed as 
Line 320: Line 311:
  
 <code console> <code console>
-[sagon@login2 ~] $ dbus-launch bash+[sagon@login1 ~] $ dbus-launch bash
 </code> </code>
  
Line 326: Line 317:
  
 <code console> <code console>
-[sagon@login2 ~] $ gio mount smb://server_name/share_name+[sagon@login1 ~] $ gio mount smb://server_name/share_name
 </code> </code>
  
Line 342: Line 333:
  
 <code console> <code console>
-[sagon@login2 ~] $ gio mount -u smb://server_name/share_name+[sagon@login1 ~] $ gio mount -u smb://server_name/share_name
 </code> </code>
  
-<note important>The data are only available on the login2 node.  +<note important>The data are only available where gio has been mounted.  
-If you need to access the data on the nodes, you need to mount them there as well in your sbatch script.</note>+If you need to access the on other nodes, you need to mount them there as well in your sbatch script.</note>
  
 If you need to script this, you can put your credentials in a file in your home directory. If you need to script this, you can put your credentials in a file in your home directory.
Line 359: Line 350:
 Mount example using credentials in a script: Mount example using credentials in a script:
 <code console> <code console>
-[sagon@login2 ~] $ gio mount smb://server_name/share_name < .credentials+[sagon@login1 ~] $ gio mount smb://server_name/share_name < .credentials
 </code> </code>
  
Line 373: Line 364:
  
 <code console> <code console>
-[sagon@login2 ~] $ ps ux | grep -e '[g]vfsd-fuse'+[sagon@login1 ~] $ ps ux | grep -e '[g]vfsd-fuse'
 sagon    196919  0.0  0.0 387104  3376 ?        Sl   08:49   0:00 /usr/libexec/gvfsd-fuse /home/sagon/.gvfs -f -o big_writes sagon    196919  0.0  0.0 387104  3376 ?        Sl   08:49   0:00 /usr/libexec/gvfsd-fuse /home/sagon/.gvfs -f -o big_writes
 </code> </code>
Line 382: Line 373:
  
 <code console> <code console>
-[sagon@login2 ~] $ pgrep -a -U $(id -u) dbus+[sagon@login1 ~] $ pgrep -a -U $(id -u) dbus
 196761 /usr/bin/dbus-daemon --fork --print-pid 4 --print-address 6 --session 196761 /usr/bin/dbus-daemon --fork --print-pid 4 --print-address 6 --session
 224317 /usr/bin/dbus-daemon --fork --print-pid 4 --print-address 6 --session 224317 /usr/bin/dbus-daemon --fork --print-pid 4 --print-address 6 --session
Line 426: Line 417:
 ====== Robinhood ====== ====== Robinhood ======
 Robinhood Policy Engine is a versatile tool to manage contents of large file systems. It daily scans the scratch beegfs filesystems. It makes it possible to schedule mass action on filesystem entries by defining attribute-based policies. Robinhood Policy Engine is a versatile tool to manage contents of large file systems. It daily scans the scratch beegfs filesystems. It makes it possible to schedule mass action on filesystem entries by defining attribute-based policies.
 +<WRAP center round important 60%>
 +We are working on the newer functionality needed to enforce our scratch data retention policy, the report are out of date until further notice
 +</WRAP>
  
 ==== Policies ==== ==== Policies ====
hpc/storage_on_hpc.1738834643.txt.gz · Last modified: 2025/02/06 09:37 by Adrien Albert