Switched-Off Filesystems (Outdated)¶
Warning
This page is deprecated! All filesystems mentioned on this page are decommissioned.
Workspaces¶
Workspace Lifetimes¶
The filesystems warm_archive, ssd, and scratch were decommissioned by the end of 2023. Do not use
them anymore!
Filesystem (use with parameter --filesystem <filesystem>) |
Duration, days | Extensions | Filesystem Feature | Remarks |
|---|---|---|---|---|
scratch (default) |
100 | 10 | fs_lustre_scratch2 |
Scratch filesystem (/lustre/scratch2, symbolic link: /scratch) with high streaming bandwidth, based on spinning disks |
ssd |
30 | 2 | fs_lustre_ssd |
High-IOPS filesystem (/lustre/ssd, symbolic link: /ssd) on SSDs. |
warm_archive |
365 | 2 | 30 | fs_warm_archive_ws |
Node Features for Selective Job Submission¶
The nodes in our HPC system are becoming more diverse in multiple aspects, e.g, hardware, mounted storage, software. The system administrators can describe the set of properties and it is up to you as user to specify the requirements. These features should be thought of as changing over time (e.g., a filesystem get stuck on a certain node).
A feature can be used with the Slurm option -C, --constraint=<ARG> like
srun --constraint="fs_lustre_scratch2" [...] with srun or sbatch.
Multiple features can also be combined using AND, OR, matching OR, resource count etc.
E.g., --constraint="fs_beegfs|fs_lustre_ssd" requests for nodes with at least one of the
features fs_beegfs and fs_lustre_ssd. For a detailed description of the possible
constraints, please refer to the Slurm documentation.
Hint
A feature is checked only for scheduling. Running jobs are not affected by changing features.
Filesystem Features¶
A feature fs_* is active if a certain (global) filesystem is mounted and available on a node.
Access to these filesystems is tested every few minutes on each node and the Slurm features are
set accordingly.
| Feature | Description | Workspace Name |
|---|---|---|
fs_lustre_scratch2 |
/scratch mounted read-write (mount point is /lustre/scratch2) |
scratch |
fs_lustre_ssd |
/ssd mounted read-write (mount point is /lustre/ssd) |
ssd |
fs_warm_archive_ws |
/warm_archive/ws mounted read-only |
warm_archive |
fs_beegfs_global0 |
/beegfs/global0 mounted read-write |
beegfs_global0 |
fs_beegfs |
/beegfs mounted read-write |
beegfs |
Hint
For certain projects, specific filesystems are provided. For those, additional features are available, like fs_beegfs_<projectname>.
Warm Archive¶
Warm Archive is End of Life
The warm_archive storage system will be decommissioned for good together with the Taurus
system (end of 2023). Thus, please do not use warm_archive any longer and migrate you
data from warm_archive to the new filesystems. We provide a quite comprehensive
documentation on the
data migration process to the new filesystems.
You should consider the new walrus storage as an substitue for jobs with moderately low
bandwidth, low IOPS.
The warm archive is intended as a storage space for the duration of a running HPC project. It does not substitute a long-term archive, though.
This storage is best suited for large files (like tgzs of input data data or intermediate results).
The hardware consists of 20 storage nodes with a net capacity of 10 PiB on spinning disks. We have seen an total data rate of 50 GiB/s under benchmark conditions.
A project can apply for storage space in the warm archive. This is limited in capacity and duration.
Access¶
As Filesystem¶
On ZIH systems, users can access the warm archive via workspaces). Although the lifetime is considerable long, please be aware that the data will be deleted as soon as the user's login expires.
Attention
These workspaces can only be written to from the login or export nodes. On all compute nodes, the warm archive is mounted read-only.
S3¶
A limited S3 functionality is available.