Skip to main content

Start Here

What is storage for a supercomputer?

Storage on every Hyak cluster (e.g., Klone and Tillicum) is physically separate. It is best practice on every supercomputer that storage live as its own infrastructure to be high-performance and able to handle the bandwidth I/O and read/write operations required by so many compute nodes attached to it. These are typically parallel file systems (e.g., GPFS, Lustre, BeeGFS).

Storage systems are mounted (i.e., accessible) from every compute node of the cluster. Each Hyak cluster (e.g., Klone and Tillicum) has its own separate parallel file system. The storage attached to each Hyak cluster has its own policies, hierarchy, etc. Please refer to their respective pages for more information:

warning

Important Storage Reminder for Hyak Klone

Hyak Klone does not provide backup, persistent storage, or archival storage. All data on Klone exists as a single copy and is therefore vulnerable to loss due to hardware failure, filesystem issues, facility damage, or natural disasters. Users are solely responsible for transferring important results to external systems (for example, Kopah S3 or Lolo Archive) during the course of their project if persistent or long-term storage is required. Retaining long-term or archival data on Klone is against administrative guidance.

While our storage systems have a track record of stability, it is important to note that STORAGE IS NOT BACKED UP by default on Hyak Klone, while backups are in place for dedicated project storage and home directories on Tillicum. It is the responsibility of the user that in the event of an incident you have a place and plan to restore their data.

On campus storage options:

Best Practices: 3-2-1 Policy

Your data is precious, in some cases completely irreplaceable. The research computing team encourages the use of the widely accepted 3-2-1 backup strategy.

note

3-2-1 is a general IT best practice see the following references: Backblaze, Acronis, and Networkworld.

The 3-2-1 backup policy suggests 3 copies of your data on 2 different types of storage media of which 1 copy is off-site. If you use both Klone Gscratch (i.e., /gscratch) and Lolo then you are already adhering to this best practice: One copy resides in /gscratch on our parallel file system, and if you archive your data to Lolo, two additional copies are created (it does automatic duplication with one copy on UW-Seattle campus and another copy in eastern Washington). /gscratch consists of spinning disk hard drives as a storage medium while Lolo is a tape-based storage medium. Lolo does one of its automatic duplication copies to a geographically remote data center in eastern Washington.

important

You have to copy your data to Lolo to be 3-2-1 compliant or to use your own archive solution, it does not happen automatically. See our Lolo documentation for more information.