Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

Users of By default, Koa users are granted access to two storage locations by default and could be granted access to two other types of storage locations.The two default storage locations users are given access to is :

  1. home storage (home)

...

  1. scratch storage (Koa Scratch)

...

The two additional types of storage locations could be granted access to include to a user, but are typically tied to a lab or project:

  1. Lab storage (free)

...

  1. KoaStore storage (for fee)

...

...

Table of Contents
maxLevel3
minLevel1
include
outlinefalse
indent
exclude
styledefault
typelist
printabletrue
class

...

Common Storage Locations

Home (Free)

Each user is provided a home storage location on Koa. Each user's home is provided 50 GB of space that can be used as the user chooses. Upon initialization, user homes are setup with an examples directory, a symlink for Koa Scratch and other basic configurations needed to allow a users to take advantage of Koa.

...

Home is not a high performance filesystem. Network File System (NFS) is tried a and true but it is not designed to handle the stresses an High Performance Computing cluster can place on it. As a result, home should not be used for writing output from multiple jobs or jobs that can generate a lot of data. Jobs of this nature should take advantage of Scratch instead, which is a file system designed for the stresses an HPC can generate.

...

Details

Home is utilizing a Zoned File System (ZFS) over NFS Remote Direct Access (RDMA). Currently, home is served out utilizing a single Virtual Host that is connected to Koa at 200 Gbit. ZFS is configured to use zstd-3 compression allowing users to take full advantage of the inline compression and space savings zstd provides. As of this writing, on average users see a 2.16x compression ratio.

File system

Per user storage quotaPer user file limit

File & Dir (inode) Quota

Compression

Persistent

ZFS + NFS v4

50 GB

N/A

zstd-3

Yes

...

...

Scratch (Free)

Scratch, also known by the symlink “koa_scratch”, provides each user access to a an 800TB pool of storage on which files may live 90 days on from the last time they were modified. In total, this file system can support up to 400,000,000 files and directories. Users are not provided an individual quota allowing for flexibility based on need. Scratch directly access accesses the underlying Koa Storage system providing the high performance possible from the storage system.

...

Scratch provides direct access to the underlying high performance file system: Koa Storage, which utilizes Lustre. Lustre is designed for situations where many servers and workloads are needing need to read and write data as quickly as possible. While Lustre works best with long sequential reads and writesread/writes, and exhibits poorer performance with small random reads, methods exist to work around this limitation when the need to read as input a lot of small files, such as using squashfs. Workarounds would include the use of squashfs as we cover in our documentation or archive formats that combine multiple small files that align with your workflow, such as HDF5.

For scalability and performance, files written to Scratch use a progressive file layout in which at certain size boundaries switch to different types of storage medium or recruit more storage targets overall are recruited to storage store parts of a file.

Automatic File Purging

Note

Scratch is not a persistent storage location for users data. Scratch provides a 90-day grace period after a file was last written to before the file is removed automatically from the file system.

Warning

The purge process cannot be paused for individual users and files that are removed cannot be recovered.

Parameters of the automated file purge

  • Only files under ~/koa_scratch/ or /mnt/lustre/koa/scratch/${USER} will be subject to purge

  • Files and folders not modified for *90 days* will be deleted from scratch

  • The purge process will run daily The process cannot be paused for individual users and files that are removed cannot be recovered

Info

In the case the file system

...

nears 85-90%

...

utilized, ITS-CI will contact users who have large occupancy of

...

scratch to voluntarily reduce their usage

...

. If we are unable to reclaim enough space to drop below 70% utilized, we will purge files from oldest to newest, regardless of time on scratch until we

...

are below 70% usage

...

.

Details

Scratch, or more specifically, Koa Storage utilizes Lustre with ZFS as its underlying file system. For transport to Koa, it also utilizes RDMA using multiple servers all connected at 200Gbit Infiniband. The ZFS components are setup to provide the same zstd-3 compression seen on the home file system providing space savings and in some cases faster access to file files as you are needing need to read less data from a slower storage medium. The underlying storage system utilizes a mixture of spinning enterprise hard drives (HDD) and Solid State storage, (SAS SSD and NVMe).

...

Metadata

Meta data Metadata is stored in separate targets from the files file data, which for performance reason reasons is entirely using Solid State Drives. The meta data metadata is split up among multiple targets, which are then served out by different servers. In case of a server failure, “failing-over” the storage to another server is possible allowing for minimal down timedowntime. Each folder is assigned to one of the Meta Data Metadata storage targets and all files under it are also assigned to that target. Load balancing is done in some cases to try and balancing the different meta data metadata targets so they do not grow too out of sync size wise.

Object Data

Object data is stored is storage on a mixture of spinning enterprise hard drives, SAS Solid State Drives and NVMe.

Info

Of the current storage (7 PB), about 1/7th

...

is either solid state (SAS SSD or NVMe

...

)

Storage targets are split up into different configurations providing at least a 2 disk parity. Current Currently, Koa has 58 Object Storage Targets, of which 10 targets are SAS SSD or NVMe.

Object Data data is written to Scratch using a progressive file layout used for scratch is currently setup to follow (PFL). The current PFL setup used by scratch follows the following rules:

  • The first 512K of every file is written to a single solid state (SSD or NVMe) storage target

  • Next, file data up to 64MB is written to a single HDD target

  • Next, file data up to 512MB is written to two HDD targets in 4 MB stripes

  • Next, file data up to 1 GB is written to four HDD targets in 4 MB stripes

  • Finally, any remaining data for a file is written to eight HDD targets in 4 MB stripes

File system

Total Storage Max Files and DirectoriesQuota

File & Dir (inode) Quota

Compression

Persistent

Lustre + ZFS

800 TB

400,000,000

zstd-3

No (90 day purge)

...

...

Per lab or project storage

Lab and purchased storage processing

Lab (Free)

Lab storage is limited to one free storage location per Lab at UH. A lab will be granted 500GB of space and a limit of 512,000 files and directories. The lab will be assigned a posix group which users can be added and removed from upon request. A lab space will allow them to share applications, data and environments between members of the lab.

Info

In the case the lab or Faculty member requesting the space already lease KoaStore storage, the same posix group will be utilized.

Details

Lab storage has direct access to the Koa Storage system, just like Scratch. It follows a different progressive file layout, in which solid state storage is not utilized.

...

KoaStore (For Fee Storage)

KoaStore is storage that is leased year to year on Koa in 1 TB increments with a limit of 1,024,000 files and folders per TB leased. Funds for this storage must come from an account at UH. Access control to this storage for a given group, lab or project is handled through a posix group which users can be added and removed from upon request.

Info

If the lab or faculty member requesting the space already has a lab storage, we will typically utilize the same posix group as the lab storage unless requested otherwise.

Details

Lab storage has direct access to the Koa Storage system, just like Scratch. It follows a different progressive file layout, in which solid state storage is not utilized.

...