Evaluating object stores with HPC science applications
Object Stores and its HPC supported version:
Swift | DAOS | Ceph | |
---|---|---|---|
Cloud | Openstack Swift | DAOS | RADOS |
HPC | HDF5_Swift-VOL | HDF5_DAOS-VOL | HDF5_RADOS-VOL |
Science Applications that represent typical HPC workload:
BOSS | VPIC | BDCATS | |
---|---|---|---|
Field | Astronomy | Physics | Plasma |
Code | H5BOSS | VPIC-IO | BDCATS-IO |
Testbed:
Swift@NERSC | Boro@Intel | Ceph@NERSC | |
---|---|---|---|
Storage Nodes | 4 | 8 | 4 |
Gateway Nodes | 2 | NA | 2 |
Compute Nodes | 12076 | 66 | 12076 |
Ram per Compute Node(GB) | 96-128 | 128 | 96-128 |
Cores per Compute Node | 32-68 | 32 | 32-68 |
Capacity per Storage Node (TB) | 280 | 2 | 280 |
Cpu | Haswell/KNL | Haswell | Haswell/KNL |
How To Use Now
module load hdf5-parallel/rados
h5pcc -o vpic vpic.c
./vpic
How To Use in the Future:
module load rados/daos/swift
./vpic
Metrics
Use at Performance Tier vs. Lustre
Metrics | Function/Scale |
---|---|
Bandwidth | single node, multi-nodes(scalability) |
IOPS | |
Metadata | File/object open/close/create |
Autentication |
Use at Warm/Cold Tier vs. HPSS?
Metrics | |
---|---|
File Scan | ls |
File Movement | cp, put/get |
File Sharing | |
User-defined Metadata | put/get |
Transition (Posix to Object API)
Metrics |
---|
? |
Admin/ Operation
Metrics |
---|
Live Expandability |
Relibility |
Rebuilding |