This articleneeds additional citations forverification. Please helpimprove this article byadding citations to reliable sources. Unsourced material may be challenged and removed. Find sources: "High Performance Storage System" – news ·newspapers ·books ·scholar ·JSTOR(August 2014) (Learn how and when to remove this message) |
High Performance Storage System | |
---|---|
Developer(s) | HPSS Collaboration (IBM,LANL,LBNL,LLNL,ORNL,SNL) |
Stable release | 11.2 / February 2025 |
Operating system | Linux |
Type | Hierarchical Storage Management |
License | Proprietary |
Website | hpss-collaboration |
High Performance Storage System (HPSS) is a flexible,scalable, policy-based, software-definedhierarchical storage management (HSM) product developed bythe HPSS Collaboration. It provides scalable HSM, archive, and file system services using cluster,LAN andstorage area network (SAN) technologies to aggregate the capacity and performance of many computers, disks, disk systems, tape drives, and tape libraries.[1]
HPSS supports a variety of methods for accessing and creating data. Among them are support forFTP, parallel FTP,FUSE (Linux), as well as a robust clientAPI with support for parallel I/O.
As of version 7.5, HPSS has full support onLinux. The HPSS clientAPI is supported onAIX,Linux, andSolaris.[1]
The implementation is built around IBM'sDb2, a scalablerelational database management system.
In early 1992, severalUnited States Department of Energy (DOE)National Laboratories —Lawrence Livermore (LLNL),Los Alamos (LANL),Oak Ridge (ORNL), andSandia (SNL) — joined with IBM to form the National Storage Laboratory (NSL).[2] The NSL's purpose was to commercialize software and hardware technologies that would overcome computing and data storage bottlenecks.[3] The NSL's research on data storage gave birth to the collaboration which produces HPSS. This collaboration began in the fall of 1992[4] and involvedIBM'sHouston Global Services and five DOE national labs (Lawrence Berkeley [LBL], LLNL, LANL, ORNL, and SNL).[1] At that time, the HPSS design team at the DOE national laboratories and IBM recognized there would be a data storage explosion driven by computing power rising toteraflops/petaflops requiring data stored in HSMs to rise to petabytes and beyond, data transfer rates with the HSM to rise to gigabytes/s and higher, and daily throughput with a HSM in tens of terabytes per day. Therefore, the collaboration set out to design and deploy a system that would scale by a factor of 1,000 or more and evolve from the base above toward these expected targets and beyond.[5]
The HPSS collaboration is based on the premise that no single organization has the experience and resources to meet all the challenges represented by the growing imbalance between computing power and data collection capabilities, and storage system I/O, capacity, and functionality. Over twenty organizations worldwide including industry, US Department of Energy (DOE), other federal laboratories, universities,National Science Foundation (NSF) supercomputer centers, FrenchCommissariat a l'Energie Atomique (CEA), and Gleicher Enterprises have contributed to various aspects of this effort.
As of 2022, the primary HPSS development team consists of: