Top HPC Storage Vendors
HPE (Cray) ClusterStor E1000 AFA, HPE Apollo 2000, Fujitsu Primergy RX2530 M5, Dell EMC PowerSwitch Z9332F-ON, NetApp E5700, NEC ScaTeFS, Mellanox HPC-X ScalableHPC, Panasas ActiveStor-18
This is a Press Release edited by StorageNewsletter.com on March 19, 2020 at 2:20 pmThis article was written on a blog posted by Ashish Sharma, content marketing manager, MSys Technologies LLC, on March 16, 2020.
High Performance Computing Storage
Hybrid Cloud, Parallel File Systems, Key Challenges, and Top Vendors’ Products
The toughest Terminator, T-1000, can demonstrate rapid shapeshifting, near-perfect mimicry, and recovery from damage.
This is because it is made of mimetic polyalloy with robust mechanical properties. T-1000s naturally require speed, hi-tech communication system, razor-sharp analytical speed, and most powerful connectors and processors. Neural networks are also critical to the functioning of terminators. It stacks an amount of data in nodes, which then communicates with the outer world depending on the input received.
We infer one important thing – these Terminators produce an arduous amount of data. Therefore, it must require a sleek data storage system that scales and carry capabilities to compute massive datasets. Which, rings a bell – just like the case of terminators, HPC also require equally robust storage to maintain compute performance.
Well, HPC has been the nodal force to path defining innovations and scientific discoveries. This is because HPC enables processing of data and powerin complex calculations at the speed of light. To give it a perspective, HPC leverages compute to deliver high performance. The rise of AI/ML, deep learning, edge computing and IoT created a need to store and process incredible amount of data. Therefore, HPC became the key enabler to bring digital technologies within the realm of daily use. In layman’s term, they can be referred as the supercomputers.
Continual Coming of Age of HPC
The first supercomputer – CDC 6600 reigned for 5 years from its inception in 1964. it was paramount to the critical operations of the US government and the US military. It was considered 10x faster to its nearest competitor – IBM 7030 Stretch. Well, it worked with a speed of up to 3 million flops.
The need for complex computer modeling and simulation never stopped over the decades. Likewise, we also witnessed evolution of HPCs. They were made of core-components, which had more power and vast memories to handle complex workloads and analyze datasets. Any new release of supercomputers would make its predecessors obsolete. Just like new robots from the Terminator series.
The latest report by Hyperion Research states that iterative simulation workloads and new workloads such as Al and other big data jobs would be driving the adoption of HPC storage.
Understanding Data Storage as Enabler for HPC
Investing in HPC is exorbitant. Therefore, one must bear in mind that it is essential to have a robust and equally proficient data storage system that runs concurrently with the HPC environment. Further some, HPC workloads differ based on its use cases. For example, HPC at the government and military secret agency consumes heavier workloads versus HPC at a national research facility. This means HPC storage require heavy customization for differential storage architecture, based on its application.
Hybrid Cloud – Optimal Solution for Data-Intensive HPC Storage
Thinking about just the perfect HPC storage will not help. There has to an optimal solution that scales based on HPC needs. Ideally, it has to the right mix of best of the both – traditional storage (on-prem disk drives) and cloud (SSDs and HDDs). Complex, data-intensive IO/s can be channeled to SSDs, while usual streaming data can be handled by disk drives. An efficient combination of hybrid cloud – SDS and hardware configuration ultimately helps scale performance, while eliminating the need to have a storage tier separately. The SDS must come with key characteristics – write back, read persistence performance statistics, dynamic flush, and I/O histogram. Finally, the HPC storage should support parallel file systems by handling complex sequential I/O.
Long Term Solution (LTS) Luster for Parallel File System
More than 50 percent of the global storage architecture prefer Luster – an open-source parallel file system to support HPC clusters. Well, for starters it offers free installation. Further, it provides massive data storage capabilities along with unified configuration, centralized management, simple installation, and powerful scalability. It is built on LTS community release allowing parallel I/O spanning multiple servers, clients, and storage devices. It offers open APIs for deep integration. The throughput is more than 1TB/s. It also offers integrated support for an application built on Hadoop MapReduce applications.
Challenges of Data Management in Hybrid HPC Storage
- Inefficient Data Handling
The key challenge in implementing hybrid HPC storage is inefficient data handling. Dealing with the large and complex dataset and accessing it over WAN is time-consuming and tedious. - Security
It is an another complex affair for HPC storage. The hybrid cloud file system also must include in-built data security. These small files must not be vulnerable to external threats. Providing SMBv3 encryption for files moving within the environment could be a great deal. Further, building the feature of snapshot replication can deliver integrated protection to the data in a seamless manner. - Right HPC product
End users usually find it difficult to choose the right product relevant to their services and industry. Hyperion Research presents an important fact. It states, “Although a large majority (82%) of respondents were relatively satisfied with their current HPC storage vendors, a substantial minority said they are likely to switch storage vendors the next time they upgrade their primary HPC system. The implication here is that a fair number of HPC storage buyers are scrutinizing vendors for competencies as well as price.”
Top HPC storage products
Let’s briefly understand the top varied HPC Storage products in the market.
ClusterStor E1000 All Flash – by Cray (HPE Company)
It enables handling of the data at the speed of exascale. Its core is a combination of SSD and HDD. It is a policy-driven architecture enablin to move data intelligently. ClusterStor E1000 HDD-based configuration offers up to 50% more performance with the same number of drives than its closest competitors. This all-flash configuration is perfect for mainly small files, random access, and terabytes to single-digit petabyte capacity requirements.
HPE Apollo 2000 System – by HPE
Th Apollo 2000 Gen10 system is designed as an enterprise-level, density-optimized, 2U shared infrastructure chassis for up to 4 HPE ProLiant Gen10 hot-plug servers with the entire traditional data center attributes – standard racks and cabling and rear-aisle serviceability access. A 42U rack fits up to 20 Apollo 2000 system chassis, accommodating up to 80 servers per rack. It delivers the flexibility to tailor the system to the precise needs of your workload with the right compute, flexible I/O, and storage options. The servers can be “mixed and matched” within a single chassis to support different applications, and it can even be deployed with a single server, leaving room to scale as customer’s needs grow.
PRIMERGY RX2530 M5 – by Fujitsu
It is a dual-socket rack server that provides high performance of the new Intel Xeon Processor Scalable Family CPUs, expandability of up to 3TB of DDR4 memory and the capability to use Intel Optane DC Persistent Memory, and up to 10×2.5-inch storage devices – all in a 1U space saving housing. The system can also be equipped with the new 2nd generation processors of the Intel Xeon Scalable Family (CLX-R) delivering industry-leading frequencies. Accordingly, the PRIMERGY RX2530 M5 is the optimal system for large virtualization and scale-out scenarios, databases and for high-performance computing.
PowerSwitch Z9332F-ON – by Dell EMC
The Z9332F-ON 100/400GbE fixed switch comprises Dell EMC’s latest disaggregated hardware and software data center networking solutions, providing high-density 100/400GbE ports and range of functionality to meet the growing demands of today’s data center environment. These innovative, next-generation open networking high-density aggregation switches offer optimum flexibility and cost effectiveness for the web 2.0, enterprise, mid-market and cloud service provider with demanding compute and storage traffic environments. The compact unit provides industry-leading density of either 32 ports of 400GbE in QSFP56-DD form factor or 128 ports of 100 or up to 144 ports of 10/25/50GbE (via breakout), in a 1RU design.
E5700 – by NetApp
These hybrid-flash storage systems deliver high IO/s with low latency and high bandwidth for your mixed workload apps. Requiring 2U of rack space, the E5700 hybrid array combines extreme IO/s, sub-100 microsecond response times, and up to 21GB/s of read bandwidth and 14GB/s of write bandwidth. Wit redundant I/O paths, advanced data protection features, and extensive diagnostic capabilities, the E5700 storage systems enable you to achieve greater than 99.9999% availability and provide data integrity and security.
ScaTeFS – by NEC Corporation
The NEC Scalable Technology File System (ScaTeFS) is a distributed and parallel file system designed for large-scale HPC systems requiring large capacity. To realize load balancing and scale-out, all typical basic functions of a file system (read/write operation, file/directory generation, etc.) are distributed to multiple IO servers uniformly since ScaTeFS does not need a master server for managing the entire file system such as a metadata server. Therefore, the throughput of the entire system increases, and parallel I/O processing can be used for large files.
HPC-X ScalableHPC – by Mellanox
This toolkit is software package that includes MPI and SHMEM/PGAS communications libraries. It also includes various acceleration packages to improve both the performance and scalability of high performance computing applications running on top of these libraries, including UCX (Unified Communication X) which accelerates point-to-point operations, and FCA (Fabric Collectives Accelerations) which accelerates collective operations used by the MPI/PGAS languages. This full-featured, tested and packaged toolkit enables MPI and SHMEM/PGAS programming languages to achieve high performance, scalability and efficiency, and to assure that the communication libraries are optimized of the Mellanox interconnect solutions.
Panasas ActiveStor-18 – by Panasas
Panasas is the performance leader in hybrid scale-out NAS for unstructured data, driving industry and research innovation by accelerating workflows and simplifying data management. ActiveStor appliances leverage the patented PanFS storage operating system and DirectFlow protocol to deliver high performance and reliability at scale from an appliance that is as easy to manage as it is fast to deploy. With flash technology speeding small file and metadata performance, ActiveStor provide improved file system responsiveness while accelerating time-to-results. Based on a fifth-generation storage blade architecture and the proven PanFS storage operating system, ActiveStor offers an attractive low TCO for the energy, government, life sciences, manufacturing, media, and university research markets.
Future Ahead
Dataset is growing enormously. And, there will be no end to it. HPC storage must be able to process data at the speed of the light to maintain compute efficiency at peak levels. HPC storage should climb to exascale from petascale. It must have robust in-built security, be fault-tolerant, be modular in design and most importantly, scale seamlessly. HPC storage based on hybrid cloud technology is a sensible path ahead; however, the efforts must be geared to control its components at runtime. Further, focus should also be on dynamic marshaling via the applet provisioning and in-built automation engine. This will improve compute performance and reduce costs.