Using Pattern-Models to Guide SSD Deployment for Big Data in HPC Systems Junjie Chen 1, Philip C. Roth 2, Yong Chen 1 1 Data-Intensive Scalable Computing Lab (DISCL) Department of Computer Science Texas Tech University 2 Oak Ridge National Laboratory IEEE BigData’13 1 Background • HPC applications are increasingly data-intensive • Scientific simulations have already reached 100TB – 1PB of data volume, projected at the scale of 10PB – 100PB for upcoming exascale • Collected data from instruments increases rapidly too, in a global climate model, with 100 × 120 km grid cell, PBs of data managed • Such trend brings a critical challenge • Efficient I/O access demands • Highly efficient storage system IEEE BigData’13 2 Storage Media • Over 90% of all data in the world is being stored on magnetic media (hard disk drives, HDDs) • IBM invented in 1956 • Mechanism remains the same since then • Various mechanical moving parts • High latency, slow random access performance, unreliable, power hungry • Large capacity, low cost (USD 0.10/GB), impressive sequential access performance IEEE BigData’13 3 Emerging Storage Media • Non-volatile storage-class memory (SCM) • Flash-memory based Solid State Drives (SSDs), PCRAM, NRAM, … • Use microchips which retain data in nonvolatile memory (array of floating gate transistors isolated by an insulating layer) 1 1 1 1 Intel® X25-E SSD • Superior performance, high bandwidth, low latency (esp. random accesses), less susceptible to physical shock, power efficient • Low capacity, high cost (USD 0.90-2/GB), block erasure, wear out (10K-100K P/E cycles) 1 0 1 1 1 0 0 1 1 1 0 1 IEEE BigData’13 4 Motivation • The challenge of leveraging SSDs and maximizing benefits remains daunting. • Deploy SSDs on different nodes can have different impacts • The fixed hardware budget Interconnect needs to be considered. • A cost-effective decision of deployment needs to be made Local SSD storage at design/deployment phase of HPC systems. IEEE BigData’13 5 Local SSD storage Our Study • To investigate different deployment strategies • Compute side and storage side • Characteristics of SSDs, ratios, access patterns • Consider a fixed hardware budget • Pattern-Model Guided Deployment Approach • Considering I/O access pattern of workloads • Considering SSD characteristics via a performance model IEEE BigData’13 6 Our Contributions • We propose a pattern-model guided deployment approach • We introduce a performance model to quantitatively analyze different SSD deployment strategies • We try to answer the questions of how SSDs show be utilized for big data applications in HPC systems • We have carried out initial experimental tests to verify the proposed approach. IEEE BigData’13 7 Pattern-Model Guided Approach Workloads I/O Requests Parallel File System Operation Types Storage Configuration Workload Size Spatial Pattern Analytical Model Workload Characterization Storage Arrays Strategy Mapping Pattern-Model Guided Approach HDDs IEEE BigData’13 8 SSDs Workload I/O Access Pattern • Workload Characterization • Request size, I/O operation type, spatial pattern, and ratio of local requests to remote requests. • Given or obtained from I/O characterization tools, like Darshan and IOSIG. • Strategy Mapping • Analysisi -> Strategyj • For a specific pattern, give a specific deployment strategy. IEEE BigData’13 9 Performance Model • R is the total response time, Rlocal is the local response time, Rremote is the remote response time and Rinter is the time spent on interconnection. • W is the workload and B is the aggregate bandwidth. R= Rlocal + Rremote + Rint er W = B´ R® B = IEEE BigData’13 10 W R Performance Model (cont.) • We characterize the three different response time respectively and estimate the total response time (1 - g ) ´W ´ g ´W ´ L + + (1 - p) ´ w ´ Lssd ssd ´ Bint er ´ ´+[(1 - g ) ´W - (1 - p) ´ w ] ´ Lhdd , (g W ´ pw ) ´ R= ´ ´ W - p ´w ´p ´ w ´ Lssd + + (1- p) ´ w ´ Lssd Bint er ´ ´+(W - w ) ´ L , ( pw < g W) hdd ´ IEEE BigData’13 11 Lssd Latency of SSD Lhdd Latency of HDD γ Percentage of workload serviced locally ω The available capacity of SSDs p Percentage of SSD budget deployed on compute nodes Performance Model (cont.) • The tradeoff analysis, C is the capacity of SSD which one compute node could utilize, G is the SSD budget, and n is the number of compute nodes. p´ G C= + (1- p) ´ G n • Compute-side deployment: all the SSDs on compute nodes • Storage-side deployment: all the SSDs on storage nodes • Compute-Storage deployment: SSDs on both types of nodes IEEE BigData’13 12 Preliminary Results and Analysis • IOR • Tested the aggregate bandwidth/execution time • File size is varied, and the performance of sequential read and write, and random read and write is tested. • MPI-IO Test • Tested the aggregate bandwidth/execution time • With different file sizes and operation types (sequential read and write, random read and write). • Both benchmark with ratio γ= ¼. IEEE BigData’13 13 Preliminary Results and Analysis (cont.) IOR IEEE BigData’13 14 Preliminary Results and Analysis (cont.) MPI-IO Test IEEE BigData’13 15 Conclusion • Flash-memory based SSDs are promising storage devices in the storage hierarchy of HPC systems. • Different deployment strategies of SSDs can impact the performance given a fixed hardware budget • We proposed a pattern-model guided approach • Model the performance impact of various deployment strategies • Considering workload characterization and device characteristics • Mapping to deployment strategy • This study provides a possible solution that guides such placement and deployment strategies IEEE BigData’13 16 Ongoing and Future Work Workloads • A Unified HPC Storage System Managing Heterogeneous Devices Write Requests • We study the needs of a wellmanaged and unified heterogeneous storage system for HPC workloads Read Requests I/O History I/O History Similar to L H Read Region A Time Window • We propose a working-set based reorganization scheme (WS-ROS) Write Region Working-Set Model Storage management • Explore the capability of SSDs and HDDs • Provide a highly efficient storage system for HPC workloads Parallel File System Read region guidance Active data directed to Write region guidance IEEE BigData’13 17 Written data directed to HDDs SSDs Q&A Thank You Please visit our website: http://discl.cs.ttu.edu ACKNOWLDEGEMENT: This research is sponsored in part by the Advanced Scientific Computing Research program, Office of Science, U.S. Department of Energy. This research is also sponsored in part by Texas Tech University startup grant and National Science Foundation under NSF grant CNS-1162488. The work was performed in part at the Oak Ridge National Laboratory, which is managed by UT-Battelle, LLC under Contract No. De-AC05-00OR22725. Accordingly, the U.S. Government retains a non-exclusive, royalty-free license to publish or reproduce the published form of this contribution, or allow others to do so, for U.S. Government purposes. IEEE BigData’13 18
© Copyright 2026 Paperzz