ABSTRACT
The advent of fast, unprecedentedly scalable, yet energy-hungry exascale supercomputers poses a major challenge consisting in sustaining a high performance per watt ratio. While much recent work has explored new approaches to I/O management, aiming to reduce the I/O performance bottleneck exhibited by HPC applications (and hence to improve application performance), there is comparatively little work investigating the impact of I/O management approaches on energy consumption.
In this work, we explore how much energy a supercomputer consumes while running scientific simulations when adopting various I/O management approaches. We closely examine three radically different I/O schemes including time partitioning, dedicated cores, and dedicated nodes. We implement the three approaches within the Damaris I/O middleware and perform extensive experiments with one of the target HPC applications of the Blue Waters sustained-petaflop supercomputer project: the CM1 atmospheric model. Our experimental results obtained on the French Grid'5000 platform highlight the differences between these three approaches and illustrate in which way various configurations of the application and of the system can impact performance and energy consumption.
- James Hamilton, Cost of Power in Large-Scale Data Centers . http://perspectives.mvdirona.com/2008/11/28/ CostOfPowerInLargeScaleDataCenters.aspx, November2008.Google Scholar
- R. Bolze, F. Cappello, E. Caron, M. Daydé F. Desprez, E. Jeannot, Y. Jégou, S. Lanteri, J. Leduc, N. Melab, et al. Grid '5000: a large scale and highly reconfigurable experimental grid testbed. International Journal of High Performance Computing Applications, 20(4):481,2006. Google ScholarDigital Library
- G. H. Bryan and J. M. Fritsch. A benchmark simulation for moist nonhydrostatic numerical models. Monthly Weather Review,130(12):2917--2928,2002.Google ScholarCross Ref
- P. H. Carns, W. B. Ligon, III, R. B. Ross, and R. Thakur. PVFS: a parallel file system for linux clusters. In Proceedings of the 4th annual Linux Showcase & Conference - Volume 4, Berkeley, CA, USA, 2000. USENIX Association. Google ScholarDigital Library
- M. Dorier, G. Antoniu, F. Cappello, M. Snir, and L. Orf. Damaris: Leveraging Multicore Parallelism to Mask I/O Jitter. Research report RR-7706, INRIA, Dec 2011.Google Scholar
- M. Dorier, G. Antoniu, F. Cappello, M. Snir, and L. Orf. Damaris: How to Efficiently Leverage Multicore Parallelism to Achieve Scalable, Jitter-free I/O. In Proceedings of the 2012 IEEE International Conference on Cluster Computing, Cluster'14, pages 155--163, Sept. 2012. Google ScholarDigital Library
- M. Gamell, I. Rodero, M. Parashar, J. C. Bennett, H. Kolla, J. Chen, P.-T. Bremer, A. G. Landge, A. Gyulassy, P. McCormick, S. Pakin, V. Pascucci, and S. Klasky. Exploringpower behaviors and trade-offs of in-situ data analytics. In Proceedings of the International Conference on High Performance Computing, Networking, Storage and Analysis, SC '13, pages 77:1--77:12, New York, NY, USA, 2013. ACM. Google ScholarDigital Library
- A. Hoisie and V. Getov. Extreme-Scale Computing - Where 'Just More of the Same' Does Not Work. Computer, 42(11):24--26, Nov. 2009. Google ScholarDigital Library
- J. H. Laros, III, K. T. Pedretti, S. M. Kelly, W. Shu, and C. T. Vaughan. Energy based performance tuning for large scale high performance computing systems. In Proceedings of the 2012 Symposium on High Performance Computing, HPC '12, pages 6:1--6:10, San Diego, CA, USA, 2012. Society for Computer Simulation International. Google ScholarDigital Library
- J. Lofstead, F. Zheng, Q. Liu, S. Klasky, R. Oldfield, T. Kordenbrock, K. Schwan, and M. Wolf. Managing Variability in the IO Performance of Petascale Storage Systems. In Proceedings of the 2010 ACM/IEEE International Conference for High Performance Computing, Networking, Storage and Analysis, SC'10, pages1--12,Washington,DC, USA, 2010. IEEE Computer Society. Google ScholarDigital Library
- NCSA. BlueWaters project, http://www.ncsa.illinois.edu/BlueWaters/.Google Scholar
- C. Patel, R. Sharma, C. Bash, and S. Graupner. Energy aware grid: Global workload placement based on energy efficiency. HPL Technical Report, HPL-2002--329, Nov. 2002.Google Scholar
- D. Skinner and W. Kramer. Understanding the Causes of Performance Variability in HPC Workloads. In Proceedings of the IEEE International Workload Characterization Symposium, pages 137--149,Oct.2005.Google ScholarCross Ref
- F. Zheng, H. Abbasi, C. Docan, J. Lofstead, Q. Liu, S. Klasky, M. Parashar, N. Podhorszki, K. Schwan, and M. Wolf. PreDatA - Preparatory Data Analytics on Peta-Scale Machines. In Proceedings of the 2010 IEEE International Symposium on Parallel Distributed Processing, IPDPS'10, pages 1--12, April 2010.Google ScholarCross Ref
Index Terms
- A performance and energy analysis of I/O management approaches for exascale systems
Recommendations
On the energy footprint of I/O management in Exascale HPC systems
The advent of unprecedentedly scalable yet energy hungry Exascale supercomputers poses a major challenge in sustaining a high performance-per-watt ratio. With I/O management acquiring a crucial role in supporting scientific simulations, various I/O ...
Damaris: Addressing Performance Variability in Data Management for Post-Petascale Simulations
With exascale computing on the horizon, reducing performance variability in data management tasks (storage, visualization, analysis, etc.) is becoming a key challenge in sustaining high performance. This variability significantly impacts the overall ...
Challenges on the road to exascale computing
ICS '08: Proceedings of the 22nd annual international conference on SupercomputingSupercomputing systems have made great strides in recent years as the extensive computing needs of cutting-edge engineering work and scientific discovery have driven the development of more powerful systems. The first teraflop computer, ASCI Red, came ...
Comments