Power Efficiency Analysis of a Deep Learning Workload on an IBM "Minsky" Platform
Title | Power Efficiency Analysis of a Deep Learning Workload on an IBM "Minsky" Platform |
Publication Type | Book Chapter |
Year of Publication | 2018 |
Authors | Pérez, MDMazueco, Seiler, NG, Bederián, CSergio, Wolovick, N, Vega, AJ |
Editor | Meneses, E, Castro, H, Hernández, CJaime Barr, Ramos-Pollán, R |
Book Title | High Performance Computing - 5th Latin American Conference, CARLA 2018, Bucaramanga, Colombia, September 26-28, 2018, Revised Selected Papers |
Series Title | Communications in Computer and Information Science |
Volume | 979 |
Pagination | 255–262 |
Publisher | Springer |
Abstract | The rise of Deep Learning techniques has attracted special attention to GPUs usage for better performance of model computation. Most frameworks for Cognitive Computing include support to offload model training and inferencing to graphics hardware, and this is so common that GPU designers are reserving die area for special function units tailored to accelerating Deep Learning computation. Measuring the capability of a hardware platform to run these workloads is a major concern for vendors and consumers of this exponentially growing market. In a previous work [9] we analyzed the execution times of the Fathom AI workloads [2] in CPUs and CPUs+GPUs. In this work we measure the Fathom workloads in the POWER8-based “Minsky” [15] platform, profiling power consumption and energy efficiency in GPUs. We explore alternative forms of execution via GPU power and frequency capping with the aim of reducing Energy-to-Solution (ETS) and Energy-Delay-Product (EDP). We show important ETS savings of up to 27% with half of the workloads decreasing the EDP. We also expose the advantages of frequency capping with respect to power capping in NVIDIA GPUs. |
URL | https://doi.org/10.1007/978-3-030-16205-4_19 |
DOI | 10.1007/978-3-030-16205-4_19 |
PDF (Full text):