Power Efficiency Analysis of a Deep Learning Workload on an IBM "Minsky" Platform

TitlePower Efficiency Analysis of a Deep Learning Workload on an IBM "Minsky" Platform
Publication TypeBook Chapter
Year of Publication2018
AuthorsPérez, MDMazueco, Seiler, NG, Bederián, CSergio, Wolovick, N, Vega, AJ
EditorMeneses, E, Castro, H, Hernández, CJaime Barr, Ramos-Pollán, R
Book TitleHigh Performance Computing - 5th Latin American Conference, CARLA 2018, Bucaramanga, Colombia, September 26-28, 2018, Revised Selected Papers
Series TitleCommunications in Computer and Information Science
Volume979
Pagination255–262
PublisherSpringer
AbstractThe rise of Deep Learning techniques has attracted special attention to GPUs usage for better performance of model computation. Most frameworks for Cognitive Computing include support to offload model training and inferencing to graphics hardware, and this is so common that GPU designers are reserving die area for special function units tailored to accelerating Deep Learning computation. Measuring the capability of a hardware platform to run these workloads is a major concern for vendors and consumers of this exponentially growing market. In a previous work [9] we analyzed the execution times of the Fathom AI workloads [2] in CPUs and CPUs+GPUs. In this work we measure the Fathom workloads in the POWER8-based “Minsky” [15] platform, profiling power consumption and energy efficiency in GPUs. We explore alternative forms of execution via GPU power and frequency capping with the aim of reducing Energy-to-Solution (ETS) and Energy-Delay-Product (EDP). We show important ETS savings of up to 27% with half of the workloads decreasing the EDP. We also expose the advantages of frequency capping with respect to power capping in NVIDIA GPUs.
URLhttps://doi.org/10.1007/978-3-030-16205-4_19
DOI10.1007/978-3-030-16205-4_19
PDF (Full text):