Analyzing the Energy-Time Trade-OFF in high performance computing applications

Presentation done by Kevin Gee
based on Freeh, Lowenthal, Pan, Kappiah, Springer paper 
published in IEEE Transactions on parallel and distributed systems, 
Vol. 18, No. 6, June 2007 

Purpose


  • Demonstrate results of energy-time trade-off of a wide range of applications - serial and parallel - on a power scalable cluster
  • Introduce metrics to predict memory and communication bottlenecks






2

Content


  • Introduction
  • Experimental methodology
  • Single node results
  • Multi node results
  • Metrics
  • Conclusion





3

Introduction: facts



  • HPC  (High Performance Computing) tends to push performance at all cost
  • Last 10% increase in performance requires disproportionally large amounts of resources
  • The Earth Simulator consumes 7 MW of power





4

INTRODUCTION: Solution



  • Low-power high-performance clusters to save energy
  • Green Destiny consumes about 3 times less energy than Accelerated Strategic Computing Initiative (ASCI) Q machine
  • ASCI_Q machine is about 15 times faster per node





5

INTRODUCTION: SOLUTION


  • Using high-performance commodity microprocessor with frequency and voltage scaling
  • Increase in frequency results in:
    • a smaller increase in application performance
    • increase in CPU stalls






6

Introduction: prediction



It is possible to reduce power consumption without a significant increase in execution time








7

Methodology: Programs


  • Single node benchmark sets
    • Nasa Advanced Supercomputing (NAS)
    • System Performance Evaluation Cooperative (SPEC) integer
    • SPEC float
  • Multiple node benchmark sets
    • NAS MPI
    • ASCI



8

METHODOLOGY: Cluster


10 nodes, connected by 100 Mbps network,
 each has:

  • AMD Athlon 64 CPU
  • 1 Gb main memory
  • 128 Kb L1 cache
  • 512 Kb L2 cache




9

METHODOLOGY: Gears











Idle and active power for each AMD-64 node


10

Single node results


Overall results







                      NAS                                        SPEC INT                                   SPEC FP



11

SINGLE NODE RESULTS


Best energy-time trade-off in each set with different programs





12

SINGLE NODE RESULTS


Worst energy-time trade-off using different programs





13

Multiple node results


Bottlenecks:

  • Memory
  • Communication







14

Multiple node results


Possible layouts of the graph :

  • Poor speedup - more energy, less time
  • Perfect or superlinear speedup
  • Speedup is good and there are significant memory operations





15

MULTIPLE NODE RESULTS


Poor speedup (more energy, less time):


16

MULTIPLE NODE RESULTS


Perfect speedup (no energy-time trade-off):



17

MULTIPLE NODE RESULTS


Good speedup:



18

Metrics: CPU criticality



Tmax - execution time in maximum frequency
T - execution time in current frequency
fmax - maximum frequency
f - current frequency

if (beta == 0) => application is independent from CPU 
(T == Tmax)
19

METrics: MPO


  • MPO - Misses Per Operation
  • Measures memory pressure
  • MPO = The number of operations retired and the L2 cache misses( which are memory accesses)
  • Constant while changing frequency
  • If MPO decreases - application becomes more CPU bound





20

Metrics: slack


  • Predicts communication bottleneck
  • Slack - ratio of the total time a node is blocked in MPI calls to the total execution time
  • Greater slack suggests reducing gear
  • Collected on the fly






21

Conclusion


  • In the best case on one node, reducing the CPU speed makes it possible to use 20 percent less energy while increasing time by only 3 percent.

  • On the other hand, a program that is largely CPU bound should be run at the fastest gear possible.

  • MPO and slack can be used to to choose appropriate gear.



22

Your Questions







23

Energy-Time Trade-OFF in high performance computing applications

By Žilvinas Kučinskas

Energy-Time Trade-OFF in high performance computing applications

  • 313