SRCMap: Energy Proportional Storage using Dynamic Consolidation - - PowerPoint PPT Presentation

srcmap energy proportional storage using dynamic
SMART_READER_LITE
LIVE PREVIEW

SRCMap: Energy Proportional Storage using Dynamic Consolidation - - PowerPoint PPT Presentation

Motivation Design Evaluation Conclusions & Future Work SRCMap: Energy Proportional Storage using Dynamic Consolidation Akshat Verma 1 Ricardo Koller 2 Luis Useche 2 Raju Rangaswami 2 1 IBM Research, India 2 School of Computing and


slide-1
SLIDE 1

Motivation Design Evaluation Conclusions & Future Work

SRCMap: Energy Proportional Storage using Dynamic Consolidation

Akshat Verma1 Ricardo Koller2 Luis Useche2 Raju Rangaswami2

1IBM Research, India 2School of Computing and Information Sciences

College of Engineering and Computing

FAST Conference, 2010

1 / 21

slide-2
SLIDE 2

Motivation Design Evaluation Conclusions & Future Work

◮ Current power density of data centers is 100 W/sq.ft & increasing 15-20% per year. ◮ Storage consume 10-25% of computing equipment. ◮ Storage load low (10-30%), but still peak power consumed. ◮ CPUs are more energy proportionality than storage. ◮ Consolidation is a well known technique for energy proportionality in virtualized servers.

2 / 21

slide-3
SLIDE 3

Motivation Design Evaluation Conclusions & Future Work

Storage Consolidation?

Can we use a storage virtualization layer to design a practical energy proportional storage system? ◮ Storage virtualization I/O indirection useful for consolidation. Challenge Moving logical volumes from one device to another is prohibitively expensive.

3 / 21

slide-4
SLIDE 4

Motivation Design Evaluation Conclusions & Future Work

Background: Storage Virtualization

4 / 21

slide-5
SLIDE 5

Motivation Design Evaluation Conclusions & Future Work

Outline

  • 1. Motivation
  • 2. Design
  • 3. Evaluation
  • 4. Conclusions & Future Work

5 / 21

slide-6
SLIDE 6

Motivation Design Evaluation Conclusions & Future Work

Workloads

mail Our department mail server. web-vm Virtual machine hosting two web-servers: CS web-mail & online course management. homes NFS server that serves the home directories for our research group. Block traces collected downstream of an active page cache for three weeks.

6 / 21

slide-7
SLIDE 7

Motivation Design Evaluation Conclusions & Future Work

Observations

Observation 1 The active data set is only a small fraction of total storage used. (about 1.5-6.5%) Observation 2 There is a significant variability in I/O load. (5-6 orders of magnitude) Observation 3 More that 99% of the working set consist of really popular & recently accessed data.

7 / 21

slide-8
SLIDE 8

Motivation Design Evaluation Conclusions & Future Work

Overview

8 / 21

slide-9
SLIDE 9

Motivation Design Evaluation Conclusions & Future Work

Our Approach

Sample Characterize the logical volume to find the working set. Replicate Create multiple working-set replicas in various physical volumes’ scratch space. Consolidate Based on I/O workload intensity, activate a sub-set

  • f physical volumes and serve workloads either from
  • riginal copies or working set replicas on these active

disks.

Initialization Every H hours

9 / 21

slide-10
SLIDE 10

Motivation Design Evaluation Conclusions & Future Work

Goals → Solutions

Goal Solution Fine grained proportionality Multiple replica targets. Low space overhead Instead of entire volumes, only working-sets are replicated. Reliability Coarse-grained consolidation

  • intervals. (hours)

Workload Adaptation Update working set replicas with new data that lead to read misses. Heterogeneity support Performance-power ratio ac- counted for in the replica place- ment benefit function.

10 / 21

slide-11
SLIDE 11

Motivation Design Evaluation Conclusions & Future Work

SRCMap work-flow

Event Response Initialization Detect working-sets of logical volumes & create replicas. Every H hours Identify what volumes and replicas to activate the next H hours. Change in workload Same as initialization.

11 / 21

slide-12
SLIDE 12

Motivation Design Evaluation Conclusions & Future Work

Replica Placement

◮ Replication benefit based on:

  • 1. Working set stability
  • 2. Average load
  • 3. Power efficiency of primary physical volume.
  • 4. Working set size

◮ Assign space with priorities based on benefit. ◮ Update replica creation benefit as additional replicas are created. ◮ Algorithm executes until scratch spaces are full.

12 / 21

slide-13
SLIDE 13

Motivation Design Evaluation Conclusions & Future Work

Active Replica Identification

◮ Calculate predicted aggregate workload IOPS. ◮ Compute minimum number of volumes to serve the aggregate IOPS. ◮ Identify replicas for inactive volumes. ◮ The number of active disks is incremented by one in case no active replica has been identified for some inactive volume.

13 / 21

slide-14
SLIDE 14

Motivation Design Evaluation Conclusions & Future Work

Workloads & Configuration

◮ 8 workloads to independent data volumes. ◮ Mix of web-servers of our CS department, and file server, SVN, and WiKi for our research group. ◮ H = 2. Change active replicas every 2 hours. ◮ Two minute disk time-outs. ◮ Working sets & replicas based on three week workload history. ◮ We report results of replaying the next 8 most active hours in the traces. ◮ We assume an oracle for estimation of load during each consolidation interval.

14 / 21

slide-15
SLIDE 15

Motivation Design Evaluation Conclusions & Future Work

Storage test-bed

◮ One machine with 8 SATA ports. ◮ Intel P4 HT 3GHz, 1GB memory. ◮ Trace played back using btreplay. ◮ Dedicated power supply for disks connected to power meter. ◮ Watts up? PRO power meter: measures power every second with resolution of 0.1W.

15 / 21

slide-16
SLIDE 16

Motivation Design Evaluation Conclusions & Future Work

Power

20 30 40 50 60 70 Watts Baseline - On SRCMap 2 4 6 1 2 3 4 5 6 7 8 # Disks On Hour

◮ Power consumption measured every second & active disks every 5 seconds. Disks off Power Saved 4.33 23.5 (35.5%)

16 / 21

slide-17
SLIDE 17

Motivation Design Evaluation Conclusions & Future Work

Response time

0.85 0.9 0.95 1 0.2 0.4 0.6 0.8 10-1 100 101 102 103 104 P(Response Time < x) Response Time (msec) SRCMap Baseline - On 0.85 0.9 0.95 1 0.2 0.4 0.6 0.8 P(Response Time < x)

◮ After 1ms, Baseline - On demonstrate better performance. ◮ 8% of requests with latencies ≥ 10ms. ◮ 2% of requests with latencies ≥ 100ms. ◮ Synchronization I/Os issued at beginning of each interval. ◮ Replaying without sync I/Os follows Baseline-On more closely.

17 / 21

slide-18
SLIDE 18

Motivation Design Evaluation Conclusions & Future Work

Energy proportionality

25 30 35 40 45 50 55 60 10 20 30 40 50 60 70 80 90 Power (Watts) Load factor (%) 25.65 + 0.393*x

◮ One point for each 2-hour interval in 24-hour duration. ◮ Load Factor: Load relative to the assumed volume maximum load capacity. SRCMap is able to achieve close to N-level proportionality for a system with N physical volumes.

18 / 21

slide-19
SLIDE 19

Motivation Design Evaluation Conclusions & Future Work

Conclusions ◮ We proposed and evaluate SRCMap, a storage virtualization solution for energy proportional storage. ◮ SRCMap establishes the feasibility of energy proportional storage systems. ◮ SRCMap meets all goals we set out to achieve energy proportional storage:

Low space overhead Reliability Workload adaptation Heterogeneity support Fine grain energy proportionality

19 / 21

slide-20
SLIDE 20

Motivation Design Evaluation Conclusions & Future Work

Future Work ◮ Models to predict I/O workload intensity. ◮ Models that estimate the performance impact of storage consolidation. ◮ Investigate the presence of workload correlation for better workload estimation and consolidation decision. ◮ Optimizing the scheduling of synchronization I/Os to minimize impact on foreground requests.

20 / 21

slide-21
SLIDE 21

Motivation Design Evaluation Conclusions & Future Work

http://dsrl.cs.fiu.edu/projects/srcmap/

Questions?

21 / 21

slide-22
SLIDE 22

Motivation Design Evaluation Conclusions & Future Work

Related Work

◮ Singly redundant schemes: Spin down volumes with redundant data during low load. ◮ Geared RAIDs: Redundancy on several disks and each disk spun down represents a gear shift. ◮ Caching systems: Cache of popular data on additional storage. ◮ Write Offloading: Increase disk idle periods by redirecting writes to alternate locations.

slide-23
SLIDE 23

Motivation Design Evaluation Conclusions & Future Work

Other Methods

90 2 4 6 8 10 12 14 16 18 20 22 24 Load Hour 2 Remaps 30 60 90 Power (Watts) SRCMap(L0) Replication Caching-1 Caching-2