File System Trace Replay Methods Through the Lens of Metrology - - PowerPoint PPT Presentation

file system trace replay methods through the lens of
SMART_READER_LITE
LIVE PREVIEW

File System Trace Replay Methods Through the Lens of Metrology - - PowerPoint PPT Presentation

File System Trace Replay Methods Through the Lens of Metrology Thiago Emmanuel Pereira, Francisco Vilar Brasileiro, Lvia Sampaio temmanuel@copin.ufcg.edu.br, fubica@computacao.ufcg.edu.br, livia@computacao.ufcg.edu.br Federal University of


slide-1
SLIDE 1

File System Trace Replay Methods Through the Lens

  • f Metrology

Thiago Emmanuel Pereira, Francisco Vilar Brasileiro, Lívia Sampaio

temmanuel@copin.ufcg.edu.br, fubica@computacao.ufcg.edu.br, livia@computacao.ufcg.edu.br

Federal University of Campina Grande, Brasil

slide-2
SLIDE 2

Very popular but ...

  • Many ad hoc trace replay tools – no description about their design and

implementation

  • Impossible to reproduce results.

"How to do this accurately is still an open question, and the best we can do right now is take results with a degree of skepticism” - Traeger, A.,

Zadok, E., Joukov, N., & Wright, C. P. (2008). A nine year study of file system and storage

  • benchmarking. ACM Transactions on Storage (TOS)
slide-3
SLIDE 3

Before creating new methods, how good are current trace based methods?

slide-4
SLIDE 4

Our take A metrology case study

slide-5
SLIDE 5

Our take A metrology case study

Single-laboratory

slide-6
SLIDE 6

Our take A metrology case study

Inter-laboratories Single-laboratory Different operators Different instruments Different environment

slide-7
SLIDE 7

Our take A metrology case study

Inter-laboratories Single-laboratory Different operators Different instruments Different environment

slide-8
SLIDE 8

Single-lab testing

  • 1. Defjne the measurand
  • The quantity intended to be measured
  • 2. Specify the measurement procedure
  • 3. Identify the uncertainty sources
  • 4. Conduct the measurement characterization
  • In terms of bias, precision, sensitivity, resolution, etc.
  • 5. Perform the calibration (or mitigation of measurement

errors)

  • 6. Calculate the measurement uncertainty
  • An interval [y – u, y + u] within the true value of

measurand y are expected to be.

slide-9
SLIDE 9

Measurand

File system response time

slide-10
SLIDE 10

Measurement procedure

Instruments ARTC replayer (compilation-based) TBBT replayer (event-based)

1.Weiss, Zev, et al. "Root: Replaying multithreaded traces with resource-

  • riented ordering." SOSP

. ACM, 2013. 2.Zhu, Ningning, Jiawu Chen, and Tzi-Cker Chiueh. "TBBT: scalable and accurate trace replay for fjle server evaluation." FAST,2005.

slide-11
SLIDE 11

ARTC Replayer

1.Weiss, Zev, et al. "Root: Replaying multithreaded traces with resource-

  • riented ordering." SOSP

. ACM, 2013.

ARTC compiler C code C compiler ARTC replayer shared library trace

slide-12
SLIDE 12

TBBT Replayer

1.Zhu, Ningning, Jiawu Chen, and Tzi-Cker Chiueh. "TBBT: scalable and accurate trace replay for fjle server evaluation." FAST,2005. 2.T arihi, Mojtaba, Hossein Asadi, and Hamid Sarbazi-Azad. "DiskAccel: Accelerating Disk-Based Experiments by Representative Sampling." SIGMETRICS , 2015.

formatter TBBT Based on TBBT design, running as a real time process to be less sensitive trace trace

slide-13
SLIDE 13

Uncertainty sources

slide-14
SLIDE 14

Characterization

ARTC Workload trace capture TBBT Measurement

slide-15
SLIDE 15

Characterization

ARTC Workload trace capture TBBT Measurement Reference values

slide-16
SLIDE 16

Characterization

  • Microbenchmark (5k ops, 4k chunks, [1-4] threads)
  • Random read (RR), Random write (RW)
  • Sequential read (SR), Sequential write (SW)
  • Filebench fjleserver workload

ARTC Workload trace capture TBBT Measurement Reference values

slide-17
SLIDE 17

Characterization

Microbenchmark

slide-18
SLIDE 18

Bias characterization

slide-19
SLIDE 19

Bias characterization

slide-20
SLIDE 20

Bias characterization

slide-21
SLIDE 21

TBBT improvements

slide-22
SLIDE 22

TBBT improvements

TBBT coordinator overhead TBBT coordinator overhead

slide-23
SLIDE 23

TBBT improvements

TBBT coordinator overhead TBBT coordinator overhead Real time scheduler

slide-24
SLIDE 24

Uncertainty

Workload TBBT ARTC Random read 22579.0 ± 2.4% (22891.6 ± 4.8%) 22243.5 ± 1.8% Random write 22946.1 ± 3.2% (24807.6 ± 18%) 23076.0 ± 4.1% Sequential read 4.0 ± 32.9% (7.8 ± 253%) 3.7 ± 18.6% Sequential write 105.6 ± 1.3% (107.7 ± 4.2%) 105.8 ± 0.6%

slide-25
SLIDE 25

Uncertainty

Workload TBBT ARTC Random read 22579.0 ± 2.4% (22891.6 ± 4.8%) 22243.5 ± 1.8% Random write 22946.1 ± 3.2% (24807.6 ± 18%) 23076.0 ± 4.1% Sequential read 4.0 ± 32.9% (7.8 ± 253%) 3.7 ± 18.6% Sequential write 105.6 ± 1.3% (107.7 ± 4.2%) 105.8 ± 0.6%

Before TBBT improvements ARTC is a clear winner

slide-26
SLIDE 26

Uncertainty

Workload TBBT ARTC Random read 22579.0 ± 2.4% (22891.6 ± 4.8%) 22243.5 ± 1.8% Random write 22946.1 ± 3.2% (24807.6 ± 18%) 23076.0 ± 4.1% Sequential read 4.0 ± 32.9% (7.8 ± 253%) 3.7 ± 18.6% Sequential write 105.6 ± 1.3% (107.7 ± 4.2%) 105.8 ± 0.6%

TBBT improvements are affective

slide-27
SLIDE 27

Uncertainty

Workload TBBT ARTC Random read 22579.0 ± 2.4% (22891.6 ± 4.8%) 22243.5 ± 1.8% Random write 22946.1 ± 3.2% (24807.6 ± 18%) 23076.0 ± 4.1% Sequential read 4.0 ± 32.9% (7.8 ± 253%) 3.7 ± 18.6% Sequential write 105.6 ± 1.3% (107.7 ± 4.2%) 105.8 ± 0.6%

How to choose between replayers?

slide-28
SLIDE 28

Characterization

Filebench fjleserver workload

slide-29
SLIDE 29

Characterization

Filebench fjleserver workload

  • 4 threads
  • creat, delete, append, read, write, stat
  • variable fjle sizes
  • Wholefjle read and write
slide-30
SLIDE 30

Uncertainty

TBBT ARTC Reference Read 20.73 ± 118.27% 27.21 ± 92.72% 50.72 Write 50.45 ± 79.81% 69.79 ± 33.79% 83.95

slide-31
SLIDE 31

Uncertainty

TBBT ARTC Reference Read 20.73 ± 118.27% 27.21 ± 92.72% 50.72 Write 50.45 ± 79.81% 69.79 ± 33.79% 83.95

Replayed response time appears better than reference TBBT and ARTC memory footprints are smaller than filebench footprint, thus more cache hits

slide-32
SLIDE 32

Uncertainty

TBBT ARTC Reference Read 20.73 ± 118.27% 27.21 ± 92.72% 50.72 Write 50.45 ± 79.81% 69.79 ± 33.79% 83.95

TBBT response time appears better than ARTC response time

slide-33
SLIDE 33

Uncertainty

Replayers are not able to match captured workload concurrency

slide-34
SLIDE 34

Conclusions

Metrology can help:

  • Choosing the best instrument for the job (based on the

measurement uncertainty)

  • The TBBT replayer, in some cases, is equivalent to

the ARTC replayer.

  • Improving tools and best practices
  • Event-based replayer needs improvement
  • Changes in OS scheduler policy may affect sensitive

metrics.

  • Spotting uncertainty sources
  • Differences in experimental environment, such as the

amount of available memory, are likely to hurt reproducibility.