ATLAS ROOT I/O pt 2 Atlas Hot Topics (with reference to CHEP - - PowerPoint PPT Presentation

atlas root i o pt 2
SMART_READER_LITE
LIVE PREVIEW

ATLAS ROOT I/O pt 2 Atlas Hot Topics (with reference to CHEP - - PowerPoint PPT Presentation

ATLAS ROOT I/O pt 2 Atlas Hot Topics (with reference to CHEP presentations) Big data interlude (not ATLAS ) ROOT I/O Monitoring and Testing on ATLAS Atlas feature requests /fixes Wahid Bhimji 6th Dec 2013 ROOT IO Workshop Hot


slide-1
SLIDE 1

6th Dec 2013 ROOT IO Workshop

ATLAS ROOT I/O pt 2

Wahid Bhimji

  • Atlas Hot Topics (with reference to CHEP presentations)
  • Big data interlude (not ATLAS )
  • ROOT I/O Monitoring and Testing on ATLAS
  • Atlas feature requests /fixes
slide-2
SLIDE 2

Hot topics

slide-3
SLIDE 3

Old (current) ATLAS data flow

RAW ESD AOD dESD dAOD D3PD Reco Reduce Analysis Analysis Bytestream Not ROOT AOD/ESD ROOT with POOL Persistency D3PD ROOT ntuples with

  • nly primitive types

and vectors of those Athena Software Framework Non-Athena User Code (standard tools and examples)

3

Some Simplification! U s e r N t u p l e s TAG Analysis

ATLAS ROOT-based data formats - CHEP 2012

See my talk - at last CHEP

slide-4
SLIDE 4

A problem with this is

✤ Much heavy-IO activity is not centrally organised ✤ Run using users own pure-ROOT code - up to them to optimise

By no. of jobs, Analysis =55% By wallclock time, Analysis=22%

4

ATLAS ROOT-based data formats - CHEP 2012

slide-5
SLIDE 5

New (future) ATLAS Model

✤ xAOD easier to read in pure ROOT than current AOD. ✤ Reduction framework centrally controlled and does heavy lifting ✤ Common analysis software

, ‘ xAOD

see Paul Laycock’s CHEP talk

slide-6
SLIDE 6

New (future) ATLAS Model

✤ Of interest to this group ✤ New data structure: the xAOD ✤ Opportunities for IO optimisations that don’t have to be in each

users code: reduction framework and common analysis tools

✤ A step on the way: NTUP_COMMON ✤ Previously many physics groups have their own (large) D3PDs

  • verlapping in content - using more space than need be

✤ So new common D3PD solves that issue. But has a huge number

(10k+) of branches. Not an optimal solution - will go to xAOD.

slide-7
SLIDE 7

ATLAS Xrootd Federation (FAX)

see Ilija VUKOTIC’s CHEP talk

Aggregating storage in global namespace with transparent failover Of interest to this group:

✤ WAN reading requires decent I/O..

And is working out OK for us (though not exposed to random users yet (except as a fallback))

slide-8
SLIDE 8

Atlas Rucio

✤ Redesign of data management system

Popularity, Accounting, Metrics, Measures, Reports Account, Scope, Data identifier, Namespace, Meta- data, Replica registry, Subscription, Rules, Locks, Quota, Accounting Authentication & Authorization File conveyor, Reaper Rucio Storage Element(RSE), FTS3, Networking Middleware Core Analytics Daemons CLIs, APIs, Python Clients Clients Rucio

Open and standard technologies:

  • RESTful APIs

(https+json)

  • http caching
  • WSGI server
  • Token-based

authentication (X509, GSS)

  • Open source data

access protocols

6

Software Stack

  • Better management of users, physics groups, ATLAS

activities, data ownership, permission, quota, etc.

  • Data hierarchy with metadata support

Files are grouped into datasets Datasets/Containers are grouped

in containers

  • Concepts covering changes in middleware

Federations Cloud storage Move towards open and widely adopted protocols

4

user.jdoe:AllPeriods user.jdoe:RunPeriodA user.jdoe:Run1 user.jdoe:Run2 user.jdoe: File_0001 user.jdoe: File_0250 user.jdoe: File_0751 ... ... ...

Concepts - Highlights

see Vincent GARONNE’s CHEP talk

Of interest to this group:

  • Supports container file (metalink)

for multiple sources / failover.

  • Possible http based “federation”

see Mario Lassing’s CHEP poster and lightening talk

slide-9
SLIDE 9

Big data interlude (not ATLAS)

slide-10
SLIDE 10

Chep theme was “BiG data ...”

✤ “Big data” in industry means Hadoop and its successors ✤ Number of CHEP presentations for physics event I/O (ie as well as the

log mining / metadata use cases that came before))- e.g.:

✤ EBKE and Waller: Drillbit column store ✤ My “Hepdoop”poster; Maaike Limper’s poster and lightning talk ✤ Many on using Hadoop processing with ROOT files ✤ Most don’t see explicit performance gains from using Hadoop ✤ My impression: scheduling tools and HDFS filessytem very mature

data structures less so (for our needs) but much of interest from Dremel

slide-11
SLIDE 11

Big data opportunities

✤ Opportunities to benefit from growth of “big data” ✤ “Impact” of our work ✤ Sharing technologies / ideas. Gaining something from the others ✤ As Fons said parts of Dremel “sound pretty much like ROOT” but

that should make us sad as well as proud.

✤ It would be great if we make ROOT usable / used by these

communities and their products useable by us: some areas are Friend trees, ROOT modular “distribution”; chunkable ROOT files etc.

✤ I realize this requires manpower but surprising if the hype can’t get

us some money for transferring LHC big data expertise to industry

slide-12
SLIDE 12

ATLAS Monitoring and Testing

slide-13
SLIDE 13

What testing /monitoring we have

✤ Still have Hammercloud ROOT I/O

tests (see previous meeting). Now testing also FAX /WAN

✤ ad hoc Hammercloud stress tests with

real analysis codes (also for FAX)

✤ But also now have server-side

detailed xrootd records

  • for federation traffic
  • Also local traffic for sites

that use xrootd

  • Regular monitoring but

also can be mined

Hammercloud Oracle Db SVN Define Code; Release; Dataset;… Uploads stats Regularly submitting single tests Sites Data mining tools Command line, Web interface, ROOT scripts ROOT source (via curl) dataset

xRootD&

Castor& Collector&at& SLAC&

xRootD&

DPM& Collector&at& CERN& UDP&

Monitor&

xroot4j& dCache&

xRootD&

posix&

xRootD&

MonaLisa& Ac=veMQ&

Consumers:&

  • Popularity&DB&
  • SSB&
  • Dashboard&

Monitor&

xroot4j1& dCache& Summary& stream& Detailed&stream&

slide-14
SLIDE 14

From mining of xrootd records from Edinburgh site - all jobs

Jobs with up to a million read operations

  • Most jobs have 0 vector operations (so not using TTC)
  • Some jobs do use it (but these mostly tests (shaded blue))

ECDF

We need to be able to switch on TTreeCache for our users!

slide-15
SLIDE 15

dCache localcopy Lustre DPM local copy xrootd/EOS GPFS dCap Direct

Cpu Eff. 0.5 0.55 0.6 0.65 0.7 0.75 0.8 0.85 0.9 0.95 1

Cpu Eff. 100% Events read

! TTreeCache

essential at some sites

! Users still

don’t set it

! Different

  • ptimal

values per site

! Ability to set

in job environment would be useful

20

300 MB TTC No TTC

ATLAS ROOT-based data formats - CHEP 2012

Old slide from CHEP 2012 to remind of the impact of TTC

slide-16
SLIDE 16

Oct HC FAX stress tests - US cloud

91 100 166 6 4 198 99 72 1 1 5 284 4 5 118 37 49 246 54 87

MWT2 AGLT2 SLAC BNL

Johannes Elmsheuser, Friedrich H¨

  • nig (LMU M¨

unchen) HammerCloud status and FAX tests 23/10/2013

Width and numbers: Event rate Green = 100% success Remote reading working well in these tests (which do use TTreeCache !)

See Johannes Elmsheuser et al.’s CHEP poster

slide-17
SLIDE 17

ROOT IO Feature Requests

✤ TTreeCache switch on/ configure in environ.: (including in ROOT 5.) ✤ This is useful even if TTC on by default or in new framework. ✤ Choices (multiple trees etc.) - but are there blockers? ✤ Support for new analysis model: generally for xAOD as it develops ✤ Specific Reflex feature “rules for the production of dictionaries for

template types with markup” in ROOT 6 (already on todo list I hear)

✤ Advice on handling Trees with 5000+ branches. ✤ Planned http access would benefit from TDavixFile: is it in now?