volley: automated data placement
for geo-distributed cloud services
sharad agarwal, john dunagan, navendu jain, stefan saroiu, alec wolman, harbinder bhogan
very rapid pace of datacenter rollout April 2007 Microsoft opens DC - - PowerPoint PPT Presentation
volley: automated data placement for geo-distributed cloud services sharad agarwal, john dunagan, navendu jain, stefan saroiu, alec wolman, harbinder bhogan very rapid pace of datacenter rollout April 2007 Microsoft opens DC in Quincy, WA
sharad agarwal, john dunagan, navendu jain, stefan saroiu, alec wolman, harbinder bhogan
sharad.agarwal@microsoft.com PAGE 2 4/29/2010
sharad.agarwal@microsoft.com PAGE 3 4/29/2010
and cost is a major concern
sharad.agarwal@microsoft.com PAGE 4 4/29/2010
sharad.agarwal@microsoft.com PAGE 6 4/29/2010
john john’s wall john’s news feed sharad sharad’s wall sharad’s news feed
sharad.agarwal@microsoft.com PAGE 7 4/29/2010
PLACING ALL DATA ITEMS IN ONE PLACE IS REALLY BAD FOR LATENCY
sharad.agarwal@microsoft.com PAGE 8 4/29/2010
ALGORITHM NEEDS TO HANDLE USER LOCATIONS THAT CAN VARY
20 40 60 80 100 1 2 3 4 5 6 7 8 9 10 % of devices or users max distance from centroid (x1000 miles) % of Mesh devices % of Messenger users
sharad.agarwal@microsoft.com PAGE 9 4/29/2010
ALGORITHM NEEDS TO HANDLE DATA ITEMS THAT ARE ACCESSED AT SAME TIME BY USERS IN DIFFERENT LOCATIONS
20 40 60 80 100 1 2 3 4 5 6 7 8 9 10 % of instances distance from device to sharing centroid (x1000 miles) % of Messenger conversations % of Mesh notification sessions
sharad.agarwal@microsoft.com PAGE 10 4/29/2010
john john’s wall john’s news feed sharad sharad’s wall sharad’s news feed
sharad.agarwal@microsoft.com PAGE 12 4/29/2010
DC Z DC Y DC X
user updates wall A with two subscribers C,D
user updates wall A with one subscriber C
user updates wall B with one subscriber D
IP 2 data B data C IP 1 data A
2 1 2 1
data D
1
frequency of operations can be weighted by importance
sharad.agarwal@microsoft.com PAGE 13 4/29/2010
sharad.agarwal@microsoft.com PAGE 14 4/29/2010
sharad.agarwal@microsoft.com PAGE 15 4/29/2010
sharad.agarwal@microsoft.com PAGE 16 4/29/2010
app-specific migration mechanism
app servers in DC n Cosmos store in DC y Volley analysis job app-specific migration mechanism app servers in DC 2 app servers in DC 1
sharad.agarwal@microsoft.com PAGE 18 4/29/2010
sharad.agarwal@microsoft.com PAGE 19 4/29/2010
sharad.agarwal@microsoft.com PAGE 20 4/29/2010
INCLUDES SERVER-SERVER (SAME DC OR CROSS-DC) AND SERVER-USER
50 100 150 200 250 300 350 400 450 50th 75th 95th user transaction latency (ms) percentile of total user transactions hash
commonIP volley
sharad.agarwal@microsoft.com PAGE 21 4/29/2010
WAN TRAFFIC IS A MAJOR SOURCE OF COST FOR OPERATORS
0.0000 0.7929 0.2059 0.1109 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
hash commonIP volley fraction of messages that are inter-DC placement real money
sharad.agarwal@microsoft.com PAGE 22 4/29/2010
COMPARED TO FIRST WEEK
0% 20% 40% 60% 80% 100% week2 week3 week4 percentage of objects
different placement
same placement new objects
sharad.agarwal@microsoft.com PAGE 23 4/29/2010
sharad agarwal john dunagan navendu jain stefan saroiu alec wolman harbinder bhogan