On the mixing advantage Kais Hamza Monash University ANZAPW, July - - PowerPoint PPT Presentation

on the mixing advantage
SMART_READER_LITE
LIVE PREVIEW

On the mixing advantage Kais Hamza Monash University ANZAPW, July - - PowerPoint PPT Presentation

On the mixing advantage Kais Hamza Monash University ANZAPW, July 2013, University of Queensland Joint work with Aidan Sudbury, Peter Jagers & Daniel Tokarev Kais Hamza On the mixing advantage Introduction Kais Hamza On the mixing


slide-1
SLIDE 1

On the mixing advantage

Kais Hamza

Monash University

ANZAPW, July 2013, University of Queensland Joint work with Aidan Sudbury, Peter Jagers & Daniel Tokarev

Kais Hamza On the mixing advantage

slide-2
SLIDE 2

Introduction

Kais Hamza On the mixing advantage

slide-3
SLIDE 3

Introduction

◮ X j i , Xi are the lifetime of an individual/unit and,

max(X 1

i , . . . , X n i ) and max(X1, X2, . . . , Xn) represent the

lifetime of population/system. All random variables are assumed to be non-negative. Xi, X 1

i , . . . , X n i are identically distributed.

Kais Hamza On the mixing advantage

slide-4
SLIDE 4

Introduction

◮ X j i , Xi are the lifetime of an individual/unit and,

max(X 1

i , . . . , X n i ) and max(X1, X2, . . . , Xn) represent the

lifetime of population/system. All random variables are assumed to be non-negative. Xi, X 1

i , . . . , X n i are identically distributed. ◮ X1

X 1

1

X 2

1

. . . X n

1

→ M1 = E[max(X 1

1 , . . . , X n 1 )]

Kais Hamza On the mixing advantage

slide-5
SLIDE 5

Introduction

◮ X j i , Xi are the lifetime of an individual/unit and,

max(X 1

i , . . . , X n i ) and max(X1, X2, . . . , Xn) represent the

lifetime of population/system. All random variables are assumed to be non-negative. Xi, X 1

i , . . . , X n i are identically distributed. ◮ X1

X 1

1

X 2

1

. . . X n

1

→ M1 = E[max(X 1

1 , . . . , X n 1 )]

X2 X 1

2

X 2

2

. . . X n

2

→ M2 = E[max(X 1

2 , . . . , X n 2 )]

Kais Hamza On the mixing advantage

slide-6
SLIDE 6

Introduction

◮ X j i , Xi are the lifetime of an individual/unit and,

max(X 1

i , . . . , X n i ) and max(X1, X2, . . . , Xn) represent the

lifetime of population/system. All random variables are assumed to be non-negative. Xi, X 1

i , . . . , X n i are identically distributed. ◮ X1

X 1

1

X 2

1

. . . X n

1

→ M1 = E[max(X 1

1 , . . . , X n 1 )]

X2 X 1

2

X 2

2

. . . X n

2

→ M2 = E[max(X 1

2 , . . . , X n 2 )]

. . .

Kais Hamza On the mixing advantage

slide-7
SLIDE 7

Introduction

◮ X j i , Xi are the lifetime of an individual/unit and,

max(X 1

i , . . . , X n i ) and max(X1, X2, . . . , Xn) represent the

lifetime of population/system. All random variables are assumed to be non-negative. Xi, X 1

i , . . . , X n i are identically distributed. ◮ X1

X 1

1

X 2

1

. . . X n

1

→ M1 = E[max(X 1

1 , . . . , X n 1 )]

X2 X 1

2

X 2

2

. . . X n

2

→ M2 = E[max(X 1

2 , . . . , X n 2 )]

. . . Xn X 1

n

X 2

n

. . . X n

n

→ Mn = E[max(X 1

n , . . . , X n n )]

Kais Hamza On the mixing advantage

slide-8
SLIDE 8

Introduction

◮ X j i , Xi are the lifetime of an individual/unit and,

max(X 1

i , . . . , X n i ) and max(X1, X2, . . . , Xn) represent the

lifetime of population/system. All random variables are assumed to be non-negative. Xi, X 1

i , . . . , X n i are identically distributed. ◮ X1

X 1

1

X 2

1

. . . X n

1

→ M1 = E[max(X 1

1 , . . . , X n 1 )]

X2 X 1

2

X 2

2

. . . X n

2

→ M2 = E[max(X 1

2 , . . . , X n 2 )]

. . . Xn X 1

n

X 2

n

. . . X n

n

→ Mn = E[max(X 1

n , . . . , X n n )] ◮ We wish to compare E[max(X1, X2, . . . , Xn)] to

Mi = E[max(X 1

i , . . . , X n i )], i = 1, . . . , n.

Kais Hamza On the mixing advantage

slide-9
SLIDE 9

Introduction

A B

Reliability – Warm Duplication Method

Kais Hamza On the mixing advantage

slide-10
SLIDE 10

Introduction

A B

A A

Reliability – Warm Duplication Method

Kais Hamza On the mixing advantage

slide-11
SLIDE 11

Introduction

A B

A B A B

Reliability – Warm Duplication Method

Kais Hamza On the mixing advantage

slide-12
SLIDE 12

Introduction

A B

A B A A B B

Reliability – Warm Duplication Method

Kais Hamza On the mixing advantage

slide-13
SLIDE 13

Introduction

Kais Hamza On the mixing advantage

slide-14
SLIDE 14

Introduction

◮ Question: Is it better to mix or go with a single type?

Kais Hamza On the mixing advantage

slide-15
SLIDE 15

Introduction

◮ Question: Is it better to mix or go with a single type? ◮ Obviously, if one type dominates all others, then choosing

that type only is optimum.

Kais Hamza On the mixing advantage

slide-16
SLIDE 16

Introduction

◮ Question: Is it better to mix or go with a single type? ◮ Obviously, if one type dominates all others, then choosing

that type only is optimum.

◮ Question: What if all types are similar (no dominant type); i.e.

E[max(X 1

1 , . . . , X n 1 )] = . . . = E[max(X 1 n , . . . , X n n )]?

Kais Hamza On the mixing advantage

slide-17
SLIDE 17

Introduction

Kais Hamza On the mixing advantage

slide-18
SLIDE 18

Introduction

Assume all random variables are independent.

Kais Hamza On the mixing advantage

slide-19
SLIDE 19

Introduction

Assume all random variables are independent.

◮ It is easy to show (direct consequence of the

arithmetic-geometric mean inequality) that E[max(X1, . . . , Xn)] ≥ E[max(X 1

i , . . . , X n i )].

In fact, the same arithmetic-geometric mean inequality shows that E[max(X1, . . . , Xn)] ≥ 1 n

n

  • i=1

E[max(X 1

i , . . . , X n i )].

In other words, mixing is advantageous.

Kais Hamza On the mixing advantage

slide-20
SLIDE 20

Introduction

Assume all random variables are independent.

◮ It is easy to show (direct consequence of the

arithmetic-geometric mean inequality) that E[max(X1, . . . , Xn)] ≥ E[max(X 1

i , . . . , X n i )].

In fact, the same arithmetic-geometric mean inequality shows that E[max(X1, . . . , Xn)] ≥ 1 n

n

  • i=1

E[max(X 1

i , . . . , X n i )].

In other words, mixing is advantageous.

◮ If Mi = E[max(X 1 i , . . . , X n i )], i = 1, . . . , n, we call mixing

factor θ = E[max(X1, . . . , Xn)] max(M1, . . . , Mn) . We show that when Mi = M, θ ≤ 2 − 1/n < 2.

Kais Hamza On the mixing advantage

slide-21
SLIDE 21

Existing literature

Kais Hamza On the mixing advantage

slide-22
SLIDE 22

Existing literature

◮ An extensive literature exists on E[max(X1, . . . , Xn)] in the iid

case – see David and Nagaraja (2003). However, very little work exists for the non-identically distributed case.

Kais Hamza On the mixing advantage

slide-23
SLIDE 23

Existing literature

◮ An extensive literature exists on E[max(X1, . . . , Xn)] in the iid

case – see David and Nagaraja (2003). However, very little work exists for the non-identically distributed case.

◮ Arnold and Groeneveld (1979) obtain upper and lower bounds

  • n E[max(X1, . . . , Xn)] even when X1, . . . , Xn are not

independent and not identically distributed, but in terms of E[X1] and var(Xi), not M1, . . . , Mn. This generalises Hartley and David (1954) and Gumbel (1954) who deal with the iid case.

Kais Hamza On the mixing advantage

slide-24
SLIDE 24

Existing literature

Kais Hamza On the mixing advantage

slide-25
SLIDE 25

Existing literature

◮ Sen (1970) shows that max(X1, . . . , Xn) stochastically

dominates max(Y 1, . . . , Y n), where Y 1, . . . , Y n are iid equally-weighted probability mixtures of X1, . . . , Xn: P(max(X1, . . . , Xn) ≤ z) ≤ P(max(Y 1, . . . , Y n) ≤ z). In particular 1 n

n

  • i=1

E[max(X 1

i , . . . , X n i )]

≤ E[max(Y 1, . . . , Y n)] ≤ E[max(X1, . . . , Xn)]. However, E[max(Y 1, . . . , Y n)] cannot be expressed in terms

  • f M1, . . . , Mn.

Kais Hamza On the mixing advantage

slide-26
SLIDE 26

Unbounded independent case

Kais Hamza On the mixing advantage

slide-27
SLIDE 27

Unbounded independent case

Theorem (H., Jagers, Sudbury & Tokarev, 2009) If X1, . . . , Xn are independent random variables with the property that E[max(X 1

i , . . . , X n i )] = Mi, i = 1, 2, ..., n , then

1 n

n

  • i=1

Mi ≤ E[max(X1, . . . , Xn)] ≤ 1 n

n

  • i=1

Mi + n − 1 n max(M1, . . . , Mn). In particular, if Mi = M, i = 1, ..., n, M ≤ E[max(X1, . . . , Xn)] ≤ (2 − 1/n)M.

Kais Hamza On the mixing advantage

slide-28
SLIDE 28

Unbounded independent case

Theorem (H., Jagers, Sudbury & Tokarev, 2009) If X1, . . . , Xn are independent random variables with the property that E[max(X 1

i , . . . , X n i )] = Mi, i = 1, 2, ..., n , then

1 n

n

  • i=1

Mi ≤ E[max(X1, . . . , Xn)] ≤ 1 n

n

  • i=1

Mi + n − 1 n max(M1, . . . , Mn). In particular, if Mi = M, i = 1, ..., n, M ≤ E[max(X1, . . . , Xn)] ≤ (2 − 1/n)M. The upper bound is obtained by letting some of the random variables be concentrated on 0 and x and letting x → ∞.

Kais Hamza On the mixing advantage

slide-29
SLIDE 29

Bounded independent case

Kais Hamza On the mixing advantage

slide-30
SLIDE 30

Bounded independent case

Theorem (H. & Sudbury, 2011) If a set of random variables X1, . . . , Xn are independent, concen- trated on [0, b] and s.t. E[max(X 1

i , . . . , X n i )] = Mi, i = 1, . . . , n,

then, putting Mn = max(M1, . . . , Mn), b −

n

  • i=1

(b − Mi)1/n ≤ E[max(X1, . . . , Xn)] ≤ b − (b − Mn)

n−1

  • i=1

(1 − Mi/b)1/n.

Kais Hamza On the mixing advantage

slide-31
SLIDE 31

Bounded independent case

Kais Hamza On the mixing advantage

slide-32
SLIDE 32

Bounded independent case

Corollary In the case Mi = M, i = 1, . . . , n we have M ≤ E[max(X1, . . . , Xn)] ≤ b − b(1 − M/b)2−1/n where the latter expression approaches (2−1/n)M as b → +∞ and M(2 − M/b) as n → +∞.

Kais Hamza On the mixing advantage

slide-33
SLIDE 33

Bounded independent case

Kais Hamza On the mixing advantage

slide-34
SLIDE 34

Bounded independent case

Changing X into b − X transforms maxima into minima immediately yielding the following result.

Kais Hamza On the mixing advantage

slide-35
SLIDE 35

Bounded independent case

Changing X into b − X transforms maxima into minima immediately yielding the following result. Corollary The equivalent result for the minima, with m1 = min(m1, . . . , mn), is m1

n

  • i=2

(mi/b)1/n ≤ E[min(X1, . . . , Xn)] ≤

n

  • i=1

m1/n

i

.

Kais Hamza On the mixing advantage

slide-36
SLIDE 36

Dependent case

Kais Hamza On the mixing advantage

slide-37
SLIDE 37

Dependent case

◮ What if the random variables are NOT independent.

Kais Hamza On the mixing advantage

slide-38
SLIDE 38

Dependent case

◮ What if the random variables are NOT independent. ◮ U, V and W are independent continuous random variables.

Let X = U ∧ W and Y = V ∧ W (a ∧ b = min(a, b)).

Kais Hamza On the mixing advantage

slide-39
SLIDE 39

Dependent case

◮ What if the random variables are NOT independent. ◮ U, V and W are independent continuous random variables.

Let X = U ∧ W and Y = V ∧ W (a ∧ b = min(a, b)).

Kais Hamza On the mixing advantage

slide-40
SLIDE 40

Dependent case

◮ What if the random variables are NOT independent. ◮ U, V and W are independent continuous random variables.

Let X = U ∧ W and Y = V ∧ W (a ∧ b = min(a, b)). U V W

Kais Hamza On the mixing advantage

slide-41
SLIDE 41

Dependent case – Copulas

Kais Hamza On the mixing advantage

slide-42
SLIDE 42

Dependent case – Copulas

◮ If X has marginal F, Y has marginal G and they assume a

copula C, then (X, Y ) has joint distribution H(x, y) = C(F(x), G(y)).

Kais Hamza On the mixing advantage

slide-43
SLIDE 43

Dependent case – Copulas

◮ If X has marginal F, Y has marginal G and they assume a

copula C, then (X, Y ) has joint distribution H(x, y) = C(F(x), G(y)).

◮ Recall that a copula is defined as satisfying:

Kais Hamza On the mixing advantage

slide-44
SLIDE 44

Dependent case – Copulas

◮ If X has marginal F, Y has marginal G and they assume a

copula C, then (X, Y ) has joint distribution H(x, y) = C(F(x), G(y)).

◮ Recall that a copula is defined as satisfying:

◮ C is defined on [0, 1] × [0, 1]; Kais Hamza On the mixing advantage

slide-45
SLIDE 45

Dependent case – Copulas

◮ If X has marginal F, Y has marginal G and they assume a

copula C, then (X, Y ) has joint distribution H(x, y) = C(F(x), G(y)).

◮ Recall that a copula is defined as satisfying:

◮ C is defined on [0, 1] × [0, 1]; ◮ C(s, 0) = C(0, t) = 0; Kais Hamza On the mixing advantage

slide-46
SLIDE 46

Dependent case – Copulas

◮ If X has marginal F, Y has marginal G and they assume a

copula C, then (X, Y ) has joint distribution H(x, y) = C(F(x), G(y)).

◮ Recall that a copula is defined as satisfying:

◮ C is defined on [0, 1] × [0, 1]; ◮ C(s, 0) = C(0, t) = 0; ◮ C(s, 1) = s and C(1, t) = t; Kais Hamza On the mixing advantage

slide-47
SLIDE 47

Dependent case – Copulas

◮ If X has marginal F, Y has marginal G and they assume a

copula C, then (X, Y ) has joint distribution H(x, y) = C(F(x), G(y)).

◮ Recall that a copula is defined as satisfying:

◮ C is defined on [0, 1] × [0, 1]; ◮ C(s, 0) = C(0, t) = 0; ◮ C(s, 1) = s and C(1, t) = t; ◮ C(s2, t2) − C(s2, t1) − C(s1, t2) + C(s1, t1) ≥ 0. Kais Hamza On the mixing advantage

slide-48
SLIDE 48

Dependent case – Copulas

◮ If X has marginal F, Y has marginal G and they assume a

copula C, then (X, Y ) has joint distribution H(x, y) = C(F(x), G(y)).

◮ Recall that a copula is defined as satisfying:

◮ C is defined on [0, 1] × [0, 1]; ◮ C(s, 0) = C(0, t) = 0; ◮ C(s, 1) = s and C(1, t) = t; ◮ C(s2, t2) − C(s2, t1) − C(s1, t2) + C(s1, t1) ≥ 0.

◮ Three examples

Kais Hamza On the mixing advantage

slide-49
SLIDE 49

Dependent case – Copulas

◮ If X has marginal F, Y has marginal G and they assume a

copula C, then (X, Y ) has joint distribution H(x, y) = C(F(x), G(y)).

◮ Recall that a copula is defined as satisfying:

◮ C is defined on [0, 1] × [0, 1]; ◮ C(s, 0) = C(0, t) = 0; ◮ C(s, 1) = s and C(1, t) = t; ◮ C(s2, t2) − C(s2, t1) − C(s1, t2) + C(s1, t1) ≥ 0.

◮ Three examples

◮ Π(s, t) = st – independent case; Kais Hamza On the mixing advantage

slide-50
SLIDE 50

Dependent case – Copulas

◮ If X has marginal F, Y has marginal G and they assume a

copula C, then (X, Y ) has joint distribution H(x, y) = C(F(x), G(y)).

◮ Recall that a copula is defined as satisfying:

◮ C is defined on [0, 1] × [0, 1]; ◮ C(s, 0) = C(0, t) = 0; ◮ C(s, 1) = s and C(1, t) = t; ◮ C(s2, t2) − C(s2, t1) − C(s1, t2) + C(s1, t1) ≥ 0.

◮ Three examples

◮ Π(s, t) = st – independent case; ◮ M(s, t) = s ∧ t – perfectly positively related case; Kais Hamza On the mixing advantage

slide-51
SLIDE 51

Dependent case – Copulas

◮ If X has marginal F, Y has marginal G and they assume a

copula C, then (X, Y ) has joint distribution H(x, y) = C(F(x), G(y)).

◮ Recall that a copula is defined as satisfying:

◮ C is defined on [0, 1] × [0, 1]; ◮ C(s, 0) = C(0, t) = 0; ◮ C(s, 1) = s and C(1, t) = t; ◮ C(s2, t2) − C(s2, t1) − C(s1, t2) + C(s1, t1) ≥ 0.

◮ Three examples

◮ Π(s, t) = st – independent case; ◮ M(s, t) = s ∧ t – perfectly positively related case; ◮ W (s, t) = (s + t − 1)+ – perfectly negatively related case; Kais Hamza On the mixing advantage

slide-52
SLIDE 52

Dependent case – Copulas

◮ If X has marginal F, Y has marginal G and they assume a

copula C, then (X, Y ) has joint distribution H(x, y) = C(F(x), G(y)).

◮ Recall that a copula is defined as satisfying:

◮ C is defined on [0, 1] × [0, 1]; ◮ C(s, 0) = C(0, t) = 0; ◮ C(s, 1) = s and C(1, t) = t; ◮ C(s2, t2) − C(s2, t1) − C(s1, t2) + C(s1, t1) ≥ 0.

◮ Three examples

◮ Π(s, t) = st – independent case; ◮ M(s, t) = s ∧ t – perfectly positively related case; ◮ W (s, t) = (s + t − 1)+ – perfectly negatively related case; ◮ K(s, t) = s ∧ t − ψ(s ∧ t) + (s ∨ t)ψ(s ∧ t) – (U ∧ W , V ∧ W ). Kais Hamza On the mixing advantage

slide-53
SLIDE 53

Dependent case – A toy example

Kais Hamza On the mixing advantage

slide-54
SLIDE 54

Dependent case – A toy example

◮ n = 2.

Kais Hamza On the mixing advantage

slide-55
SLIDE 55

Dependent case – A toy example

◮ n = 2. ◮ X1, X2 take at most 2 values and assume a copula C.

Kais Hamza On the mixing advantage

slide-56
SLIDE 56

Dependent case – A toy example

◮ n = 2. ◮ X1, X2 take at most 2 values and assume a copula C. ◮ pi = P(Xi = ai), P(Xi = xi) = 1 − pi, ai ≤ xi.

Kais Hamza On the mixing advantage

slide-57
SLIDE 57

Dependent case – A toy example

◮ n = 2. ◮ X1, X2 take at most 2 values and assume a copula C. ◮ pi = P(Xi = ai), P(Xi = xi) = 1 − pi, ai ≤ xi. ◮ X 1 i and X 2 i inherit the copula of X1 and X2, C:

P(X 1

i = ai, X 2 i = ai)

= C(pi, pi) P(X 1

i = ai, X 2 i = xi)

= pi − C(pi, pi) P(X 1

i = xi, X 2 i = ai)

= pi − C(pi, pi) P(X 1

i = xi, X 2 i = xi)

= 1 − 2pi + C(pi, pi)

Kais Hamza On the mixing advantage

slide-58
SLIDE 58

Dependent case – A toy example

◮ n = 2. ◮ X1, X2 take at most 2 values and assume a copula C. ◮ pi = P(Xi = ai), P(Xi = xi) = 1 − pi, ai ≤ xi. ◮ X 1 i and X 2 i inherit the copula of X1 and X2, C:

P(X 1

i = ai, X 2 i = ai)

= C(pi, pi) P(X 1

i = ai, X 2 i = xi)

= pi − C(pi, pi) P(X 1

i = xi, X 2 i = ai)

= pi − C(pi, pi) P(X 1

i = xi, X 2 i = xi)

= 1 − 2pi + C(pi, pi)

◮ Mi = E[max(X 1 i , X 2 i )] = C(pi, pi)ai + (1 − C(pi, pi))xi,

i = 1, 2.

Kais Hamza On the mixing advantage

slide-59
SLIDE 59

Dependent case – A toy example

Kais Hamza On the mixing advantage

slide-60
SLIDE 60

Dependent case – A toy example

Assumption We assume that for any (s, t), C(s, t) − sC(t, t) ≥ 0 and C(s, t) − tC(s, s) ≥ 0. (⋆)

◮ Π, M and K satisfy this condition; W does not.

Kais Hamza On the mixing advantage

slide-61
SLIDE 61

Dependent case – A toy example

Assumption We assume that for any (s, t), C(s, t) − sC(t, t) ≥ 0 and C(s, t) − tC(s, s) ≥ 0. (⋆)

◮ Π, M and K satisfy this condition; W does not. ◮ If (U, V ) are uniform (0, 1) and have copula C, then (⋆)

translates to P(U ≤ s| max(U, V ) ≤ t) ≥ P(U ≤ s), s < t.

Kais Hamza On the mixing advantage

slide-62
SLIDE 62

Dependent case – A toy example

Kais Hamza On the mixing advantage

slide-63
SLIDE 63

Dependent case – A toy example

Assume WLOG that a1 ≤ a2. We need to consider 3 cases:

Kais Hamza On the mixing advantage

slide-64
SLIDE 64

Dependent case – A toy example

Assume WLOG that a1 ≤ a2. We need to consider 3 cases:

x1 a1 a2 x2

Kais Hamza On the mixing advantage

slide-65
SLIDE 65

Dependent case – A toy example

Assume WLOG that a1 ≤ a2. We need to consider 3 cases:

x1 a1 a2 x2

Kais Hamza On the mixing advantage

slide-66
SLIDE 66

Dependent case – A toy example

Assume WLOG that a1 ≤ a2. We need to consider 3 cases:

x1 a1 a2 x2

Kais Hamza On the mixing advantage

slide-67
SLIDE 67

Dependent case – A toy example

Kais Hamza On the mixing advantage

slide-68
SLIDE 68

Dependent case – A toy example

a1 a2 x2 x1

The case a1 ≤ x1 ≤ a2 ≤ x2 is trivial since in this case X2 dominates X1.

Kais Hamza On the mixing advantage

slide-69
SLIDE 69

Dependent case – A toy example

Kais Hamza On the mixing advantage

slide-70
SLIDE 70

Dependent case – A toy example

a1 a2 x2 x1 M2

Assume a1 ≤ a2 ≤ x2 ≤ x1.

Kais Hamza On the mixing advantage

slide-71
SLIDE 71

Dependent case – A toy example

a1 a2 x2 x1 M2

Assume a1 ≤ a2 ≤ x2 ≤ x1. E[max(X1, M2)] − E[max(X1, X2)] = p1M2 − C(p1, p2)a2 − (p1 − C(p1, p2))x2 =

  • C(p1, p2) − p1C(p2, p2)
  • (x2 − a2) ≥ 0.

Kais Hamza On the mixing advantage

slide-72
SLIDE 72

Dependent case – A toy example

a1 a2 x2 x1 M2

Assume a1 ≤ a2 ≤ x2 ≤ x1. E[max(X1, M2)] − E[max(X1, X2)] = p1M2 − C(p1, p2)a2 − (p1 − C(p1, p2))x2 =

  • C(p1, p2) − p1C(p2, p2)
  • (x2 − a2) ≥ 0.

Therefore we may replace X2 with M2.

Kais Hamza On the mixing advantage

slide-73
SLIDE 73

Dependent case – A toy example

Kais Hamza On the mixing advantage

slide-74
SLIDE 74

Dependent case – A toy example

a1 a2 x2 x1 Assume a1 ≤ a2 ≤ x1 ≤ x2.

Kais Hamza On the mixing advantage

slide-75
SLIDE 75

Dependent case – A toy example

a1 a2 x2 x1 Assume a1 ≤ a2 ≤ x1 ≤ x2. We vary a2 and x2 keeping p2 (and a1, x1, p1) constant: E[max(X 1

2 , X 2 2 )] = C(p2, p2)a2 + (1 − C(p2, p2))x2 = M2.

Kais Hamza On the mixing advantage

slide-76
SLIDE 76

Dependent case – A toy example

a1 a2 x2 x1 Assume a1 ≤ a2 ≤ x1 ≤ x2. We vary a2 and x2 keeping p2 (and a1, x1, p1) constant: E[max(X 1

2 , X 2 2 )] = C(p2, p2)a2 + (1 − C(p2, p2))x2 = M2.

Then, the linear function E[max(X1, X2)] = C(p1, p2)a2 + (1 − p2)x2 + (p2 − C(p1, p2))x1 is maximum at one of the 3 boundary points.

Kais Hamza On the mixing advantage

slide-77
SLIDE 77

Dependent case – A toy example

Kais Hamza On the mixing advantage

slide-78
SLIDE 78

Dependent case – A toy example

a1 a2 x2 x1 Assume a1 ≤ a2 ≤ x1 ≤ x2. We vary a2 and x2 keeping p2 (and a1, x1, p1) constant: E[max(X 1

2 , X 2 2 )] = C(p2, p2)a2 + (1 − C(p2, p2))x2 = M2.

Then, the linear function E[max(X1, X2)] = C(p1, p2)a2 + (1 − p2)x2 + (p2 − C(p1, p2))x1 is maximum at one of the 3 boundary points.

Kais Hamza On the mixing advantage

slide-79
SLIDE 79

Dependent case – A toy example

a1 a2 x2 x1

◮ a2 = x1. In this case X2 dominates X1.

Kais Hamza On the mixing advantage

slide-80
SLIDE 80

Dependent case – A toy example

a1 a2 x2 x1

◮ a2 = x1. In this case X2 dominates X1. ◮ a2 = a1. In this case we may collapse X1 into M1.

Kais Hamza On the mixing advantage

slide-81
SLIDE 81

Dependent case – A toy example

a1 a2 x2 x1

◮ a2 = x1. In this case X2 dominates X1. ◮ a2 = a1. In this case we may collapse X1 into M1. ◮ x2 = x1. In this case we may collapse X2 into M2.

Kais Hamza On the mixing advantage

slide-82
SLIDE 82

Dependent case – A toy example

a1 a2 x2 x1

◮ a2 = x1. In this case X2 dominates X1. ◮ a2 = a1. In this case we may collapse X1 into M1. ◮ x2 = x1. In this case we may collapse X2 into M2. ◮ In any case, we may assume that X2 = M2 and a1 ≤ M2 ≤ x1.

Kais Hamza On the mixing advantage

slide-83
SLIDE 83

Dependent case – A toy example

1 1 1 1 1

a x M x p − − =

a1 x1 M2

Kais Hamza On the mixing advantage

slide-84
SLIDE 84

Dependent case – A toy example

1 1 1 1 1

a x M x p − − =

a1 x1 M2

We vary a1 and p1 keeping x1 constant: E[max(X 1

1 , X 2 1 )] = C(p1, p1)a1 + (1 − C(p1, p1))x1 = M1.

Kais Hamza On the mixing advantage

slide-85
SLIDE 85

Dependent case – A toy example

1 1 1 1 1

a x M x p − − =

a1 x1 M2

We vary a1 and p1 keeping x1 constant: E[max(X 1

1 , X 2 1 )] = C(p1, p1)a1 + (1 − C(p1, p1))x1 = M1.

Then, E[max(X1, M2)] = p1M2 + (1 − p1)x1 = x1 − (x1 − M2)p1 is maximum for p1 minimum i.e. a1 = 0.

Kais Hamza On the mixing advantage

slide-86
SLIDE 86

Dependent case – A toy example

a1=0 x1 M2

Therefore we may assume that X2 = M2, a1 = 0 and 0 ≤ M2 ≤ x1.

Kais Hamza On the mixing advantage

slide-87
SLIDE 87

Dependent case – A toy example

a1=0 x1 M2

Therefore we may assume that X2 = M2, a1 = 0 and 0 ≤ M2 ≤ x1. In this case E[max(X1, X2)] = p1M2 + (1 − p1) M1 1 − C(p1, p1) Theorem E[max(X1, X2)] ≤ sup

0≤r<1

  • M2r + M1

1 − r 1 − C(r, r)

  • .

Kais Hamza On the mixing advantage

slide-88
SLIDE 88

The upper bound for Π, M and K

Kais Hamza On the mixing advantage

slide-89
SLIDE 89

The upper bound for Π, M and K

Let γ(r) = C(r, r) and assume that M1 ≤ M2.

◮ C = Π. In this case γ(r) = r2, γ′(1) = 2 and

E[max(X1, X2)] ≤ 1 2M1 + M2.

Kais Hamza On the mixing advantage

slide-90
SLIDE 90

The upper bound for Π, M and K

Let γ(r) = C(r, r) and assume that M1 ≤ M2.

◮ C = Π. In this case γ(r) = r2, γ′(1) = 2 and

E[max(X1, X2)] ≤ 1 2M1 + M2.

◮ C = M. In this case γ(r) = r, γ′(1) = 1 and

E[max(X1, X2)] ≤ M1 + M2.

Kais Hamza On the mixing advantage

slide-91
SLIDE 91

The upper bound for Π, M and K

Let γ(r) = C(r, r) and assume that M1 ≤ M2.

◮ C = Π. In this case γ(r) = r2, γ′(1) = 2 and

E[max(X1, X2)] ≤ 1 2M1 + M2.

◮ C = M. In this case γ(r) = r, γ′(1) = 1 and

E[max(X1, X2)] ≤ M1 + M2.

◮ C = K. In this case γ(r) = r − ψ(r) + rψ(r), γ′(1) = 2 and

E[max(X1, X2)] ≤ 1 2M1 + M2.

Kais Hamza On the mixing advantage

slide-92
SLIDE 92

The upper bound for Π, M and K

Let γ(r) = C(r, r) and assume that M1 ≤ M2.

◮ C = Π. In this case γ(r) = r2, γ′(1) = 2 and

E[max(X1, X2)] ≤ 1 2M1 + M2.

◮ C = M. In this case γ(r) = r, γ′(1) = 1 and

E[max(X1, X2)] ≤ M1 + M2.

◮ C = K. In this case γ(r) = r − ψ(r) + rψ(r), γ′(1) = 2 and

E[max(X1, X2)] ≤ 1 2M1 + M2.

◮ Note that the definition of Mi depends on C and the three

bounds cannot be compared.

Kais Hamza On the mixing advantage

slide-93
SLIDE 93

References

◮ Arnold B.C. and Groeneveld R.A. (1979), Bounds on

Expectations of Linear Systematic Statistics Based on Dependent Samples, Ann. Stat. 7, pp 220–223..

◮ Balakrishnan N. and Balasubramanian K. (2008), Revisiting

Sen’s inequalities on order statistics, Statist. Probab. Letters 78, pp 616–621.

◮ David H.A. and Nagaraja H.N. (2003), Order Statistics, John

Wiley & Sons, Hoboken, New Jersey, 3rd ed.

◮ Hamza K., Jagers P., Sudbury A. and Tokarev D. (2009), The

mixing advantage is less than 2, Extremes 12, no. 1, pp 19–31.

◮ Hamza K. and Sudbury A. (2011), The mixing advantage for

bounded random variables, Statist. Probab. Letters 81, no. 8, pp 1190–1195.

Kais Hamza On the mixing advantage

slide-94
SLIDE 94

Refrences

◮ Hartley H.O. and David H.A. (1954), Universal Bounds for

Mean Range and Extreme Observation, Ann. Math. Stat. 25, pp 85–89..

◮ Gumbel E.J. (1954), The Maxima of the Mean Largest Value

and of the Range, Ann. Math. Stat. 25, pp 76–84..

◮ Sen P.K. (1970), A Note on Order Statistics for

Heterogeneous Distributions, Ann. Math. Stat. 41, pp 2137–2139..

◮ Tokarev D. (2007), Galton-Watson processes and extinction in

population systems, PhD Thesis, Monash University.

Kais Hamza On the mixing advantage