From Arabidopsis roots to bilinear equations Dustin Cartwright 1 - - PowerPoint PPT Presentation

from arabidopsis roots to bilinear equations
SMART_READER_LITE
LIVE PREVIEW

From Arabidopsis roots to bilinear equations Dustin Cartwright 1 - - PowerPoint PPT Presentation

From Arabidopsis roots to bilinear equations Dustin Cartwright 1 October 22, 2008 1 joint with Philip Benfey, Siobhan Brady, David Orlando (Duke University) and Bernd Sturmfels (UC Berkeley), research supported by the DARPA project Fundamental


slide-1
SLIDE 1

From Arabidopsis roots to bilinear equations

Dustin Cartwright 1 October 22, 2008

1joint with Philip Benfey, Siobhan Brady, David Orlando (Duke University)

and Bernd Sturmfels (UC Berkeley), research supported by the DARPA project Fundamental Laws of Biology

slide-2
SLIDE 2

Arabidopsis root

slide-3
SLIDE 3

Arabidopsis root

Gene expression microarrays are a tool to understand dynamics and regulatory processes.

slide-4
SLIDE 4

Arabidopsis root

Gene expression microarrays are a tool to understand dynamics and regulatory processes. Two ways of separating cells in the lab:

◮ Chemically, using

18 markers (colors in diagram A)

slide-5
SLIDE 5

Arabidopsis root

Gene expression microarrays are a tool to understand dynamics and regulatory processes. Two ways of separating cells in the lab:

◮ Chemically, using

18 markers (colors in diagram A)

◮ Physically, using

13 longitudinal sections (red lines in diagram B)

slide-6
SLIDE 6

Measurement along two axes

◮ Markers measure variation among cell types.

slide-7
SLIDE 7

Measurement along two axes

◮ Markers measure variation among cell types. ◮ Longitudinal sections measure variation along developmental

stage.

slide-8
SLIDE 8

Measurement along two axes

◮ Markers measure variation among cell types. ◮ Longitudinal sections measure variation along developmental

stage. Na¨ ıve approach would use variation among each set of experiments as proxies for variation along each of the two axes.

slide-9
SLIDE 9

Problem with na¨ ıve approach

Correspondence between markers and cell types is imperfect.

slide-10
SLIDE 10

Problem with na¨ ıve approach

Correspondence between markers and cell types is imperfect. For example, the sample labelled APL consists of mixture of two cell types: cell type section phloem phloem companion cells 12

1 16 1 16

. . . . . . . . . 7

1 16 1 16

6

1 16

. . . . . . . . . 3

1 16

2 1 columella

slide-11
SLIDE 11

Problem with na¨ ıve approach

Similarly, the longitudinal sections do not have the same mixture of

  • cells. For example:

◮ In each of sections 1-5, 30-50% of the cells are lateral root

cap cells.

slide-12
SLIDE 12

Problem with na¨ ıve approach

Similarly, the longitudinal sections do not have the same mixture of

  • cells. For example:

◮ In each of sections 1-5, 30-50% of the cells are lateral root

cap cells.

◮ In sections 6-12, there are no lateral root cap cells.

slide-13
SLIDE 13

Problem with na¨ ıve approach

Similarly, the longitudinal sections do not have the same mixture of

  • cells. For example:

◮ In each of sections 1-5, 30-50% of the cells are lateral root

cap cells.

◮ In sections 6-12, there are no lateral root cap cells.

Conclusion: Need to analyze each transcript across all 31 (= 13 + 18) experiments to model the expression pattern in the whole root.

slide-14
SLIDE 14

Model

◮ A cluster consists of cells of the same type in the same

  • section. Each cluster has an expression level.
slide-15
SLIDE 15

Model

◮ A cluster consists of cells of the same type in the same

  • section. Each cluster has an expression level.

◮ For each marker and each longitudinal section, we have a

measurement functional, a linear combination of the expression levels in different clusters.

slide-16
SLIDE 16

Model

◮ A cluster consists of cells of the same type in the same

  • section. Each cluster has an expression level.

◮ For each marker and each longitudinal section, we have a

measurement functional, a linear combination of the expression levels in different clusters. The coefficients of these functionals can be determined from:

◮ Numbers of cells present in each section ◮ Marker selection patterns

slide-17
SLIDE 17

Model

◮ A cluster consists of cells of the same type in the same

  • section. Each cluster has an expression level.

◮ For each marker and each longitudinal section, we have a

measurement functional, a linear combination of the expression levels in different clusters. The coefficients of these functionals can be determined from:

◮ Numbers of cells present in each section ◮ Marker selection patterns

Under-constrained system: 31 (= 13 + 18) functionals and 129 clusters.

slide-18
SLIDE 18

Assumption

Since the system is under constrained, we make the following assumption.

slide-19
SLIDE 19

Assumption

Since the system is under constrained, we make the following assumption.

◮ The dependence on the expression level on the section is

independent of the dependence on the cell type.

slide-20
SLIDE 20

Assumption

Since the system is under constrained, we make the following assumption.

◮ The dependence on the expression level on the section is

independent of the dependence on the cell type.

◮ More precisely, the expression level of cluster in section i and

type j is xiyj for some vectors x and y.

slide-21
SLIDE 21

Assumption

Since the system is under constrained, we make the following assumption.

◮ The dependence on the expression level on the section is

independent of the dependence on the cell type.

◮ More precisely, the expression level of cluster in section i and

type j is xiyj for some vectors x and y.

Example

If the expression level is either 0 or 1 (off or on), then our assumption says that it is 1 for the combination of some subset of the sections and some subset of the cell types.

slide-22
SLIDE 22

Non-negative bilinear equations

A(1), . . . , A(k) n × m non-negative matrices (cell mixture)

  • 1, . . . , ok

non-negative scalars (expression levels) Solve (approximately) f1(x, y) := xtA(1)y = o1 . . . fk(x, y) := xtA(k)y = ok x1 + · · · + xn = 1

slide-23
SLIDE 23

Non-negative bilinear equations

A(1), . . . , A(k) n × m non-negative matrices (cell mixture)

  • 1, . . . , ok

non-negative scalars (expression levels) Solve (approximately) f1(x, y) := xtA(1)y = o1 . . . fk(x, y) := xtA(k)y = ok x1 + · · · + xn = 1 for x and y non-negative vectors of dimensions n × 1 and m × 1 respectively.

slide-24
SLIDE 24

Probabilistic interpretation

fℓ(x, y) :=

  • i,j

A(ℓ)

ij xiyj for ℓ = 1, . . . , k

Up to scaling, this vector has the form of the family of probability distributions (depending on vectors x and y)

slide-25
SLIDE 25

Probabilistic interpretation

fℓ(x, y) :=

  • i,j

A(ℓ)

ij xiyj for ℓ = 1, . . . , k

Up to scaling, this vector has the form of the family of probability distributions (depending on vectors x and y) coming from the following process:

  • 1. Pick a pair of integers from {1, . . . , n} × {1, . . . , m} with (i, j)

having probability proportional to

ℓ A(ℓ) ij

  • xiyj
slide-26
SLIDE 26

Probabilistic interpretation

fℓ(x, y) :=

  • i,j

A(ℓ)

ij xiyj for ℓ = 1, . . . , k

Up to scaling, this vector has the form of the family of probability distributions (depending on vectors x and y) coming from the following process:

  • 1. Pick a pair of integers from {1, . . . , n} × {1, . . . , m} with (i, j)

having probability proportional to

ℓ A(ℓ) ij

  • xiyj
  • 2. Output an integer from {1, . . . , k}. Conditional on having

picked i and j in the previous step, the probability of

  • utputing ℓ is:

A(ℓ)

ij / ℓ A(ℓ) ij

slide-27
SLIDE 27

Maximum Likelihood Estimation

Rescaling both sides of our system of equations: fℓ(x, y)

  • ℓ′ fℓ′(x, y) =
  • ℓ′ oℓ′ for ℓ = 1, . . . , k
slide-28
SLIDE 28

Maximum Likelihood Estimation

Rescaling both sides of our system of equations: fℓ(x, y)

  • ℓ′ fℓ′(x, y) =
  • ℓ′ oℓ′ for ℓ = 1, . . . , k

Finding an approximate solution to these equations is known as Maximum Likelihood Estimation.

slide-29
SLIDE 29

Kullback-Leibler divergence

Kullback-Leibler divergence gives a way of comparing two probability distributions: D(zf (x, y)) :=

zℓ log zℓ fℓ(x)

slide-30
SLIDE 30

Kullback-Leibler divergence

Kullback-Leibler divergence gives a way of comparing two probability distributions: D(zf (x, y)) :=

zℓ log zℓ fℓ(x)

  • − zℓ + fℓ(x, y)

We generalize divergence to any pair of non-negative vectors.

slide-31
SLIDE 31

Kullback-Leibler divergence

Kullback-Leibler divergence gives a way of comparing two probability distributions: D(zf (x, y)) :=

zℓ log zℓ fℓ(x)

  • − zℓ + fℓ(x, y)

We generalize divergence to any pair of non-negative vectors. By approximate solution to a system, we will mean the a solution which minimizes the Kullback-Leibler divergence.

slide-32
SLIDE 32

Expectation Maximization

Want to solve:

  • i,j

A(ℓ)

ij xiyj = oℓ for ℓ = 1, . . . , k

(1)

slide-33
SLIDE 33

Expectation Maximization

Want to solve:

  • i,j

A(ℓ)

ij xiyj = oℓ for ℓ = 1, . . . , k

(1)

◮ Start with guesses ˜

x, ˜ y

slide-34
SLIDE 34

Expectation Maximization

Want to solve:

  • i,j

A(ℓ)

ij xiyj = oℓ for ℓ = 1, . . . , k

(1)

◮ Start with guesses ˜

x, ˜ y

◮ Estimate contribution of (i, j) term of left side of equation 1

needed to obtain equality: A(ℓ)

ij ˜

xi ˜ yj

  • i′j′ A(ℓ)

i′j′˜

xi ˜ yj

  • ℓ =: eijℓ
slide-35
SLIDE 35

Expectation Maximization

Want to solve:

  • i,j

A(ℓ)

ij xiyj = oℓ for ℓ = 1, . . . , k

(1)

◮ Start with guesses ˜

x, ˜ y

◮ Estimate contribution of (i, j) term of left side of equation 1

needed to obtain equality: A(ℓ)

ij ˜

xi ˜ yj

  • i′j′ A(ℓ)

i′j′˜

xi ˜ yj

  • ℓ =: eijℓ

◮ Find approximate solution to system:

A(ℓ)

ij

  • xiyj ≈

eijℓ =: eij

slide-36
SLIDE 36

Expectation Maximization

Want to solve:

  • i,j

A(ℓ)

ij xiyj = oℓ for ℓ = 1, . . . , k

(1)

◮ Start with guesses ˜

x, ˜ y

◮ Estimate contribution of (i, j) term of left side of equation 1

needed to obtain equality: A(ℓ)

ij ˜

xi ˜ yj

  • i′j′ A(ℓ)

i′j′˜

xi ˜ yj

  • ℓ =: eijℓ

◮ Find approximate solution to system:

A(ℓ)

ij

  • xiyj ≈

eijℓ =: eij

◮ Repeat until convergence

slide-37
SLIDE 37

Likelihood maximization for monomial models

g : Rn × Rm → Rnm (xi), (yj) → Aijxiyj where Aij =

ℓ A(ℓ) ij .

slide-38
SLIDE 38

Likelihood maximization for monomial models

g : Rn × Rm → Rnm (xi), (yj) → Aijxiyj where Aij =

ℓ A(ℓ) ij .

Moment map (taking row sums and column sums): µ: Rnm → Rn × Rm bij →

j

bij

  • ,

i

bij

slide-39
SLIDE 39

Likelihood maximization for monomial models

g : Rn × Rm → Rnm (xi), (yj) → Aijxiyj where Aij =

ℓ A(ℓ) ij .

Moment map (taking row sums and column sums): µ: Rnm → Rn × Rm bij →

j

bij

  • ,

i

bij

  • Theorem

Kullback-Leibler divergence D(zg(x, y)) is minimized over all x and y when µ(z) equals µ(g(x, y)).

slide-40
SLIDE 40

Inverting the moment map: Iterative Proportional Fitting

Rnm µ Rn × Rm g(x, y) b µ(g(x, y)) µ(b)

slide-41
SLIDE 41

Inverting the moment map: Iterative Proportional Fitting

◮ Adjust ˜

xi: ˜ xi ← ˜ xi

  • j bij
  • j aij˜

xi ˜ yj

◮ Adjust ˜

yi: ˜ yj ← ˜ yj

  • i bij
  • i aij˜

xi ˜ yj

◮ Iterate until convergence

Rnm µ Rn × Rm g(x, y) b µ(g(x, y)) µ(b)

slide-42
SLIDE 42

Validation: Preliminary results

On the left is a visual representation

  • f the reconstructed expression

levels. On the right, the expression levels for the same transcript are visualized using GFP.