Machine learning for CalabiYau manifolds Harold Erbin Asc , Lmu - - PowerPoint PPT Presentation

machine learning for calabi yau manifolds
SMART_READER_LITE
LIVE PREVIEW

Machine learning for CalabiYau manifolds Harold Erbin Asc , Lmu - - PowerPoint PPT Presentation

Machine learning for CalabiYau manifolds Harold Erbin Asc , Lmu (Germany) Machine Learning Landscape, Ictp , Trieste 12th December 2018 1 / 35 Outline: 1. Motivations Motivations Machine learning CalabiYau 3-folds Data analysis ML


slide-1
SLIDE 1

Machine learning for Calabi–Yau manifolds

Harold Erbin

Asc, Lmu (Germany)

Machine Learning Landscape, Ictp, Trieste – 12th December 2018

1 / 35

slide-2
SLIDE 2

Outline: 1. Motivations

Motivations Machine learning Calabi–Yau 3-folds Data analysis ML analysis Conclusion

2 / 35

slide-3
SLIDE 3

String phenomenology

Goal

Find “the” Standard Model from string theory. Method:

◮ type II / heterotic strings, M-theory, F-theory: D = 10, 11, 12 ◮ vacuum choice (flux compactification):

◮ (typically) Calabi–Yau (CY) 3- or 4-fold ◮ fluxes and intersecting branes

→ reduction to D = 4

◮ check consistency (tadpole, susy. . . ) ◮ read the D = 4 QFT (gauge group, spectrum. . . )

3 / 35

slide-4
SLIDE 4

String phenomenology

Goal

Find “the” Standard Model from string theory. Method:

◮ type II / heterotic strings, M-theory, F-theory: D = 10, 11, 12 ◮ vacuum choice (flux compactification):

◮ (typically) Calabi–Yau (CY) 3- or 4-fold ◮ fluxes and intersecting branes

→ reduction to D = 4

◮ check consistency (tadpole, susy. . . ) ◮ read the D = 4 QFT (gauge group, spectrum. . . )

No vacuum selection mechanism ⇒ string landscape

3 / 35

slide-5
SLIDE 5

Landscape mapping

String phenomenology:

◮ find consistent string models ◮ find generic/common features ◮ reproduce the Standard Model

4 / 35

slide-6
SLIDE 6

Landscape mapping

String phenomenology:

◮ find consistent string models ◮ find generic/common features ◮ reproduce the Standard Model

Typical challenges: properties and equations involving many integers

4 / 35

slide-7
SLIDE 7

Types of data

Calabi–Yau (CY) manifolds

◮ CICY (complete intersection in products of projective spaces):

7890 (3-fold), 921,497 (4-fold)

◮ Kreuzer–Skarke (reflexive polyhedra):

473,800,776 (d = 4) String and F-theory models involve huge numbers

◮ 10500 ◮ 10755 ◮ 10272,000 ◮ . . .

5 / 35

slide-8
SLIDE 8

Types of data

Calabi–Yau (CY) manifolds

◮ CICY (complete intersection in products of projective spaces):

7890 (3-fold), 921,497 (4-fold)

◮ Kreuzer–Skarke (reflexive polyhedra):

473,800,776 (d = 4) String and F-theory models involve huge numbers

◮ 10500 ◮ 10755 ◮ 10272,000 ◮ . . .

→ use machine learning

5 / 35

slide-9
SLIDE 9

Plan

Analysis of CICY 3-fold

◮ ML methodology ◮ results and discussions of Hodge numbers

In progress with: Vincent Lahoche, Mohamed El Amine Seddik, Mohamed Tamaazousti (List, Cea).

6 / 35

slide-10
SLIDE 10

Outline: 2. Machine learning

Motivations Machine learning Calabi–Yau 3-folds Data analysis ML analysis Conclusion

7 / 35

slide-11
SLIDE 11

Definition

Machine learning (Samuel)

The field of study that gives computers the ability to learn without being explicitly programmed.

Machine learning (Mitchell)

A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P if its performance at tasks in T, as measured by P, improves with experience E.

8 / 35

slide-12
SLIDE 12

Deep neural network

Architecture:

◮ 1–many hidden layers ◮ link: weighted input ◮ neuron: non-linear "activation

function" Summary: x(n+1) = g(n+1)(W (n)x(n)). Generic method: fixed functions g(n), learn weights W (n)

9 / 35

slide-13
SLIDE 13

Deep neural network

x(1)

i1

≡ xi1 x(2)

i2

= g(2)W (1)

i2i1 x(1) i1

  • fi3(xi1) ≡ x(3)

i3

= g(3)W (2)

i3i2 x(2) i2

  • i1 = 1, 2, 3; i2 = 1, . . . , 4; i3 = 1, 2

Summary: x(n+1) = g(n+1)(W (n)x(n)). Generic method: fixed functions g(n), learn weights W (n)

9 / 35

slide-14
SLIDE 14

Learning method

◮ define a loss function L

L =

Ntrain

  • i=1

distance

y(train)

i

, y(pred)

i

  • ◮ minimize the loss function (iterated gradient descent. . . )

10 / 35

slide-15
SLIDE 15

Learning method

◮ define a loss function L

L =

Ntrain

  • i=1

distance

y(train)

i

, y(pred)

i

  • ◮ minimize the loss function (iterated gradient descent. . . )

◮ main risk: overfitting (= cannot generalize)

→ various solutions (regularization, dropout. . . ) → split data set in two (training and test)

10 / 35

slide-16
SLIDE 16

ML workflow

“Naive” workflow:

  • 1. get raw data
  • 2. write neural network with

many layers

  • 3. feed raw data to neural

network

  • 4. get nice results

(or give up)

11 / 35

slide-17
SLIDE 17

ML workflow

Real-world workflow:

  • 1. understand the problem
  • 2. exploratory data analysis

◮ feature engineering ◮ feature selection

  • 3. baseline model

◮ full working pipeline ◮ lower-bound on accuracy

  • 4. validation strategy
  • 5. machine learning model
  • 6. ensembling

Pragmatic ref.: coursera.org/learn/competitive-data-science

11 / 35

slide-18
SLIDE 18

Complex neural network

12 / 35

slide-19
SLIDE 19

Complex neural network

Particularities:

◮ fi(I) : engineered features ◮ identical outputs (stabilisation)

12 / 35

slide-20
SLIDE 20

Outline: 3. Calabi–Yau 3-folds

Motivations Machine learning Calabi–Yau 3-folds Data analysis ML analysis Conclusion

13 / 35

slide-21
SLIDE 21

Calabi-Yau

Complete intersection Calabi–Yau (CICY) 3-fold:

◮ CY: complex manifold with vanishing first Chern class ◮ complete intersection: non-degenerate hypersurface in

products of projective spaces

◮ hypersurface = solution to system of homogeneous

polynomial equations

14 / 35

slide-22
SLIDE 22

Calabi-Yau

Complete intersection Calabi–Yau (CICY) 3-fold:

◮ CY: complex manifold with vanishing first Chern class ◮ complete intersection: non-degenerate hypersurface in

products of projective spaces

◮ hypersurface = solution to system of homogeneous

polynomial equations

◮ described by configuration matrix m × k

X =

  

Pn1 a1

1

· · · a1

k

. . . . . . ... . . . Pnm am

1

· · · am

k

  

dimC X =

m

  • r=1

nr − k = 3, nr + 1 =

k

  • α=1

ar

α ◮ ar α power of coordinates on Pnr in αth equation

14 / 35

slide-23
SLIDE 23

Configuration matrix

Examples

◮ quintic

  • P4

x

5

  • =

  • a

(X a)5 = 0

◮ 2 projective spaces, 3 equations

  • P3

x

3 1 P3

y

3 1

  • =

      

fabc X aX bX c = 0 gαβγY αY βY γ = 0 haα X aY α = 0

15 / 35

slide-24
SLIDE 24

Configuration matrix

Examples

◮ quintic

  • P4

x

5

  • =

  • a

(X a)5 = 0

◮ 2 projective spaces, 3 equations

  • P3

x

3 1 P3

y

3 1

  • =

      

fabc X aX bX c = 0 gαβγY αY βY γ = 0 haα X aY α = 0 Classification

◮ invariances (→ huge redundancy)

◮ permutation of lines and columns ◮ identities between subspaces

◮ but:

◮ constraints ⇒ bound on matrix size ◮ ∃ “favourable” configuration 15 / 35

slide-25
SLIDE 25

Topology

Why topology?

◮ no metric known for compact CY (cannot perform KK

reduction explicitly)

◮ topological numbers → 4d properties (number of fields,

representations, gauge symmetry. . . )

16 / 35

slide-26
SLIDE 26

Topology

Why topology?

◮ no metric known for compact CY (cannot perform KK

reduction explicitly)

◮ topological numbers → 4d properties (number of fields,

representations, gauge symmetry. . . ) Topological properties

◮ Hodge numbers hp,q (number of harmonic (p, q)-forms)

here: h1,1, h2,1

◮ Euler number χ = 2(h11 − h21) ◮ Chern classes ◮ triple intersection numbers ◮ line bundle cohomologies

16 / 35

slide-27
SLIDE 27

Topology

Why topology?

◮ no metric known for compact CY (cannot perform KK

reduction explicitly)

◮ topological numbers → 4d properties (number of fields,

representations, gauge symmetry. . . ) Topological properties

◮ Hodge numbers hp,q (number of harmonic (p, q)-forms)

here: h1,1, h2,1

◮ Euler number χ = 2(h11 − h21) ◮ Chern classes ◮ triple intersection numbers ◮ line bundle cohomologies

16 / 35

slide-28
SLIDE 28

Datasets

CICY have been classified

◮ 7890 configurations (but ∃ redundancies) ◮ number of product spaces: 22 ◮ h1,1 ∈ [0, 19], h2,1 ∈ [0, 101] ◮ 266 combinations (h1,1, h2,1) ◮ ar α ∈ [0, 5]

Original [Candelas-Dale-Lutken-Schimmrigk ’88][Green-Hubsch-Lutken ’89]

◮ maximal size: 12 × 15 ◮ number of favourable matrices: 4874

Favourable [1708.07907, Anderson-Gao-Gray-Lee]

◮ maximal size: 15 × 18 ◮ number of favourable matrices: 7820

17 / 35

slide-29
SLIDE 29

Data

0.0 2.5 5.0 7.5 10.0 12.5 15.0 17.5 h11 10

1

100 101 102 103 frequency h11 20 40 60 80 100 h21 10

1

100 101 102 frequency h21 5 10 15 h11 20 40 60 80 100 h21 Sizes 1 10 100 500 Sizes 1 10 100 500 50 100 150 200 250 300 350

18 / 35

slide-30
SLIDE 30

Goal and methodology

Philosophy

Start with the original dataset, derive everything else from configuration matrix and machine learning only.

Current goal

Input: configuration matrix − → Output: Hodge numbers

  • 1. CICY: well studied, all topological quantities known

→ use as a sandbox

  • 2. h2,1: more difficult than h1,1

→ prepare for studying CICY 4-folds

  • 3. both original and favourable datasets

Continue the analysis from:

[1706.02714, He] [1806.03121, Bull-He-Jejjala-Mishra]

19 / 35

slide-31
SLIDE 31

Outline: 4. Data analysis

Motivations Machine learning Calabi–Yau 3-folds Data analysis ML analysis Conclusion

20 / 35

slide-32
SLIDE 32

Feature engineering

Process of creating new features derived from the raw input data. Some examples:

◮ number of projective spaces (rows), m = num_cp ◮ number of equations (columns), k ◮ number of CP1 ◮ number of CP2 ◮ number of CPn with n = 1 ◮ Frobenius norm of the matrix ◮ list of the projective space dimensions and statistics thereof

(min, max, mean, median)

◮ K-nearest neighbour (KNN) clustering (with K = 2, . . . , 5)

21 / 35

slide-33
SLIDE 33

Feature selection

Select the most important features to draw attention of the ML algorithm to salient features in order to ease the learning. Discovery methods:

◮ correlation matrix ◮ random forests ◮ scatter plots ◮ trial and error ◮ etc.

22 / 35

slide-34
SLIDE 34

Correlation matrix

Original dataset

h21 h11 num_cp num_cp_1 num_cp_2 num_cp_neq1 rank_matrix norm_matrix num_eqs num_ex h21 h11 num_cp num_cp_1 num_cp_2 num_cp_neq1 rank_matrix norm_matrix num_eqs num_ex 1.00 0.75 0.50 0.25 0.00 0.25 0.50 0.75 1.00

Favourable dataset

h21 h11 num_cp num_cp_1 num_cp_2 num_cp_neq1 rank_matrix norm_matrix num_eqs num_ex h21 h11 num_cp num_cp_1 num_cp_2 num_cp_neq1 rank_matrix norm_matrix num_eqs num_ex 1.00 0.75 0.50 0.25 0.00 0.25 0.50 0.75 1.00

23 / 35

slide-35
SLIDE 35

Random forest

Large number of decision trees trained on different subsets and averaged on the outputs. The most relevant features appear at the top of the trees. = ⇒ classify feature importance

Original

num_cp num_cp_1 num_cp_2 num_cp_neq1 num_eqs num_ex rank_matrix norm_matrix 0.0 0.1 0.2 0.3 0.4 0.5 0.6 Importance h11 h21

Favourable

num_cp num_cp_1 num_cp_2 num_cp_neq1 num_eqs num_ex rank_matrix norm_matrix 0.0 0.2 0.4 0.6 0.8 Importance h11 h21

24 / 35

slide-36
SLIDE 36

Scatter plots: h1,1

Original

2 4 6 8 10 12 num_cp 0.0 2.5 5.0 7.5 10.0 12.5 15.0 17.5 20.0 h11 Sizes 1 10 100 1000 Sizes 1 10 100 1000 200 400 600 800 1000

Favourable

2 4 6 8 10 12 14 num_cp 0.0 2.5 5.0 7.5 10.0 12.5 15.0 17.5 20.0 h11 Sizes 1 10 100 1000 Sizes 1 10 100 1000 200 400 600 800 1000 1200 1400

25 / 35

slide-37
SLIDE 37

Scatter plots: h2,1

Original

2 4 6 8 10 12 num_cp 20 40 60 80 100 h21 Sizes 1 10 100 1000 Sizes 1 10 100 1000 50 100 150 200 250

Favourable

2 4 6 8 10 12 14 num_cp 20 40 60 80 100 h21 Sizes 1 10 100 1000 Sizes 1 10 100 1000 50 100 150 200 250 300 350

26 / 35

slide-38
SLIDE 38

Outline: 5. ML analysis

Motivations Machine learning Calabi–Yau 3-folds Data analysis ML analysis Conclusion

27 / 35

slide-39
SLIDE 39

Strategy

Questions:

◮ data diminution: remove outliers? (0.74%) ◮ data augmentation: use data invariance to generate more

inputs?

◮ classification or regression? ◮ normalise inputs/outputs? (shift by mean, divide by variance)

Classification vs regression:

◮ classification: assume knowledge of boundaries ◮ regression: outputs of different size

→ normalize data ≈ use continuous variable Regression: better for generalization

28 / 35

slide-40
SLIDE 40

Algorithms

Possibilities (starting from original dataset):

◮ neural network with trivial architecture

(matrix → hodges)

◮ neural network with non-trivial architecture

(matrix + engineered features → hodges and tuned topology)

◮ boosting:

  • 1. linear regression: hlin

p,q = a × num_cp + b

  • 2. neural network for hp,q − hlin

p,q

◮ other ensemble methods

(average different ML models, train on different subsets. . . )

◮ convert dataset

  • 1. find favourable representation
  • 2. apply any method

29 / 35

slide-41
SLIDE 41

Results (1)

Implementation and training

◮ sets: training (20%), test (80%) ◮ training time: few minutes

Accuracy:

◮ linear regression:

◮ orig.: h1,1 ≈ 61%, h2,1 ≈ 8.5% ◮ fav.: h1,1 ≈ 99.5%, h2,1 ≈ 4.5%

(note: regression on several scalars → h2,1 ≈ 12.5%)

◮ basic neural network (regression)

◮ orig.: h1,1 ≈ 68% (split: 30%), ≈ 78% (split: 80%) ◮ fav.: h1,1 ≈ 93%, h2,1 ≈ 16%

◮ boosting

◮ orig.: h1,1 ≈ 72%, h2,1 ≈ 15% ◮ fav.: h1,1 ≈ 99.5%, h2,1 ≈ 16% 30 / 35

slide-42
SLIDE 42

Results (2)

2 4 6 8 10 12 num_cp 0.0 2.5 5.0 7.5 10.0 12.5 15.0 17.5 Sizes 1 10 100 1000 Lines 0.00 + 1.00 x h11 h11_lin Lines 0.00 + 1.00 x 2 4 6 8 10 12 14 num_cp 0.0 2.5 5.0 7.5 10.0 12.5 15.0 17.5 Sizes 1 10 100 1000 Lines 0.00 + 1.00 x h11 h11_lin Lines 0.00 + 1.00 x

31 / 35

slide-43
SLIDE 43

Results (3)

2.5 5.0 7.5 10.0 12.5 15.0 17.5 h11 200 400 600 800 1000 1200 1400 true pred 20 30 40 50 60 70 h21 10−1 100 101 102 103 CY number

CICY3 Hodge number distribution (test set)

pred true

◮ Hodge numbers not exactly reproduced ◮ but distribution quite well learned

(ex.: within ±5% error, h2,1 is accurate more than 70%)

32 / 35

slide-44
SLIDE 44

Discussion

In progress: test different architectures (multi-inputs, multi-tasks. . . ) Possible extensions:

◮ neural network performs very badly on h2,1

→ challenge for ML community

◮ find a mapping original → favourable (GAN, cyclic GAN. . . ) ◮ representation learning: find better / invariant representation

(PCA, autoencoder. . . )

33 / 35

slide-45
SLIDE 45

Outline: 6. Conclusion

Motivations Machine learning Calabi–Yau 3-folds Data analysis ML analysis Conclusion

34 / 35

slide-46
SLIDE 46

Conclusion

◮ machine learning = extremely promising tool ◮ can help to learn how computer scientists / engineers work ◮ possible wide range of applications ◮ need to define clearly the (short- and long-term) objectives

35 / 35