Synergies in learning words and their referents
Mark Johnson1, Katherine Demuth1, Michael Frank2 and Bevan Jones3
1Macquarie University 2Stanford University 3University of Edinburgh
NIPS 2010
1/15
Synergies in learning words and their referents Mark Johnson 1 , - - PowerPoint PPT Presentation
Synergies in learning words and their referents Mark Johnson 1 , Katherine Demuth 1 , Michael Frank 2 and Bevan Jones 3 1 Macquarie University 2 Stanford University 3 University of Edinburgh NIPS 2010 1/15 Two hypotheses about language
1Macquarie University 2Stanford University 3University of Edinburgh
1/15
I “Semantic bootstrapping”: semantics is learnt rst, and used to predict
I “Syntactic bootstrapping”: syntax is learnt rst, and used to predict
I Conventional view of lexical acquisition, e.g., Kuhl (2004)
I corresponds to joint inference for all components of language I stages in language acquisition might be due to:
2/15
I partial knowledge of component A provides information about
I partial knowledge of component B provides information about
3/15
I word sequence: Is that the pig? I objects in nonlinguistic context: ,
I identify utterance topic: I identify word-topic mapping: pig → 4/15
5/15
I child-directed utterance: Is that the pig? I broad phonemic representation: ɪz ðæt ðə pɪg I input to learner:
ɪ △ z △ ð △ æ △ t △ ð △ ə △ p △ ɪ △ g
6/15
I AGs learn probability of entire
I AGs are hierarchical Dirichlet or
I Prob. of adapted subtree ∝
7/15
8/15
I equivalent to Jones et al (2010)
9/15
I at most one topical word per sentence, or I at most one topical word per topical collocation 10/15
I Child-directed speech corpus collected by Fernald et al (1993) I Objects in visual context annotated by Frank et al (2009)
I Uniform prior on PYP a parameter I “Sparse” Gamma(100, 0.01) on PYP b parameter
I collected word segmentation and topic assignments at every 10th
I computed and evaluated the modal (i.e., most frequent) sample
11/15
I consistent with results from Jones et al (2010)
I most gain with one topical word per topical collocation
12/15
13/15
14/15
I generic AG inference code makes it easy to explore models
I incorporating word-topic mapping improves segmentation accuracy (at
I improving segmentation accuracy improves topic detection and acquisition
I extend expressive power of AGs (e.g., phonology, syntax) I richer data (e.g., more non-linguistic context) I more realistic data (e.g., phonological variation) 15/15