1 / 50

MDL and the complexity of natural language

MDL and the complexity of natural language. John Goldsmith University of Chicago/CNRS MoDyCo September 2006. Thanks. Carl de Marcken, Partha Niyogi, Antonio Galves, Yu Hu…. Naïve model of language.

brandi
Download Presentation

MDL and the complexity of natural language

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. MDL and the complexity of natural language John Goldsmith University of Chicago/CNRS MoDyCo September 2006

  2. Thanks • Carl de Marcken, Partha Niyogi, Antonio Galves, Yu Hu…

  3. Naïve model of language There exists an alphabet A = {a…z}, and a finite lexicon W  A*, where A* is the set of all concatenations of elements of A. There exist a (potentially unbounded) set of sentences of a language, L  W*. An utterance is a set (or string) of sentences, that is, an element of L*.

  4. Picture of naïve view

  5. “Naïve” view? The naïve view is still interesting – even if it is a great simplification. We can ask: if we embed the naïve view inside an MDL framework, do the results resemble known words (in English, Italian, etc.)? What if we apply it to DNA or protein sequences?

  6. Word segmentation Work by Michael Brent and by Carl de Marcken in the mid-1990s at MIT. A lexicon Lis a pair of objects (L, pL ): a set L A*, and a probability distribution pL that is defined on A* for which L is the support of pL. We call L the words. • We insist that A L: all individual letters are words. • We define a language as a subset of L*; its members are sentences. • Each sentence can be uniquely associated with an utterance (an element in A *) by a mapping F:

  7. F:L*A*

  8. F:L*A* S If F(S) = U then we say that S is a parse of U. U

  9. F:L*A* S U We pull back the measure from the space of letters to thespace of words.

  10. Different lexicons lead to different probabilities of the data Given an utterance U The probability of a string of letters is the probabilityassigned to its best parse.

  11. How do these two multigram models of English compare? Why is Number 2 better? A little example, to fix ideas Lexicon 2: {a,b,…s, t, th, u…z} Lexicon 1: {a,b,…s, t, u…z}

  12. A bit of notation Notation: [t] = count of t [h] = count of h [th] = count of th Z = total number of words (tokens) Log probability of corpus:

  13. Log probof sentence C All lettersare separate th is treatedas a separate chunk

  14. th is treatedas a separate chunk All lettersare separate This is positive if Lexicon 2 is better

  15. Effect of having fewer “words” altogether This is positive if Lexicon 2 is better

  16. Effect of frequency of /t/ and /h/ decreasing This is positive if Lexicon 2 is better

  17. Effect /th/ beingtreated as a unitrather than separate pieces This is positive if Lexicon 2 is better

  18. Description Length We need to account for the increase in length of the Lexicon, which is our model of the data. We add “th” to the lexicon: This is the generic form of the MDL criterion foradding a new word to the lexicon.

  19. Results • The Fulton County Grand Ju ry said Friday an investi gation of At l anta 's recent prim arye lectionproduc ed no e videnc e that any ir regul ar it i e s took place . • Thejury further s aid in term - end present ment sthatthe City Ex ecutiveCommit t e e ,which had over - all charg eofthee lection , d e serv e s the pra is e and than ksofthe City of At l antafortheman ner in whichthee lection was conduc ted. Chunks are too big Chunks are too small

  20. From Encarta: trained on the first 150 articles La funzione originari a dell'abbigliament o fu for s e quell a di pro t egger e il corpo dalle av vers i tà del c li ma . Ne i paesi cal di uomini e donn e indoss ano gonn ell in i mor bi di e drappeggi ati e per i z om i . In generale gli abit ant i delle zon ecal d e non port ano ma i più di due stra t i di vestit i. Al contr ari o, nei luog h i dove il c li ma è più rigid o sono diffus i abiti ader enti e a più stra ti . C omun e alle due tradizion i è tuttavi a l' abitudin e di ricor re re a mantell i per ri par arsi dagli e le ment i.

  21. What do we conclude? • From the point of view of linguistics, this does not teach us something about language (at least, not directly). • From the point of view of statistical learning, this does not teach us about statistical learning procedures.

  22. What do we conclude? What is most interesting about the results is that the linguist sees the errors committed by the system (by comparison with standard spelling, e.g.) as the result of a specification of a model set which fails to allow a method to capture the structure that linguistics has analyzed in language.

  23. We return to this… …in a moment. First, an observation the behavior of MDL in this process, so far.

  24. Usage of MDL? If description length of data D, given model M, is equal to the inverse log probability assigned to D by M + compressed length of M, then The process of word-learning is unambiguously one of increasing the probability of the data, and using the length of M as a stopping criterion.

  25. Discovering words from letters: Decrease compressed length of data, Use length of model as a stopping criterion. Linguistic cases we will see below: Decrease length of model, Use data compression improvement as a stopping criterion.

  26. Naïve model (assumed so far): Words are sequences of letters (1st order Markov model) Sentences are sequences of words (1st order Markov model) (1st order Markov: frequency is taken into consideration, but not context) Linguistics says: Sentences are best modeled with phrase-structure grammars; 2nd order or 3rd order Markov model is a poor substitute Words have internal structure, of two distinct sorts: syllable structure and phonological structure; we return to this shortly. Comparison

  27. Conjecture Suppose: the data we wish to account for is all of the textual data on the Internet in the world’s various languages, plus the alignment between corresponding sentences in the case of texts appearing in more than one language. We wish to find the minimal description of all of this data.

  28. Conjecture Conjecture (version 1): if we find the optimal compression, we will discover the traditional categories of linguistic analysis inside it (morphology, syntax, semantics, etc.). Conjecture (version 2): in order to approach this optimum in a tractable fashion with an automatic learning algorithm, we need to explicitly include categories of linguistic analysis.

  29. 3 major categories of failures of MDL word-discovery • Many failures of word-discovery are correct discovery of morphemes (word-pieces) investi-gation, pro-t-egger-e. • Many (thought fewer) failures of word-discovery are discovery of pairs of words that frequently appear together (for example, ofthe). • Many failures are too short to be likely words.

  30. Morphology Ask a linguist: it is the study of word-internal structure Ask a statistician: it is the extraction of certain aspects of redundancy in the vocabulary of a language. We describe a morphology analyzer (Linguistica) that learns morphology with no knowledge of the language.

  31. In order to shrink ||G||… There are about 26 different forms of each verb (parlo parli parla parliamo …parlando). Each letter takes 3~4 bits to encode; there are a total of 194 letters ~ 680 bits. parl- is 4 letters long; each letter takes 3~4 bits to encode; hence each appearance of parl requires 12~16 bits. Why repeat parl each time? Language allows a data structure at least this complex:

  32. We could shrink the morphology: Compared to a simple word list, we save 25 repetitions of parl (= 25*4.5 bits = 112.5 bits), minus the price p of the data structure represented by “___{ }”.

  33. Order of magnitude Using this data structure allows us to save roughly 112 bits out of 680 (16%). How much do we have “pay” in order to encode the data structure? We called this p…

  34. How much is p? • Notice that it’s not the cost of expressing those suffixes (that cost would have to be paid anyway): it’s the cost of expressing the notion “this stem may be followed be these suffixes”. • There are hundreds of verb stems in Italian that will use exactly the same data structure, because they accept exactly the same suffixes.

  35. More generally

  36. Finite state automaton (FSA)

  37. DL savings and costs Specification of the vocabulary of a lexicon of a language by a finite state automaton can lead to considerable savings in description length. 1. We must make explicit the cost of an FSA; 2. And the change in the compression of the original data.

  38. Cost of an FSA For each FSA, we “pay for” the information required to specifyeach state, each transition, and each label of each transition. [s] = Number of times a signatureis used in the data. Z= size of data. Size of pointer to first state of eachsignature =

  39. Initial approximation • We assume a morphology is a collection of 3 state FSAs, all sharing a unique final state. • Then the cost is the sum of the costs of the pointers to the first states, plus the cost of labeling the edges.

  40. Benefits of re-using labels for affixes There is considerable benefit to labeling the affixes not with strings, but with pointers to strings. The information cost of such a label more expensive if it is used only once, but if it isre-used a great deal, there is rapid gain to the MDL system: in short, the model demands generalizations in the grammar.

  41. How? Not all analyses are correct: But some are:

  42. The difference lies in the very low cost associated with creating and the relatively high cost associated with creating in which l and p are extremely rare (unique) suffixes: hence a pointer to each of them is very costly in bits.

  43. Most generalizations that are “discovered” are spurious (wrong) • It is hard to emphasize enough that the challenge is not finding generalizations—it is discarding 99% of the generalizations. • Consider the result of looking for patterns of this sort:

  44. Wrong: Wrong:

  45. Correct generalizations? Have the property that they interact with other generalizations across the grammar. This notion is largely formalized by the notion of reuse of morphemes, and the replacement of edge-labels by edge-pointers to morphemes.

  46. Generalizing beyond the data It is convenient to think of this object: as having a label “ed.ing” consisting of its suffixes (concatenated), and forming a subset of the lattice of all sets of suffixes. For example…

  47. Each node is an FSA;Each FSA is a node NULL.ed.ing.s Embed the nodes in the lattice generated by the set of suffixes.

  48. NULL.ation.ed.ing.s 3:776 NULL.ment.ed.ing.s 4:61 NULL.ed.ing.s 43:1110 ed.ing.s 2:7 NULL.ing.s 25:458 NULL.ed.s 46:564 NULL.ed.ing 38:508 NULL.s 442:4406 Notation: Suffix1.Suffix2 #stems: # occurrences

  49. NULL.ation.ed.ing.s 3:776 NULL.ment.ed.ing.s 4:61 [verbs] NULL.ed.ing.s 43:1110 ed.ing.s 2:7 NULL.ing.s 25:458 NULL.ed.s 46:564 NULL.ed.ing 38:508 [nouns] NULL.s 442:4406 Generalization consists of eliminating nodes, and push their stems upward

  50. Conclusion Linguistica Project: under way since 1997 at http://linguistica.uchicago.edu Developed to rapidly discover morphological structure in an increasingly large number of natural languages with no prior knowledge of the languages.

More Related