1 / 9

Stat. inference: n-gram models over sparce data

Stat. inference: n-gram models over sparce data. Stat nlp function. Taking some data(generated in accordance with some unknown probability distribution) and then making some inferences about this distribution. Ex

Télécharger la présentation

Stat. inference: n-gram models over sparce data

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Stat. inference: n-gram models over sparce data

  2. Stat nlp function • Taking some data(generated in accordance with some unknown probability distribution) and then making some inferences about this distribution. Ex • We might look at lots of prepositional phrase attachments in a corpus and use them to try to predict prepositional phrase attachments for English in general.

  3. We will examine the classic task of language modelling (aka Shannon game) where the problem is to predict the next word given the previous words. • Importance: • Speech or optical recognition, SMT, spelling correction, and handwriting recognition.

  4. Uses • Word sense disambiguation • Probabilistic parsing

  5. Building n-gram models

  6. http://svr-www.eng.cam.ac.uk/~prc14/toolkit.html • Preprocess the corpus using ASCII files • Check this • http://books.google.com/ngrams

  7. MLE • Come across • 10 times of come across • 8 of which were followed by as • Once by more and once by a • PMLE (wn|w1 ….. Wn-1) = C(w1……….wn • ________________ • C(w1……. W-1

More Related