Deep Learning

LM, Language Model, Language Modeling, Conditional Probability, Statistical Language Model, n-gram

Naranjito 2021. 3. 9. 17:53
  • LM

Language Model, is a model assigns probability to sequence in order to modeling language, in other words, finding the most natural word sequence

 

  • Language Modeling

Prediction to unknown word from given words. 

 

  • Conditional Probability

It is the probability of an event occurring given that another event has already occurred. In this theory, mutually exclusive events are events that cannot occur simultaneously.

 

The probability of event A occurring given that event B has already occurred

 

- P(A|B) – the conditional probability; the probability of event A occurring given that event B has already occurred

- P(A ∩ B) – the joint probability of events A and B; the probability that both events A and B occur

- P(B) – the probability of event B

 

Pr(A|B)

- B : assume B occurred

- A : conditional probability to A occurs assums B occurred

 

Probability occurance w5 assumed w1 to w4 occurred

 

- P : Probability

- w : word

 

  • SLM

Statistical Language Model

- n-gram

It is a sequence of n words, for example,

Please turn your homework ...

a 2-gram (which we’ll call bigram) is a two-word sequence of words like “please turn”, “turn your”, or ”your homework”, and a 3-gram (a trigram) is a three-word sequence of words like “please turn your”, or “turn your homework”. We’ll see how to use n-gram models to estimate the probability of the last word of an n-gram given the previous words, and also to assign probabilities to entire sequences.