Deep Learning 61

LSA, SVD, Orthogonal matrix, Transposed matrix, Identity matrix, Inverse matrix, Diagonal matrix, Truncated SVD

LSALatent Semantic Analysis, substitute for DTM, TF-IDF(2021.03.10 - [Deep Learning] - BoW, CountVectorizer, fit_transform, vocabulary_, DTM, TDM, TF-IDF, TfidfVectorizer, isnull, fillna, pd.Series) which has not consider meaning of terms. It applies SVD based on DTM, TF-IDF and reduce dimensions, eliciting potential meaning of words.1. SVD Singular Value Decomposition, it refers to the decompos..

Deep Learning 2021.03.11

LM, Language Model, Language Modeling, Conditional Probability, Statistical Language Model, n-gram

LM Language Model, is a model assigns probability to sequence in order to modeling language, in other words, finding the most natural word sequence Language Modeling Prediction to unknown word from given words. Conditional Probability It is the probability of an event occurring given that another event has already occurred. In this theory, mutually exclusive events are events that cannot occur s..

Deep Learning 2021.03.09

normalization, WordNetLemmatizer, PorterStemmer, LancasterStemmer, Storword

normalization Integrate different words to make them the same word-such as US is same as USA integrate them as US. 1. WordNetLemmatizer If words have different forms, find the root word-such as the root of 'am, are, is' is 'be'. from nltk.stem import WordNetLemmatizer lemmatizer=WordNetLemmatizer() words=[ 'have', 'going', 'loves', 'lives', 'flies', 'dies', 'watched', 'has', 'starting'] print('b..

Deep Learning 2021.03.05