Introduction
- Term Frequency - Inverse Document Frequency
- t-distributed Stochastic Neighbor Embedding
- Article Spinning
- Text classification -- sentence embedding average embeddings of all words + naive classifier/logistic regression
- Bigram model -- probability of transition occurence / softmax logistic regression / 1 hidden layer neural network
- Word2Vec Embedding -- Continuous Bag Of Words (Many to one), Skip-Gram (One to many), Hierachical Softmax (like Huffman Code), Negative Sampling (like Triplet Loss)
- GloVe Embedding -- like Recommendation System
- POS NER -- HMM, DNN with Embedding and tree embedded RNN
- Bidirectional RNN -- Text generation
- Seq2seq -- Encoder + Decoder Structure + Teacher Forcing (feedback right word even if the network has wrong previous word prediction)
- Attention -- Encoder + Context + Decoder
- Memory Networks -- train to make question embedding close to answer embedding + 2 Supporting Fact (pass first fact first, next next)
References