4.3.3. 技术¶
4.3.3.1. 文字嵌入¶
Text embeddings allow deep learning to be effective on smaller datasets. These are often first inputs to a deep learning archiectures and most popular way of transfer learning in NLP. Embeddings are simply vectors or a more generically, real valued representations of strings. Word embeddings are considered a great starting point for most deep NLP tasks.
The most popular names in word embeddings are word2vec by Google (Mikolov) and GloVe by Stanford (Pennington, Socher and Manning). fastText seems to be a fairly popular for multi-lingual sub-word embeddings.
4.3.3.1.1. 单词嵌入¶
E m b e d d i n g |
Paper |
O r g a n i s a t i o n |
g e n s i m - T r a i n i n g S u p p o r t |
Blogs |
---|---|---|---|---|
w o r d 2 v e c |
Official Implementation, T.Mikolove et al. 2013. Distributed Representations of Words and Phrases and their Compositionality. pdf |
G o o g l e |
Y e s : h e a v y _ c h e c k _ m a r k : |
Visual explanations by colah at Deep Learning, NLP, and Representations; gensim’s Making Sense of word2vec |
G l o V e |
Jeffrey Pennington, Richard Socher, and Christopher D. Manning. 2014. GloVe: Global Vectors for Word Representation. pdf |
S t a n f o r d |
N o : n e g a t i v e _ s q u a r e d _ c r o s s _ m a r k : |
Morning Paper on GloVe by acoyler |
f a s t T e x t |
Official Implementation, T. Mikolov et al. 2017. Enriching Word Vectors with Subword Information. pdf |
F a c e b o o k |
Y e s : h e a v y _ c h e c k _ m a r k : |
Notes for Beginners:
Thumb Rule: fastText >> GloVe > word2vec
You can find pre-trained fasttext Vectors in several languages
If you are interested in the logic and intuition behind word2vec and GloVe: The Amazing Power of Word Vectors and introduce the topics well
arXiv: Bag of Tricks for Efficient Text Classification, and arXiv: FastText.zip: Compressing text classification models were released as part of fasttext
4.3.3.1.2. 基于句子和语言模型的词嵌入¶
ElMo from Deep Contextualized Word Represenations - PyTorch implmentation - TF Implementation
ULimFit aka Universal Language Model Fine-tuning for Text Classification by Jeremy Howard and Sebastian Ruder
InferSent from Supervised Learning of Universal Sentence Representations from Natural Language Inference Data by facebook
CoVe from Learned in Translation: Contextualized Word Vectors
Pargraph vectors from Distributed Representations of Sentences and Documents. See doc2vec tutorial at gensim
sense2vec - on word sense disambiguation
Skip Thought Vectors - word representation method
Adaptive skip-gram - similar approach, with adaptive properties
Sequence to Sequence Learning - word vectors for machine translation
4.3.3.2. 问答和知识提取¶
DrQA: Open Domain Question Answering by facebook on Wikipedia data
DocQA: Simple and Effective Multi-Paragraph Reading Comprehension by AllenAI
Markov Logic Networks for Natural Language Question Answering
Relation extraction with matrix factorization and universal schemas
Privee: An Architecture for Automatically Analyzing Web Privacy Policies
Teaching Machines to Read and Comprehend - DeepMind paper
Relation Extraction with Matrix Factorization and Universal Schemas
Towards a Formal Distributional Semantics: Simulating Logical Calculi with Tensors