No edit summary |
|||
Line 6: | Line 6: | ||
Made by Google, uses Neural Net, performs good on semantics. | Made by Google, uses Neural Net, performs good on semantics. | ||
=== Installation + getting started: === | === Installation + getting started: === | ||
pip install gensim | <code>pip install gensim</code> | ||
==Fastword== | ==Fastword== |
Revision as of 09:55, 8 May 2017
General Information on word embeddings
For a general explanation look here: [1]
Word2vec
Made by Google, uses Neural Net, performs good on semantics.
Installation + getting started:
pip install gensim
Fastword
Made by Facebbok based on word2vec. Better at capturing syntactic relations (like apparent ---> apparently) see here: [2] Pretrained model files are HUGE
GloVe
pre trained models
- https://github.com/Kyubyong/wordvectors: Word2Vec and FastText, Multiple languages, no english, trained on Wikipedia
- https://github.com/3Top/word2vec-api Mostly GloVe, some word2vec, English, Trained on News, Wikipedia, Twitter
- https://github.com/facebookresearch/fastText/blob/master/pretrained-vectors.md: Fasttext, all imaginable languages, trained on Wikipedia