This is an implementation of the LexVec word embedding model (similar to word2vec and GloVe) that achieves state of the art results in multiple NLP tasks, as described in these papers.
Subword LexVec (paper)
- Common Crawl - 58B tokens, cased - 2,000,000 words - 300 dimensions
-
Common Crawl - 58B tokens, lowercased - 2,000,000 words - 300 dimensions
-
English Wikipedia 2015 + NewsCrawl - 7B tokens - 368,999 words - 300 dimensions
Model | GSem | GSyn | MSR | RW | SimLex | SCWS | WS-Sim | WS-Rel | MEN | MTurk |
---|---|---|---|---|---|---|---|---|---|---|
LexVec | 72.6% | 73.8% | 73.2% | .539 | .477 | .687 | .809 | .696 | .814 | .717 |
fastText | 75.0% | 72.1% | 71.8% | .522 | .424 | .673 | .810 | .724 | .805 | .717 |
-
Both models use vectors with 300 dimensions.
-
Both models use character n-grams of length 3-6 as subwords.
-
All tasks are evaluated using cased words (
"Toronto" != "toronto"
). -
GSem, GSyn, and MSR analogies were solved using 3CosMul.
-
Both models were trained using this release of Common Crawl which contains 58B tokens, restricting the vocabulary to the 2 million most frequent cased words.
-
Subword LexVec was trained using the following command:
$ OUTPUT=output scripts/em_lexvec.sh -corpus common_crawl_cased.txt -negative 3 -dim 300 -subsample 1e-5 -minfreq 0 -window 2 -minn 3 -maxn 6
-
fastText was trained using the following command:
$ ./fasttext skipgram -input common_crawl_cased.txt -minCount 0 -t 1e-5 -dim 300 -lr 0.025 -minn 3 -maxn 6
Model | GSem | GSyn | MSR | RW | SimLex | SCWS | WS-Sim | WS-Rel | MEN | MTurk |
---|---|---|---|---|---|---|---|---|---|---|
LexVec, Word | 81.1% | 68.7% | 63.7% | .489 | .384 | .652 | .727 | .619 | .759 | .655 |
LexVec, Word + Context | 79.3% | 62.6% | 56.4% | .476 | .362 | .629 | .734 | .663 | .772 | .649 |
word2vec Skip-gram | 78.5% | 66.1% | 56.0% | .471 | .347 | .649 | .774 | .647 | .759 | .687 |
-
All three models were trained using the same English Wikipedia 2015 + NewsCrawl corpus.
-
GSem, GSyn, and MSR analogies were solved using 3CosMul.
-
LexVec was trained using the default parameters, expanded here for comparison:
$ OUTPUT=output scripts/im_lexvec.sh -corpus enwiki+newscrawl.txt -dim 300 -window 2 -subsample 1e-5 -negative 5 -iterations 5 -minfreq 100 -model 0 -minn 0
-
word2vec Skip-gram was trained using:
$ ./word2vec -train enwiki+newscrawl.txt -output sgnsvectors -size 300 -window 10 \ -sample 1e-5 -negative 5 -hs 0 -binary 0 -cbow 0 -iter 5 -min-count 100
Model | GSem | GSyn | MSR | RW | SimLex | SCWS | WS-Sim | WS-Rel | MEN | MTurk |
---|---|---|---|---|---|---|---|---|---|---|
LexVec, Word | 76.4% | 71.3% | 70.6% | .508 | .444 | .667 | .762 | .668 | .802 | .716 |
LexVec, Word + Context | 80.4% | 66.6% | 65.1% | .496 | .419 | .644 | .775 | .702 | .813 | .712 |
word2vec | 73.3% | 75.1% | 75.1% | .515 | .436 | .655 | .741 | .610 | .699 | .680 |
GloVe | 81.8% | 72.4% | 74.3% | .384 | .374 | .540 | .698 | .571 | .743 | .645 |
-
All models use vectors with 300 dimensions.
-
GSem, GSyn, and MSR analogies were solved using 3CosMul.
-
LexVec was trained using this release of Common Crawl which contains 58B tokens, restricting the vocabulary to the 2 million most frequent words, using the following command:
$ OUTPUT=output scripts/em_lexvec.sh -corpus common_crawl.txt -negative 3 -dim 300 -subsample 1e-5 -minfreq 0 -window 2 -minn 0
-
The pre-trained word2vec vectors were trained using the unreleased Google News corpus containing 100B tokens, restricting the vocabulary to the 3 million most frequent words.
-
The pre-trained GloVe vectors were trained using Common Crawl (release unknown) containing 42B tokens, restricting the vocabulary to the 1.9 million most frequent words.
-
Install the Go compiler and clang.
-
Make sure your
$GOPATH
is set -
Execute the following commands in your terminal:
$ go get github.com/alexandres/lexvec $ cd $GOPATH/src/github.com/alexandres/lexvec $ make
To get started, run $ scripts/demo.sh
which trains a model using the small text8 corpus (100MB from Wikipedia).
Basic usage of LexVec is:
$ OUTPUT=dirwheretostorevectors scripts/im_lexvec.sh -corpus somecorpus
Run $ ./lexvec -h
for a full list of options.
By default, LexVec stores the sparse matrix being factorized in-memory. This can be a problem if your training corpus is large and your system memory limited. We suggest you first try using the in-memory implementation. If you run into Out-Of-Memory issues, use the External Memory variant with the -memory
option specifying how many GBs of memory to use for the sort buffer.
$ OUTPUT=dirwheretostorevectors scripts/em_lexvec.sh -corpus somecorpus -memory 4. ...exactsameoptionsasinmemory
Subword information is controlled by the options -minn
, -maxn
, and -subword
.
-
To disable the use of subword information, specify
-minn 0
. -
To use character n-grams of length 3-6, specify
-minn 3 -maxn 6
(this is the default configuration). -
To provide your own subword information (such as morphological segmentation), specify
-minn 0 -subword subwords.txt
, where the subwords file contains one line for each vocabulary word (vocabulary must match that of-vocab
), each line containing a word followed by each of its subwords, separated by spaces.
By default, the binary model used for computing OOV word vectors is saved to $OUTPUT/model.bin
. Set -outputsub ""
to disable saving this model.
Use the binary model to compute vector for OOV words:
-
Using the Go executable by providing one word per line on
stdin
:$ echo "marvelicious" | ./lexvec embed -outputsub pathtomodel.bin
-
Using the Python lib:
import lexvec model = lexvec.Model('pathtomodel.bin') vector = model.word_rep('marvelicious')
Note: You can also use these commands to get vectors for in-vocabulary words as the binary model stores the vocabulary used for training.
Alexandre Salle, Marco Idiart, and Aline Villavicencio. "Matrix Factorization using Window Sampling and Negative Sampling for Improved Word Representations." ACL (2016). (pdf)
Alexandre Salle, Marco Idiart, and Aline Villavicencio. "Enhancing the LexVec Distributed Word Representation Model Using Positional Contexts and External Memory." arXiv preprint arXiv:1606.01283 (2016). (pdf)
Alexandre Salle and Aline Villavicencio. "Incorporating Subword Information into Matrix Factorization Word Embeddings." Second Workshop on Subword and Character LEvel Models in NLP (2018). (pdf)
Alexandre Salle and Aline Villavicencio. "Why So Down? The Role of Negative (and Positive) Pointwise Mutual Information in Distributional Semantics." arXiv preprint arXiv:1908.06941 (2019). (pdf)
Copyright (c) 2016-2018 Salle, Alexandre [email protected]. All work in this package is distributed under the MIT License.