wego is the implementations from scratch for word embeddings (a.k.a word representation) models in Go.
Word embeddings make words' meaning, structure, and concept mapping into vector space with a low dimension. For representative instance:
Vector("King") - Vector("Man") + Vector("Woman") = Vector("Queen")
Like this example, the models generate word vectors that could calculate word meaning by arithmetic operations for other vectors.
The following models to capture the word vectors are supported in wego:
-
Word2Vec: Distributed Representations of Words and Phrases and their Compositionality [pdf]
-
GloVe: Global Vectors for Word Representation [pdf]
-
LexVec: Matrix Factorization using Window Sampling and Negative Sampling for Improved Word Representations [pdf]
Also, wego provides nearest neighbor search tools that calculate the distances between word vectors and find the nearest words for the target word. "near" for word vectors means "similar" for words.
Please see the Usage section if you want to know how to use these for more details.
Inspired by Data Science in Go @chewxy
Use go
command to get this pkg.
$ go get -u github.com/ynqa/wego
$ bin/wego -h
wego provides CLI and Go SDK for word embeddings.
Usage:
wego [flags]
wego [command]
Available Commands:
console Console to investigate word vectors
glove GloVe: Global Vectors for Word Representation
help Help about any command
lexvec Lexvec: Matrix Factorization using Window Sampling and Negative Sampling for Improved Word Representations
query Query similar words
word2vec Word2Vec: Continuous Bag-of-Words and Skip-gram model
word2vec
, glove
and lexvec
executes the workflow to generate word vectors:
- Build a dictionary for vocabularies and count word frequencies by scanning a given corpus.
- Start training. The execution time depends on the size of the corpus, the hyperparameters (flags), and so on.
- Save the words and their vectors as a text file.
query
and console
are the commands which are related to nearest neighbor searching for the trained word vectors.
query
outputs similar words against a given word using sing word vectors which are generated by the above models.
e.g. wego query -i word_vector.txt microsoft
:
RANK | WORD | SIMILARITY
-------+-----------+-------------
1 | hypercard | 0.791492
2 | xp | 0.768939
3 | software | 0.763369
4 | freebsd | 0.761084
5 | unix | 0.749563
6 | linux | 0.747327
7 | ibm | 0.742115
8 | windows | 0.731136
9 | desktop | 0.715790
10 | linspire | 0.711171
wego does not reproduce word vectors between each trial because it adopts HogWild! algorithm which updates the parameters (in this case word vector) async.
console
is for REPL mode to calculate the basic arithmetic operations (+
and -
) for word vectors.
It can define the hyper parameters for models by functional options.
model, err := word2vec.New(
word2vec.Window(5),
word2vec.Model(word2vec.Cbow),
word2vec.Optimizer(word2vec.NegativeSampling),
word2vec.NegativeSampleSize(5),
word2vec.Verbose(),
)
The models have some methods:
type Model interface {
Train(io.ReadSeeker) error
Save(io.Writer, vector.Type) error
WordVector(vector.Type) *matrix.Matrix
}
As training word vectors wego requires the following file formats for inputs/outputs.
Input corpus must be subject to the formats to be divided by space between words like text8.
word1 word2 word3 ...
After training wego save the word vectors into a txt file with the following format (N
is the dimension for word vectors you given):
<word> <value_1> <value_2> ... <value_N>