I am implementing a paper to compare our performance. In the paper, the uathor says
300-dimensional pre-trained word2vec vectors (Mikolov et al., 2013)
I am wondering whether the pretrained word2vec Gensim model here is same as the pretrained embeddings on the official Google site (the GoogleNews-vectors-negative300.bin.gz file)
My source of doubt arises from this line in Gensim documentation (in Word2Vec Demo section)
We will fetch the Word2Vec model trained on part of the Google News dataset, covering approximately 3 million words and phrases
Does this mean the model on gensim is not fully trained? Is it different from the official embeddings by Mikolov?