Explanation:
The denominator of the softmax requires summing over all vocabulary words. If vocabulary size is 1 million, 1 million dot products must be computed for every update, making training computationally expensive.
Full softmax requires computing the denominator over the entire vocabulary: Σ exp(vwT vwc) Time complexity = O(|V|) per training example.
Explanation:
If vectors are orthogonal, their dot product is zero. Since sigmoid(0) = 0.5, the gradient is small but not zero. The model still updates the vectors to push negative samples away.
Explanation:
When frequency is very high, t/f(w) becomes very small, making the discard probability approach 1. Thus very frequent words like "the" are removed most of the time.
Explanation:
The input matrix W represents center-word embeddings and captures semantic structure. The output matrix W' represents context embeddings and is usually discarded after training.
Explanation:
According to the distributional hypothesis, words appearing in similar contexts obtain similar embeddings, resulting in high cosine similarity.
Explanation:
Skip-gram generates more training signals per word and performs better for rare words, while CBOW is generally faster and smoother for frequent words.
Explanation:
Raising frequencies to the power 3/4 reduces dominance of very frequent words and increases medium-frequency sampling, improving embedding quality.
Explanation:
Word2Vec embeddings capture linear semantic relationships, allowing vector arithmetic to represent analogies like gender direction in embedding space.
Explanation:
Full softmax requires computation across entire vocabulary for each update, making complexity proportional to vocabulary size and thus very slow.
Explanation:
Classic Word2Vec learns a single static vector representation for each word type, regardless of context. Therefore, polysemous words like “bank” (river bank vs financial bank) receive only one embedding and cannot represent different meanings based on context.
No comments:
Post a Comment