Hi Everyone,
Supposedly while using CBoW, while saying happy and sad if both of them have same context words, then according to Word2Vec they will be nearer in vector space whereas they differ in meaning
That is a good question. I think your confusion lies in synonymy vs. similarity concepts.
You’re correct that Word2Vec is based on distributional hypothesis - the link between similarity is in how words are distributed (or in simple words, words that occur in similar contexts tend to have similar meanings). But note that similarity is not the same as synonymy.
Two words are synonymous if they can be substituted for one another in any sentence without changing the “truth” of the sentence. For example, “car” and “automobile” or “water” and “H2O” have the same propositional meaning - you can substitute these two words in most sentences without altering the “truth”.
Most words don’t have many synonyms, but words have many other types of similarities with other words. For example, “H2O” and “automobile” might be more common in scientific documents hence more “similar” in that regard. Or, “Cat” is not a synonym of “Dog”, but is more “similar” to “Dog” than to “Airplane”.
Loosely speaking, that is why we let the embeddings have as many dimensions as we think is necessary. Every word can be represented in this multi dimensional space so that the model could choose one of the dimensions to represent “happiness”, another to represent “adjectiveness” etc.
So the words “happy” and “sad” would be near in vector space on “adjectiveness” and probably other dimensions but would have opposite values on “happiness”. But overall - they would be pretty similar.
For example:
I feel so ______. I’m crying.
You could agree that word “happy” is more probable here than many other English words like “nutrition”, “PC” and others.