I believe I understand the concepts of CBOW and skip-gram. But I'm a little bit stuck. I kind of don't understand this . In fact I understand it so poorly that I can't even formulate a question around it.
Now what do we do?
Edit: An attempt at formulating a question: is it the process of feeding the model with the [context][context][output] vector that you are depicting?
I'll be honest, I personally found this figure puzzling. Still not 100% clear on it, but I don't believe it refers to the negative sampling approach. My best guess is that it's referring to earlier word2vec variants where the input in skipgram (or sum of inputs in CBOW) are multiplied by a weights matrix that projects the input to an output vector.