Implementing Deep Learning Methods and Feature Engineering for Text Data: The Skip-gram Model
Editor's note: This post is only one part of a far more thorough and in-depth original, found here, which covers much more than what is included here. The Skip-gram model architecture usually tries to achieve the reverse of what the CBOW model does. It tries to predict the source context words (surrounding words) given a target word (the center word). If we used the CBOW model, we get pairs of (context_window, target_word)where if we consider a context window of size 2, we have examples like ([quick, fox], brown), ([the, brown], quick), ([the, dog], lazy) and so on. Now considering that the skip-gram model's aim is to predict the context from the target word, the model typically inverts the contexts and targets, and tries to predict each context word from its target word.
Apr-10-2018, 17:40:45 GMT
- Technology: