Complementary Learning of Word Embeddings

Complementary Learning of Word Embeddings

Yan Song, Shuming Shi

Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence
Main track. Pages 4368-4374. https://doi.org/10.24963/ijcai.2018/607

Continuous bag-of-words (CB) and skip-gram (SG) models are popular approaches to training word embeddings. Conventionally they are two standing-alone techniques used individually. However, with the same goal of building embeddings by leveraging surrounding words, they are in fact a pair of complementary tasks where the output of one model can be used as input of the other, and vice versa. In this paper, we propose complementary learning of word embeddings based on the CB and SG model. Specifically, one round of learning first integrates the predicted output of a SG model with existing context, then forms an enlarged context as input to the CB model. Final models are obtained through several rounds of parameter updating. Experimental results indicate that our approach can effectively improve the quality of initial embeddings, in terms of intrinsic and extrinsic evaluations.
Keywords:
Natural Language Processing: Natural Language Processing
Natural Language Processing: Embeddings