JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE) ›› 2019, Vol. 54 ›› Issue (7): 100-105.doi: 10.6040/j.issn.1671-9352.1.2018.104

Previous Articles    

A global word vector model based on pointwise mutual information

LI Wan-li1, TANG Jing-yao1, XUE Yun1,2*, HU Xiao-hui1, ZHANG Tao3   

  1. 1. School of Physics and Telecommunication Engineering, South China Normal University, Guangzhou 510006, Guangdong, China;
    2. Guangdong Provincial Engineering Technology Research Center for Data Science, Guangzhou 510006, Guangdong, China;
    3. Guangdong CON-COM Technology CO., LTD, Guangzhou 510640, Guangdong, China
  • Published:2019-06-27

Abstract: A global word vector training model based on pointwise mutual information was presented. The model used another correlation information, the ratio of the joint probability and the product of the marginal probability, to depict the relationship between words and avoid the shortcoming of conditional probability. In order to verify the validity of our model, we trained word embedding by GloVe and our model in the same situation and then carried out experiments with word analogy and word similarity separately using these two word embeddings. Experiments showed our model has achieved 10.50%, 4.43%, 1.02% better accuracy rate than the GloVe model does in sematic experiments of word analogy at 100 dimensionality, 200 dimensionality, 300 dimensionality respectively. Accuracy rate has also gained 5%-6% rise in word similarity experiments. The results show that our model can capture semantics of words more effectively.

Key words: pointwise mutual information, word vector, GloVe

CLC Number: 

  • TP391
[1] HINTON G E. Learning distributed representations of concepts[C] //Proceedings of the Eighth Annual Conference of the Cognitive Science Society. Amherst, Mass: Erlbaum, 1986:1-12.
[2] BENGIO Y, DUCHARME R, VINCENT P, et al. A neural probabilistic language model[J]. Journal of Machine Learning Research, 2003, 3: 1137-1155.
[3] MIKOLOV T, CHEN K, CORRADO G, et al. Efficient estimation of word representations in vector space[J]. arXiv,2013, arXiv:1301.3781.
[4] PENNINGTON J, SOCHER R, MANNING C. Glove: global vectors for word representation[C] //Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing(EMNLP). USA: ACL, 2014: 1532-1543.
[5] DEERWESTER S, DUMAIS S T, FURNAS G W, et al. Indexing by latent semantic analysis[J]. Journal of the American Society for Information Science, 1990, 41(6): 391-407.
[6] LEVY O, GOLDBERG Y. Neural word embedding as implicit matrix factorization[C] //Advances in Neural Information Processing Systems. USA: Curran Associates, Inc, 2014: 2177-2185.
[7] LEVY O, GOLDBERG Y, DAGAN I. Improving distributional similarity with lessons learned from word embeddings[J]. Transactions of the Association for Computational Linguistics, 2015, 3(1): 211-225.
[8] JAMEEL S, BOURAOUI Z, SCHOCKAERT S. Unsupervised learning of distributional relation vectors[C] //Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics(Long Papers).Melbourne: ACL, 2018:23-33.
[9] ARORA S, LI Y, LIANG Y, et al. A latent variable model approach to pmi-based word embeddings[J]. Transactions of the Association for Computational Linguistics, 2016, 4(1): 385-399.
[10] MORIN F, BENGIO Y. Hierarchical probabilistic neural network language model[C] //Proceedings of the 10th International Workshop on Artificial Intelligence and Statistics. Aistats. The Society for Artificial Intelligence and Statistics, 2005, 246-252.
[11] HARRIS Z S. Distributional structure[J]. Word, 1954, 10(2/3): 146-162.
[1] Yang XU,Jian-zhong SUN,Lei HUANG,Xiao-yao XIE. Trajectory model of area crowd based on WiFi positioning [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2019, 54(5): 8-20.
[2] Zhe-jin DONG,Jian WANG,Ling-fei QIAN,Hong-fei LIN. A modeling method of user growth profile [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2019, 54(3): 38-45.
[3] Xue-mei WANG,Xing-shu CHEN,Hai-zhou WANG,Wen-xian WANG. Automatic extraction of key information for news web pages based on tag and block features [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2019, 54(3): 67-74.
[4] Heng-ze BAO,Dong ZHOU,Tan WU. Tag recommendation with multi-source heterogeneous networked information [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2019, 54(3): 56-66.
[5] Jie WU,Xiao-fei ZHU,Yi-hao ZHANG,Jian-wu LONG,Xian-ying HUANG,Wu YANG. User sentiment tendency aware based Micro-blog sentiment analysis method [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2019, 54(3): 46-55.
[6] . Design and implementation of topic detection in Russian news based on ontology [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2018, 53(9): 49-54.
[7] YU Chuan-ming, ZUO Yu-heng, GUO Ya-jing, AN Lu. Dynamic discovery of authors research interest based on the combined topic evolutional model [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2018, 53(9): 23-34.
[8] . Reader emotion classification with news and comments [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2018, 53(9): 35-39.
[9] GONG Shuang-shuang, CHEN Yu-feng, XU Jin-an, ZHANG Yu-jie. Extraction of Chinese multiword expressions based on Web text [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2018, 53(9): 40-48.
[10] ZHANG Jun, LI Jing-fei, ZHANG Rui, RUAN Xing-mao, ZHANG Shuo. Community detection algorithm based on effective resistance of network [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2018, 53(3): 24-29.
[11] YU Chuan-ming, FENG Bo-lin, TIAN Xin, AN Lu. Deep representative learning based sentiment analysis in the cross-lingual environment [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2018, 53(3): 13-23.
[12] PANG Bo, LIU Yuan-chao. Fusion of pointwise and deep learning methods for passage ranking [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2018, 53(3): 30-35.
[13] LIAO Xiang-wen, ZHANG Ling-ying, WEI Jing-jing, GUI Lin, CHENG Xue-qi, CHEN Guo-long. User influence analysis of social media with temporal characteristics [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2018, 53(3): 1-12.
[14] CHEN Xin, XUE Yun, LU Xin, LI Wan-li, ZHAO Hong-ya, HU Xiao-hui. Text feature extraction method for sentiment analysis based on order-preserving submatrix and frequent sequential pattern mining [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2018, 53(3): 36-45.
[15] WANG Tong, MA Yan-zhou, YI Mian-zhu. Speech recognition of Russian short instructions based on DTW [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2017, 52(11): 29-36.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!