Word embedding for semantically relative words: an experimental study
Modelirovanie i analiz informacionnyh sistem, Tome 25 (2018) no. 6, pp. 726-733.

Voir la notice de l'article provenant de la source Math-Net.Ru

The ability to identify semantic relations between words has made a word2vec model widely used in NLP tasks. The idea of word2vec is based on a simple rule that a higher similarity can be reached if two words have a similar context. Each word can be represented as a vector, so the closest coordinates of vectors can be interpreted as similar words. It allows to establish semantic relations (synonymy, relations of hypernymy and hyponymy and other semantic relations) by applying an automatic extraction. The extraction of semantic relations by hand is considered as a time-consuming and biased task, requiring a large amount of time and some help of experts. Unfortunately, the word2vec model provides an associative list of words which does not consist of relative words only. In this paper, we show some additional criteria that may be applicable to solve this problem. Observations and experiments with well-known characteristics, such as word frequency, a position in an associative list, might be useful for improving results for the task of extraction of semantic relations for the Russian language by using word embedding. In the experiments, the word2vec model trained on the Flibusta and pairs from Wiktionary are used as examples with semantic relationships. Semantically related words are applicable to thesauri, ontologies and intelligent systems for natural language processing.
Keywords: word embedding, semantic relations, thesaurus, hyponymy
Mots-clés : word2vec, hypernymy, synonymy.
@article{MAIS_2018_25_6_a8,
     author = {M. S. Karyaeva and P. I. Braslavski and V. A. Sokolov},
     title = {Word embedding for semantically relative words: an experimental study},
     journal = {Modelirovanie i analiz informacionnyh sistem},
     pages = {726--733},
     publisher = {mathdoc},
     volume = {25},
     number = {6},
     year = {2018},
     language = {ru},
     url = {http://geodesic.mathdoc.fr/item/MAIS_2018_25_6_a8/}
}
TY  - JOUR
AU  - M. S. Karyaeva
AU  - P. I. Braslavski
AU  - V. A. Sokolov
TI  - Word embedding for semantically relative words: an experimental study
JO  - Modelirovanie i analiz informacionnyh sistem
PY  - 2018
SP  - 726
EP  - 733
VL  - 25
IS  - 6
PB  - mathdoc
UR  - http://geodesic.mathdoc.fr/item/MAIS_2018_25_6_a8/
LA  - ru
ID  - MAIS_2018_25_6_a8
ER  - 
%0 Journal Article
%A M. S. Karyaeva
%A P. I. Braslavski
%A V. A. Sokolov
%T Word embedding for semantically relative words: an experimental study
%J Modelirovanie i analiz informacionnyh sistem
%D 2018
%P 726-733
%V 25
%N 6
%I mathdoc
%U http://geodesic.mathdoc.fr/item/MAIS_2018_25_6_a8/
%G ru
%F MAIS_2018_25_6_a8
M. S. Karyaeva; P. I. Braslavski; V. A. Sokolov. Word embedding for semantically relative words: an experimental study. Modelirovanie i analiz informacionnyh sistem, Tome 25 (2018) no. 6, pp. 726-733. http://geodesic.mathdoc.fr/item/MAIS_2018_25_6_a8/

[1] Mikolov T., Yih W., Zweig G., “Linguistic Regularities in Continuous Space Word Representations”, HLT-NAACL 2013, 746–751

[2] Sienčnik S.K., “Adapting word2vec to named entity recognition”, Proceedings of the 20th nordic conference of computational linguistics, 2015, 239–243

[3] Lilleberg J., Zhu Y., Zhang Y., “Support vector machines and word2vec for text classification with semantic features”, Cognitive Informatics Cognitive Computing, IEEE 14th International Conference, 2015, 136–140

[4] Ling W. et al., “Two/too simple adaptations of word2vec for syntax problems”, Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2015, 1299–1304 | DOI

[5] Najafabadi M.M. et al., “Deep learning applications and challenges in big data analytics”, Journal of Big Data, 2 (2015), 1 | DOI

[6] Kutuzov A., Andreev I., Texts in, meaning out: neural language models in semantic similarity task for Russian, 2015, arXiv: 1504.08183 [cs.CL]

[7] Hearst M. A., “Automatic acquisition of hyponyms from large text corpora”, Proceedings of the 14th conference on Computational linguistics, v. 2, Association for Computational Linguistics, 1992, 539–545 | DOI

[8] Klaussner C., Zhekova D., “Lexico-syntactic patterns for automatic ontology building”, Proceedings of the Second Student Research Workshop associated with RANLP, 2011, 109–114

[9] Maedche A., Pekar V., Staab S., “Ontology learning part one–on discovering taxonomic relations from the web”, Web Intelligence, 2003, 301–319 | DOI

[10] Snow R., Jurafsky D., Ng A. Y., “Learning syntactic patterns for automatic hypernym discovery”, Advances in Neural Information Processing Systems, 2005, 1297–1304

[11] Panchenko A., et al., “Human and Machine Judgements for Russian Semantic Relatedness”, Analysis of Images, Social Networks and Texts, 5th International Conference, AIST 2016, Revised Selected Papers (Yekaterinburg, Russia, April 7–9, 2016), 2017, 221–235 | DOI

[12] Fu R., et al., “Learning semantic hierarchies via word embeddings”, Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics, v. 1, 2014, 1199–1209

[13] Ustalov D., Arefyev N., Biemann C., Panchenko A., Negative sampling improves hypernymy extraction based on projection learning, 2017, arXiv: 1707.03903 [cs.CL]

[14] Wang C., Cao L., Zhou B., Medical Synonym Extraction with Concept Space Models, 2015, arXiv: 1506.00528 [cs.CL]

[15] Rei M., Briscoe T., “Looking for hyponyms in vector space”, Proceedings of the Eighteenth Conference on Computational Natural Language Learning, 2014, 68–77 | DOI

[16] Turney P., Pantel P., “From frequency to meaning: Vector space models of semantics”, Journal of artificial intelligence research, 37 (2010), 141–188 | DOI | MR | Zbl

[17] Matsuo Y., Ishizuka M., “Keyword extraction from a single document using word co-occurrence statistical information”, International Journal on Artificial Intelligence Tools, 13:1 (2004), 157–169 | DOI