Evaluation of Stacked Embeddings for Arabic Word Sense Disambiguation

Rim Laatar, Chafik Aloulou, Lamia Hadrich Belguith

Abstract


Word Sense Disambiguation (WSD) aims to determine the correct meaning of words that can have multiple interpretations. Recently, contextualized word embeddings, whose goal is to give different representations of the same word in diverse contexts, have been shown to have a tremendous impact on several natural language processing tasks including question answering, semantic analysis and even word sense disambiguation. This paper reports on experiments with different stacks of word embeddings and evaluation of their usefulness for Arabic word sense disambiguation. Word embeddings stay in the core of the development of NLP, with multiple key language models being created over the last two years like FastText, ElMo, BERT and Flair.\newlineIt's worth pointing out that the Arabic language can be divided into three major historical periods: old Arabic, middle-age Arabic, and Modern Arabic. Actually, contemporary Arabic has proved to be the greatest concern of many researchers. The main gist of our work is to disambiguate Arabic words according to the historical period in which they appeared. To perform the WSD task, we suggest a method that deploys stacked embeddings models. The experimental evaluation demonstrates that stacked embeddings outperforms the previously proposed methods for Arabic WSD.

Keywords


Arabic language, historical dictionary, modern standard arabic, old arabic, middle age arabic, contextualized embeddings, stacked embeddings

Full Text: PDF