Automated Amharic News Categorization Using Deep Learning Models

For decades, machine learning techniques have been used to process Amharic texts. The potential application of deep learning on Amharic document classification has not been exploited due to a lack of language resources. In this paper, we present a deep learning model for Amharic news document classification. The proposed model uses fastText to generate text vectors to represent semantic meaning of texts and solve the problem of traditional methods. The text vectors matrix is then fed into the embedding layer of a convolutional neural network (CNN), which automatically extracts features. We conduct experiments on a data set with six news categories, and our approach produced a classification accuracy of 93.79%. We compared our method to well-known machine learning algorithms such as support vector machine (SVM), multilayer perceptron (MLP), decision tree (DT), XGBoost (XGB), and random forest (RF) and achieved good results.


Introduction
Text categorization is the process of predicting the domain area of a given document [1]. With increasing in volume of text in today's computing world, text classification is highly required. Text classification can be performed either manually or automatically. Manual classification of documents requires a large workforce and wastes money. Furthermore, automatic document classification plays an important role in a variety of applications, including e-mail spam filtering, electronic document indexing, and online tendering. As a result, automatic text categorization is essential for dynamic management of information on Internet.
Amharic is a Semitic language that is also known as "Amarigna" and is one of Ethiopia's most widely spoken languages. It is spoken by up to 35 million Ethiopians, accounting for roughly one-third of the country's population [2]. e language's history can be traced back to the first millennium BC when refugees from southwestern Arabia crossed the Red Sea into modern-day Eritrea and mingled with the Cushitic people. e number of Amharic electronic documents is growing at the moment. Due to this, automatic Amharic text classification is highly required.
Text classification is well studied in many languages and, in particular, the English language [3]. Despite the fact that various works have been done to develop an Amharic document classifier, they all rely on machine learning models [4][5][6]. One of the most difficult problems in machine learning is reducing the dimensionality of a high-dimensional feature space. Dimensionality reduction has a direct impact on the model's overall performance. is problem can be solved by defining threshold values on the feature set. However, defining a threshold value for a machine learning model for dimensionality reduction, on the other hand, is complicated and time-consuming and may not yield consistent results for complex and massive data sets [7].
CNN uses convolving filters in the convolutional layer to extract high-level features. It is extremely difficult to determine the number of filters and the appropriate size of a filter for a specific task. Larger filters consume more resources during training, whereas small filters may produce inaccurate results [8,9]. CNN is capable of capturing contextual information for each feature. Figure 1 depicts the overall process of text classification using CNN.
Since the past two decades, machine learning methods have been widely used for Amharic document classification, but the potential of deep learning has not been exploited yet. To the best of our knowledge, no deep learning models for long Amharic document classification have been applied. Deep learning models have drawn the attention of researchers from all fields of science due to their dependable and superior performance [10]. e aim of this study is to improve the classification accuracy of Amharic documents using deep learning and pretrained fastText. Our contributions are listed as follows: (1) To the best of our knowledge, it is the first study of document-level deep-learning-based Amharic text classification (2) We compared the deep learning model's performance to that of well-known machine learning models and found that the deep learning model outperforms e rest of the paper is organized as follows. Related works are listed and explored in Section 2.
e detail methodology of the proposed work is presented in Section 3. e experimental findings are defined in Section 4, we provide the results in Section 5, and the paper is concluded in Section 6.

Related Works
In recent years, many researchers have worked on deep learning-based document classification. Using a deep learning model for document classification improves classification accuracy, according to these recent studies. Even though, there is a lack of attempts on deep-learning-based text classification for Amharic language, we review existing works related to the proposed scheme mainly considering previous attempts made on text classification with Semitic language groups.
In the work of [11], authors investigated the performance of deep learning models by introducing new single label and multilabel data sets for Arabic text categorization. To improve the classification accuracy of Arabic text classification using deep learning models, they employ the word2vec embedding. e authors in [12] proposed text classification based on CNN and word embedding for low-resource languages (Tigrinya). ey used manually annotated data sets of 30,000 Tigrinya news texts from various sources, which are divided into six categories: sport, agriculture, politics, religion, education, and health. e authors construct CNNs using a variety of techniques, including continuous bag-ofwords (CBOW), skip-gram, and CNNs with and without word2vec and fastText. fastText and word2vec pretrained vectors are compared in terms of their impact on text classification in their work. CNN with word2vec outperforms the other methods in terms of classification accuracy for Tigrinya news article classification, according to their experiment. e works of [13] proposed a noble incremental learning strategy to solve the feature extraction problem in deep learning in text classification. eir model consists of four components: a student model, a reinforcement module, a teacher module, and a discriminator module. e learn# method significantly reduced the classifier's training time, reduced it by 80%. e author did not discuss the impact of their model on classifier performance in terms of accuracy, precision, recall, and F-measure.
In [14], the authors proposed an attention-based bidirectional long short-term memory (Bi-LSTM) +CNN hybrid model text classification. e proposed model capitalizes on the advantages of LSTM and CNN by including an additional attention mechanism. ey tested their model using movie review data from the Internet Movie Database (IMDB), and the results showed that their proposed hybrid attention-based Bi-LSTM + CNN model outperforms MLP, CNN, and LSTM models in terms of classification accuracy, recall, and F1 scores. e work of [15] studies the application of deep learning in text categorization. ey combine it with textual characteristics and use the double bi-directional gated recurrent unit (GRU) + attention deep learning model to predict hotspots, and they achieved good results.
In reviewing the above papers, we noted that using deep learning model for text classification can enhance its performance. As a result, we proposed CNN + pretrained fastText that aimed for improving the performance of Amharic text classification.

Proposed Method
e components of the proposed document classification model for Amharic language are explained in this section. To evaluate and classify Amharic news documents, the proposed model includes four basic modules: data collection and preprocessing, data representation, classification module, and evaluation module. e proposed model of text classification is implemented according to the Algorithm 1.
e basic tasks or modules we used to predict the category label of raw Amharic document is described in Figure 2.
e description and function of each of the modules shown in Figure 2 is presented as follows.

Data Collection.
We collected 3,600 Amharic news documents from 6 major news categories to train and test the designed deep learning model, since there is no publicly available Amharic document classification corpus. e six news categories are chosen based on the most commonly used category in previous studies. Each document file is saved in separate file name within the corresponding category's directory, that is, all documents in the data set are single labeled. e main goal of the proposed model is to predict the category of Amharic news document collected from different news publishing mass media such as Ethiopian News Agency (ENA), Fana Broadcasting Corporate (FBC), and 2 Computational Intelligence and Neuroscience Walta Information Center (WIC). ENA publishes news documents on a wide range of topics, such as sports, business, education, technology, health, and politics. e data sets for this analysis were collected from the FBC, WIC, and ENA websites. e six news categories used in this analysis are business, education, health, technology, politics, and sports. Figure 3 shows the percentage of each news category in the data set as a pie chart.

Data Preprocessing.
e process of preparing raw data for a machine learning or deep learning model is known as data preprocessing. Text cleaning is the first step of preprocessing. Before forwarding a raw data set to the learning model, it must be preprocessed. Since there may be noise and irrelevant data in a raw data set, it must be preprocessed before passing to the learning model, which increases the training time and degrades the model's efficiency. e following are some of the text cleaning techniques we used in this study.

Normalization.
In the Amharic writing system, there are different characters with the same sound, and there is no written rule for using them in Amharic words. As a result, there is inconsistency in writing of terms. For  Computational Intelligence and Neuroscience example, the word power ("hayil") can be written as ሀይል ፣ ሐይል ፤ ኀይል፤ ሃይል፤ሓይል፤ኃይል፤ኻይል. Characters of the same sound are replaced with the canonical form used in this study. Table 1 shows normalization of Amharic characters having the same sound with different symbolic representation.

Stemming.
e process of removing infixes (ላ፣ጃ፣ጫ), prefixes (በ፣ስለ፣እንደ፣ከ፣ወደ), and suffixes (ኦች፣ኦችም፣ን፣ም፣ንም) from words to produce the stem or root form is known as stemming [16]. In information processing, morphological variants of Amharic words have the same interpretation. HornMorpho 4, developed by Gasser [17], was used for stemming. HornMorpho generates morphemes from a given word, and we only use the root or stem form for the next text processing step. For example, it produced the lemma "ጨዋታ" for the word "ጨዋታዎች" (games), and we used only the term's lemma for the remaining tasks.

Word
Embedding. Document classification involves the transformation of documents into feature vectors. is can be done via bag of words (BoW), word2vec, TF-IDF, and fastText. Word embedding is foundational to natural language processing and represents the words in a text in an R-dimensional vector space, thereby enabling the capture of semantics, semantic similarity between words, and syntactic information for words.

Bag of Word.
BoW counts the number of times a word appears in a document. Table 2 shows the BoW representation of the sentence "the club is effective" across three documents: doc1, doc2, and doc3.

Term Frequency-Inverse Document Frequency.
e document's terms are not all equally important in distinguishing one document from the others. To determine the significance of terms in a classification, the TF-IDF method is used. is method does not rely on frequency, but rather on the TF-IDF score of each term or word throughout the document. e algorithm works similarly to bag of words, but the word count is replaced with the TF-IDF score of each term. TF-IDF score of a given term t in a document can be formulated as follows: where N denotes the total number of documents in the document, DF denotes document frequency, t denotes the term, and d denotes the document.
3.5.3. word2vec. Word embedding via word2vec was proposed by Mikolov [18]. word2vec is a model that helps us represent the distributed representation of words in a corpus. In general, word2vec is an algorithm that accepts text as input and returns vectored representations of that text as output. word2vec starts with a set of random vector terms that scan the data set in a logical order while maintaining a background window around each word and its neighbors. e target word and its context are used by Business 17% Education 17% Health 17% Sport 16% Politics 16% Technology 17% Figure 3: Data set composition.  4 Computational Intelligence and Neuroscience word2vec to decide how they act when they traverse the corpus [18]. e algorithm calculates the dot product between the target word and the background words and attempts to reduce the stochastic gradient descent (SGD) metric efficiency. Even when in a similar context, two words meet their relation, or spatial distance is enhanced.

fastText.
Facebook provides pretrained word vectors for 157 languages, including Amharic [19]. ese models were trained with CBOW with position weights in dimension 300, character n-grams of length 5, a window of size 5, and 10 negatives. fastText is a Facebook-developed model for word representation and text classification that employs unsupervised approach to word representation. It is an extension of word2vec that views the representation of words from different directions. One of fastText's unique features is its ability to generate vectors for out-of-vocabulary words such as unknown words and tokens.
During word representation learning, fastText considers not only the word itself but also groups of characters from that word and subword information such as character unigrams, bigrams, and trigrams [20]. However, GloVe and word2vec fail to provide any vector representation for words that are not in the model dictionary [21]. As a result, in this study, fastText is used as a word representation model. We downloaded the fastText pretrained embedding model for Amharic language from the link https://fasttext.cc/docs/en/ crawl-vectors.html and then used it to generate the vector of words found in our preprocessed data set. e pretrained word vectors were found in two formats: binary and text. We used the binary format for this study. e algorithm to get the vector of each word in the data set is shown in Algorithm 2.
en the generated matrix is used by the embedding layer of CNN. e modules used to generate word vectors for each term in the corpus are depicted in Figure 4.

Classification.
With the increasing availability of Amharic documents on the internet, the proposed method uses convolution neural networks (CNNs) as a classifier method to classify certain documents. CNNs are advanced neural network models that are used to discover patterns and relationships between data items based on their relative positions [11]. CNNs can automatically learn effective text features representation from massive text using a 1D structure (word order) in the convolutional layer. It captures local relationships among the neighbor words in terms of context windows, and by using pooling layers, it extracts global features. CNN is a neural network made up of several convolutional and pooling layers. e block diagram of our proposed CNN-based text classification is shown in Figure 5. e layers used in our models are described as follows.

Embedding Layer.
e words of a document are represented as word vectors by the embedding layer. Every word in the text document is converted into a fixed-size dense vector. To map input token sequences to word vectors, the embedding layer employs various embedding techniques. Word embedding generates word representations that can be fed into the convolution layer, which learns to identify text classes based on the input vector sequences.

Convolutional Layer.
A convolutional layer is made up of neurons that calculate the output in two steps: first calculate a weighted sum and then output using a nonlinear activation function on the sum. e convolution layer applies different-width filters to the news embedding matrix to extract distinct features as vectors corresponding to each filter and generates a feature map. e convolution word filter considers positions that are independent for every word, and filters at higher layers capture syntactic or semantic associations between phrases that are far apart in a text.

Pooling Layer.
e pooling layer shrinks or reduces the size of the feature map by merging or selecting a highvalued feature from neighboring features, allowing the higher-level layers to deal with more global values. We employed 1-max-pooling. e main advantage of incorporating such a max-pooling layer into the network is that it reduces the number of parameters or weights and controls overfitting.

Fully Connected Layer.
In this layer, each neuron has full connections with all of the neurons in the previous layer. e connection structure is the same as with layers in classic Computational Intelligence and Neuroscience 5 neural network models. Dropout regularization is applied to the fully connected layer to avoid overfitting and therefore improve the generalization performance. When the training is finished, the parameters and learned weights are saved into a file, which can later be loaded and used to predict classes of unlabeled news. e fully connected layer is a traditional feedforward neural network (FNN) hidden layer. It can be viewed as a subset of the convolutional layer with kernel size 1. is layer belongs to the trainable layer weights class and is used in the final stages of CNNs.

Softmax Function.
Finally, the vector representations in the fully connected layer are passed to the Softmax function, which returns the probability distribution over the labels as an output.

Experiment
In this section, we investigate CNN's performance with trained fastText for Amharic news documents. e results of the experiments are compared to current state-of-the-art machine learning models. All experiments are carried out in a Windows 10 environment on a machine equipped with a Core i7 processor and 16 GB of RAM. e number of epochs was used as a predictor in experiments. e details, settings, and evaluation methods used in the experiment are listed in the following.

Data Set.
e total number of news documents used in our experiment is 3,600 (108,000 sentences) from the six news categories: business, education, health, politics, technology, and sport. We cleaned the data set and save in the.XLSX, where the label of the document was represented as integer value from 1 to 6. For business, education, health, politics, technology, and sport, we used 1, 2, 3, 4, 5, and 6, respectively. e total number of unique terms in our data set is 795,536.

CNN Training Parameters.
We tuned hyperparameters of CNN models on data sets using a grid search to find the best value for each model parameter. Choosing the best model parameter equals getting the best model performance. We treat Amharic documents as an array of sentences separated by four points (:) for training purposes. We used a CNN-based approach to learn automatically and categorize those sentences into one of the six evaluation categories. CNNs require inputs to be a fixed size, whereas sentence lengths can vary greatly. As a result, we used a word length of 300 on average. For sentences with fewer than 300 words, the index is filled by appending a zero at the end until it reaches 299 index. e network parameters were regularized with a dropout rate of 0.5 and a training epoch typical value of 50. We divide the data set in an 80-20 ratio and evaluate the model in terms of accuracy, F-measure, recall, and precision. Table 3 shows the CNN hyperparameters used in this study.

Performance Measure.
We used the most widely used performance assessment metrics to assess classification performance: accuracy, precision, recall, and F-measure [22]. Accuracy is the most common metric for classifier efficiency and can be calculated as follows: Precision is used to determine the correctness of a classifier's result and can be determined as follows: e completeness of the classifier results is measured by recall. e following equation is used to calculate it: F-measure is the harmonic mean of precision and recall and can be calculated as follows: where TP denotes true positive, TN denotes true negative, FP denotes false positive, and FN denotes false negative.

Results and Discussion
e proposed CNN model trained by fastText is evaluated in terms of accuracy, precision, recall, and F-measure using six Amharic news categories. e performance of the proposed CNN-based Amharic news document classification on six news categories is shown in Table 4. In the four evaluation matrices, the system produced nearly identical results. As a result, the number of false positives and false negatives is approximately equal. Furthermore, the model's performance is evaluated using unlabeled documents, and the results are summarized in Table 5. Table 5 shows that the proposed model correctly predicts label of three news. In addition to this, we conduct experiment on the performance of the new model by varying number of epochs.
As shown in Figure 6, we measured classification accuracy by varying the number of epochs. We used epoch 100 for the remaining experiments because it has the highest accuracy.

Comparison with Common Machine Learning Models.
We compared our CNN model with the state-of-art machine learning models. e CNN model was evaluated and compared to the three most commonly used machine learning models, MLP, SVM, XGB, RF, and decision tree (DT) classifiers. We used models from the scikit-learn library from the "sklearn" package for the five machine learning models. To account for this variation in results, we ran it ten times without changing any parameters and took the average of the results. In addition, we run an experiment on SVM, MLP, and RF classifiers, varying the hyperparameter values, and the results are shown in Table 6.
From the above results, we took classifier parameters that resulted in better classification accuracy for SVM, MLP, and    Table 7 shows a comparison of classification accuracy between the proposed mode and other machine learning classifiers.
In Table 7, the accuracy of the CNN-based method outperforms other methods, proving the effectiveness of this classification method. e following is an explanation of why performance should be improved in the CNN model: (1) using fastText to produce word vectors can get relevant features, and (2) processed text features through CNN can be better in representing high-level features of a document.

Conclusion
is paper focuses on a text categorization method based on CNN that does not require the extraction of text features in advance. CNN is used to extract higher-level features and improve network recognition. Simultaneously, prevent data from overfitting in the dropout method to improve the network's capacity. CNN's text classification performance is evaluated using various CNN parameters and epochs, and it is also compared to well-known machine learning classifiers. According to the experimental results, the proposed CNNbased classifier produced a better classification accuracy of 93.79%. As a result, the proposed model can be used in a variety of applications that require Amharic document classification, such as automatic document organization, topic extraction, and information retrieval. We aimed to examine additional categories and data sets in the future.

Data Availability
is research work's data set and source code are publicly available on GitHub (https://github.com/demekeendalie/ feature-selection).

Conflicts of Interest
e authors declare that they have no conflicts of interest.

Acknowledgments
is study was carried out by two academic staff members from Jimma Institute of Technology, Jimma University, Ethiopia. e authors would like to express their gratitude to