key: cord-0030465-6xnrj4z5 authors: Li, Xin; Hu, Lianting; Lu, Peixin; Huang, Tianhui; Yang, Wei; Lu, Quan; Liang, Huiying; Lu, Long title: A Novel Approach of Feature Space Reconstruction with Three-Way Decisions for Long-Tailed Text Classification date: 2022-04-16 journal: Comput Intell Neurosci DOI: 10.1155/2022/3183469 sha: 75dd43c9539c7ff9f7de59d6f36a5b9e8e93d479 doc_id: 30465 cord_uid: 6xnrj4z5 Text classification is widely studied by researchers in the natural language processing field. However, real-world text data often follow a long-tailed distribution as the frequency of each class is typically different. The performance of current mainstream learning algorithms in text classification suffers when the training data are highly imbalanced. The problem can get worse when the categories with fewer data are severely undersampled to the extent that the variation within each category is not fully captured by the given data. At present, there are a few studies on long-tailed text classification which put forward effective solutions. Encouraged by the progress of handling long-tailed data in the field of image, we try to integrate effective ideas into the field of long-tailed text classification and prove the effectiveness. In this paper, we come up with a novel approach of feature space reconstruction with the help of three-way decisions (3WDs) for long-tailed text classification. In detail, we verify the rationality of using a 3WD model for feature selection in long-tailed text data classification, propose a new feature space reconstruction method for long-tailed text data for the first time, and demonstrate how to effectively generate new samples for tail classes in reconstructed feature space. By adding new samples, we enrich the representing information of tail classes, to improve the classification results of long-tailed text classification. After some comparative experiments, we have verified that our model is an effective strategy to improve the performance of long-tailed text classification. Due to the rapid development of Internet technology and information construction, it becomes easier to obtain valuable text data to study and analyze problems in certain fields. As a consequence, text classification is widely studied by researchers in the natural language processing field [1] [2] [3] , and many applications based on text classification technology have been developed, such as news filtering and organization [4] , e-mail classification and spam filtering [5] , web page mining [6] , and medical documents' classification [3] . Text classification is a task that assigns textual main steps based on these models in text classification are shown in Figure 1 . As we all know, sufficient data are the premise of superior performances of these artificial intelligence learning methods; especially, deep learning models are developed to train massive data initially [18, 19] . But real-world text data often follow a long-tailed distribution as the frequency of each class is typically different, such as news topic classification, clinical name entities recognition, and disease diagnosis for electronic medical records [20] [21] [22] . It means that a dataset can have a large number of under-represented classes (tail classes) and a few classes with more than sufficient data (head classes). e performance of supervised learning algorithms in the process of text classification suffers when the training data are highly imbalanced [23, 24] . e problem can get worse when the categories with a fewer data are severely undersampled to the extent that the variation within each category is not fully captured by the given data. At present, few studies have studied the problem of longtailed text classification and put forward effective solutions. Some methods have been proposed to handle the problem of imbalanced datasets in text classification [25] [26] [27] [28] [29] . Finding and fine tuning the network model of classifiers suitable for specific data sets is one of these methods, such as [25] describing the proposed approach for text classification in an unbalanced data environment based on an implementing individual LSTM neural network. Data augmentation is also a frequently used way in scenarios with few samples or unbalanced categories [26] [27] [28] [29] , such as back translation [26] , which has become an effective way of data augmentation; for example, researchers translate some English text into Chinese and then translate them back into English, they can get various new training data, and the size of the dataset is doubled. Oversampling [27] and undersampling [28] are both common methods to deal with unbalanced datasets. But those methods which repeatedly take samples only consider the processing at the data level, so the generated data are still redundant and invalid in the feature space, and the improvement of classification effect is not obvious. e method of finding and fine tuning appropriate classifiers in specific data sets does not have universal applicability. At present, research studies on real-world long-tailed data mainly focus on the field of image classification [30] [31] [32] [33] [34] . ose techniques for learning long-tailed distributions generally fall into three groups: resampling [30] , reweighting and cost-sensitive learning [31, 32] , and feature manipulation [33, 34] . Especially, methods based on feature space augmentation of long-tailed image data have effectively improved the classification effect recently [33, 34] . Encouraged by the progress of handling long-tailed data in the field of image [30] [31] [32] [33] [34] , considering texts carry a lot of semantic information, analysis is more complex than image data, we try to integrate effective ideas into the field of longtailed text classification, and prove the effectiveness. In this paper, we come up with a novel approach of feature space reconstruction with 3WD for long-tailed text classification in Figure 2 . In detail, with the help of 3WD model, the main innovations of this paper can be summarized into three aspects: (1) We verify the effectiveness of using a 3WD model for feature selection in long-tailed text data classification. (2) We propose a new feature space reconstruction method for long-tailed text classification and demonstrate how to effectively generate new samples for tail classes in reconstructed feature space. By adding new samples, we enrich the representing information of tail classes, to improve the classification results of long-tailed text classification. (3) After some comparative experiments, we have verified that our model is an effective strategy to improve the performance of long-tailed text classification, which can effectively improve the accuracy values of tail classes. e rest of paper is organized as follows: Section 2 introduces the novel approach of feature space reconstruction with 3WD for long-tailed text classification related works. In Section 3, we design some experiments and provide the corresponding results in detail. Finally, we draw some conclusions and some possible future works in Section 4. In this section, we propose a novel approach of feature space reconstruction with 3WD for long-tailed text classification. As mentioned above, calculating text representations, training classification models, and predicting class labels for class-unknown documents are the main steps of text classification. Firstly, with the help of 3WD, two evaluation functions are used to generate the optimal set of features, which can effectively reduce the number of feature words, make the extracted feature words more representative of categories, and text representations can be calculated accordingly. en, because texts containing sufficient information in tail-class data only account for a small part, some information will be lost in the process of text representations, and it is easy to find that the representation vectors of tail classes are very sparse. us, we augment the data in tail classes by reconstructing the feature space with the features learned from the head classes with ample samples. In particular, we decompose the features of each class into classgeneric features and class-specific features. Novel samples of tail classes are then generated randomly by fusing the classspecific features from the tail classes with the class-generic features from head classes. Finally, the new samples are added to existing data to train the classification models and predict class labels to verify the effect of our method. Considering the decision risk and uncertainty, the 3WD method was initially proposed by Yao [35] based on rough sets and Bayesian decision procedure. e 3WD method which has attracted extensive attention in various fields provides a new tool for text classification [36] [37] [38] . e core idea of the 3WD method is to divide the universe into three disjoint regions, including an acceptance decision region, a deferment decision region, and a rejection decision region. If the decision maker has enough information, he (or she) can quickly make a decision, i.e., acceptance and rejection. Otherwise, the decision maker can choose to postpone the decision. Computational Intelligence and Neuroscience Let the dataset of long-tailed texts denotes as X � X Head , X Tail � x 1 , x 2 , . . . , x i , . . . , x n , where X Head is the set of the classes with ample samples (head classes), X Tail is the set of the samples of under-represented classes (tail classes), and the total amount of all texts in X is n. C � C Head , C Tail � C 1 , C 2 , . . . , C k , . . . , C K , C k denotes the kth class of X, and there are K classes in X. Firstly, we need to preprocess the long-tailed samples and do word e 3WD model used in this paper uses the TF-IDF algorithm [39] and Chi2 Statistics [40] as double-decision functions to select features in the long-tailed dataset, which can consider the frequency and distribution of vocabularies, as well as label-related information. M μ is defined as the feature set selected by decision function μ, and M ω is defined Computational Intelligence and Neuroscience 3 as the feature set selected by decision function ω. For ∀v j ∈ V: (1) Here, n i,j denotes the number of times vocabulary v j appeared in document x i , k n k,i is the total times of all vocabularies that appeared in document x i , and the total amount of all texts in X is |n|. | i: v j ∈ x i | is the total number of documents with v j . Adding 1 to prevent v j is not in X, resulting in a calculation error caused by zero denominators. where A jk is the number of documents that contain v j and belong to C k , B jk is the number of documents that contain v j but do not belong to C k , C jk is the number of documents that do not contain v j but belong to C k , and D jk is the number of documents that do not contain v j and not belong to C k . e feature sets M μ and M ω of the longed-tailed text data are generated by two evaluation functions, we map the feature words existing in both feature sets to the positive field POS(μ, ω) of the 3WD model. e feature words that only exist in one feature set M μ or M ω are added to the boundary domain BND(μ, ω) of the 3WD model, further processing is carried out with these features according to the classification results to decide to retain them or not, and the feature words with no appearance in feature sets are added to the negative domain NEG(μ, ω). 3WD rules can be constructed through three fields: the positive field POS(μ, ω) corresponds to acceptance, the negative field NEG(μ, ω) corresponds to rejection, and the boundary field BND(μ, ω) corresponds to non commitment. en, ∀v j ∈ V: and NEG(μ, ω) � (POS(μ, ω)⋃BND(μ, ω)) C . e final feature set T � t 1 , t 2 , . . . , t m is generated accordingly, which is a collection of m feature words used for classification. For the effectiveness of this method for feature selection in the process of long-tailed text data classification, we evaluate in Section 3.2 of the paper. In Section 2.1, we obtain the final feature set T � t 1 , t 2 , . . . , t m with m optimal feature words; thus, document x i can be represented as a vector of features . . , f m , and the collection of long-tailed text data X can be represented as equation: where element f ij represents the binary weight of t j from F i . If t j appears in x i , f ij � 1, and otherwise f ij � 0. us, for a given class C k , the matrix of C k can be expressed as equation: en, we can get the class feature vector of C k which is denoted as , sgn(x) � 1 for x > 0, sgn(x) � 0 for x � 0. F C k represents the binary weight of feature words in all documents of class C k , namely, if t j appears in x i , f ij � 1, otherwise f ij � 0. By comparing with head classes, we find that the class feature vectors of tail classes are extremely sparse, which contain small numbers of nonzero eigenvalues. is is because texts containing sufficient information in tail-class data only account for a small part, and some information is lost in the process of text representations, which leads to poor classification results. e data in the head class are sufficient. It seems natural to use the knowledge learned from the head class to help recover the missing information in the tail class. In [34] , authors use deep convolution neural networks to augment the feature space of long-tailed image data. rough image feature sampling, the class features of these images are divided into class-generic and class-specific features. It is proved that the class general features from the head class are regarded as transferable knowledge for the feature space expansion of the tail class, and the effectiveness of this idea is proved in the image filed. In this paper, we explore the effectiveness of feature reconstruction in long-tailed text data classification. For a given head class and a tail class, . . , f t m are class feature vectors of them correspondingly, and m is the total number of feature words. By analyzing their class feature sets, we map the feature words existing in both class feature sets to the class-generic feature set M G , and the feature words only 4 Computational Intelligence and Neuroscience existing in tail-class feature set to tail class-generic feature set M TS , thus for ∀t i ∈ T: Using the class-generic feature set M G and t tail classgeneric feature set M TS , we reconstruct the feature space of tail-class documents. e class-specific features from the class are then combined with the class-generic features from the Na classes linearly. A random combination ratio is generated to guide the fusion by randomly drawing classgeneric and class-specific feature vectors to form an augmented sample for the tail class. In Section 2.2, novel samples of tail classes are generated by fusing the classspecific features from the tail classes with the class-generic features from classes with ample data. Finally, the new samples are added to existing data to train the classification models and predict class labels to verify the effect of our method. In our paper, we use XGBoost as our classifier which is a method of gradient boosting decision tree. We evaluate the classification performance of the proposed method in our paper and compare it with other existing machine learning and deep learning methods. For each class C k , we use accuracy precision, recall, and F1measure as experiment metrics defined in Table 1 . In multiclass classification problems, the overall performance can be measured by averaging the evaluation methods. Microaverage and macroaverage are used widely for this purpose. In this study, in cases of evaluating longtailed distribution, it is better to use macroaverage scores than microaverage scores since the data size of categories is not considered in the microaverage calculation. Taking the definition of F1-value as an example, the definitions of macro-F1 and micro-F1 are shown in equations as follows: In this paper, we use Fudan University corpus as a long-tailed text classification corpus. e Fudan University TC corpus is from the Chinese NLP group in the Department of Computer Information and Technology, Fudan University of China. We randomly selected some category data to simulate the distribution of real-world longtailed datasets. e numbers and distribution of our dataset can be seen in Table 2 and Figure 3 . ere are 4 head classes and 12 tail classes. e ratio of training set to test set is 7 : 3. In this study, we carried out 3 experiments on the dataset. All experiments were implemented on a 64 bit MAC computer with 8 GB internal storage. e experimental code was written in Python language using Scikit-learn (sk-learn) and TensorFlow. Sk-learn and TensorFlow are commonly used third-party modules in machine learning and deep learning which encapsulate many commonly used machine learning and deep learning algorithms, such as SVM, XGBoost, RNN, CNN, and others. In preprocessing, all documents were segmented into words by the opensource tool Jieba, and stop words were removed in this process. By the final feature set obtained by the 3WD model, we can calculate the text representations of our long-tailed text dataset described in Section 2.1. And the number of nonzero features of each class is obtained, as shown in Figure 3 (b). It can be seen that the tail classes contain less information, and some information is lost in the process of text representations; it is easy to find that the representation vectors of tail class are very sparse. In this paper, we use the 3WD model to generate the final feature set T � t 1 , t 2 , . . . , t m of long-tailed text data, which is a collection of m feature words used for classification. In this section, we use word frequency, CHI2 [40] , and TF-IDF [39] as comparative methods for feature selection, to test the effectiveness of the 3WD model in our dataset. For all methods mentioned above, we select the top 2,000 feature words and use SVM as a classification model. e corresponding results are shown in Table 2 . From the results of Table 3 , we can see that the 3WD algorithm has improved the long-tailed text classification performance in the scores of accuracy, precision, recall, and F1-value than other methods. It is about 10~20% improvement than other methods on average. From the average point of view, 3WD model's overall performance is better than other methods. Computational Intelligence and Neuroscience In Section 2.2, we augment the tail-class data by reconstructing the feature space with the features learned from the head classes with ample samples. In particular, we decompose the features of each class into class-generic features and class-specific features. Novel samples of tail classes are then generated randomly by fusing the class-specific features from the tail classes with the class-generic features from head classes. In this section, the new samples are added to existing data to train the classification models and predict class labels to verify the effect of our method. According to the number of samples generated, it is divided into 6 levels, namely, 0%, 10%, 20%, 30%, 40%, 50%, and 60%, which represent the degrees of sample generation. In the first stage of our experiment, we use the 3WD algorithm as a feature selection method to select 2,000 effective feature words; the introduction and distribution of our long-tailed text dataset can be seen in Section 3.1. From the results of Figure 4 , we can see that when we add new samples of tail-class data by reconstructing the feature space by learning information from head-classes based on the original samples, the classification accuracy is significantly 6 Computational Intelligence and Neuroscience improved from 0 to 30% generation degree; the experiment shows that the classification results are not significantly improved after 30-40% generation degree. Especially, the accurate score of tail-class data has improved from 0.752 to 0.84 when the sample generation degree increased to 30%, from 30% degree to 60% degree, and the accuracy is not improved. e precision score of tail-class data has improved from 0.725 to 0.813, the recall score of tail-class data has improved from 0.72 to 0.797, the F1-value score of tailclass data has improved from 0.722 to 0.801 when the sample generation degree increased to 30%. It can also be seen from Figure 4 that the classification results of tail-class data have improved without reducing the accuracy of all data and head-class data. Methods. In our experiments, we use XGBoost as our classifier and compare it with four baseline training methods: TF-IDF with Linear SVM, CNN, RNN, and Bi-LSTM. XGBoost is a method of gradient boosting decision tree. It strives to maximize speed and efficiency, so it is called X (extreme) gboosted. Parameter settings for XGBoost in our experiments: max_depth:6, eta: 0.5, objective: multi: SoftMax, and num_class: 16. Convolutional neural network (CNN) is a variant of multilayer perceptron (MLP), which essentially is a multilayer perceptron. e key to its success lies in its way of local connection and sharing weights. On the one hand, it reduces the number of weights, which makes the network easy to optimize and reduces the risk of overfitting. Parameter settings for CNN in our experiments: embedding_dim � 100, seq_length � 600, num_classes � 16, num_filters � 256, ker-nel_size � 5, hidden_dim � 128, dropout_keep_prob � 0.5, learning_rate � 1e-3, and batch_size � 64. RNN (recurrent neural network) is a kind of a neural network used to process sequence data. e biggest difference of RNN is that it also establishes weight connections between neurons and between layer to capture the information at previous time points. Parameter settings for RNN in our experiments: embedding_dim � 100, seq_length � 600, num_classes � 16, num_layers � 3, hidden_dim � 128, rnn � "gru," dropout_keep_prob � 0.8, learning_rate � 1e-3, and batch_size � 128. Bi-LSTM is the abbreviation of bidirectional long short term memory, which is a combination of forwarding LSTM and backwarding LSTM. Both are often used to model context information in natural language processing tasks. LSTM is a kind of RNN that is very suitable for modeling time-series data, such as text data. Parameter settings refer to RNN. In this section, the results of accuracy, macroprecision, macrorecall, and macro-F1 score in our experiment are shown in Table 4 and Figure 5 . It is easy to see that the proposed method in our paper outperformed baseline training methods in all cases, especially in tail classes. e macro-F1 score obtained by our method in tail classes is 80.1%, which is 5∼10% higher than that obtained by deep Computational Intelligence and Neuroscience Computational Intelligence and Neuroscience learning models and 20% higher than that obtained by TF-IDF. Meanwhile, other results reflect similar improvements to our method. Long tail text classification is a problem that needs to be solved in the real world. e performance of current mainstream learning algorithms in text classification suffers when the training data are highly imbalanced. In this paper, a novel approach of feature space reconstruction with the help of 3WD for long-tailed text classification is proposed to greatly reduce the impact from the long-tailed distribution of datasets. In detail, we use Fudan University corpus as a long-tailed text classification corpus with 4 head classes and 12 tail classes, and verify the effectiveness of using the 3WD model for feature selection in long-tailed text data. en, we propose a new feature space reconstruction method for long-tailed text data and demonstrate how to effectively generate new samples for tail classes in reconstructed feature space. By adding new samples, we enrich the representing information of tail classes, to improve the classification results of long-tailed text classification. After performing some comparative experiments, we have verified that our model is an effective strategy to improve the performance of long-tailed text classification with the comparison with mainly used deep learning methods. As mentioned above, real-world text data often follow a long-tailed distribution as the frequency of each class is typically different. For example, the classification of diseases in the auxiliary diagnosis model based on electronic medical records usually presents a long tail distribution; the distribution of different types of entities extracted from biomedical texts often shows long-tailed state. Rather than using back translation for data augmentation, this paper verifies the effectiveness of feature space reconstruction for long-tailed data classification. In the next research, we will apply the method in this paper to the construction of an auxiliary diagnosis model to verify the effectiveness of our method and whether it can improve the accuracy of disease diagnosis. is study uses the Fudan University TC corpus which is from the Chinese NLP group in Department of Computer Information and Technology, Fudan University of China. All data are public and available online. e authors declare that they have no conflicts of interest. Xin Li and Lianting Hu designed the study, implemented the algorithm, preprocessed the data, and contributed equally to this paper. Long Lu, Quan Lu, and Huiying Liang gave critical suggestions. Xin Li and Tianhui Huang wrote the paper. Peixin Lu, Wei Yang, and Xin Li proofread the manuscript. Long Lu supervised the project and secured the grant funding. Text classification using novel term weighting scheme-based improved TF-IDF for internet media reports An in-text citation classification predictive model for a scholarly search system Evaluation and accurate diagnoses of pediatric diseases using artificial intelligence A comparison of classifiers and features for authorship authentication of social networking messages A new semantic-based feature selection method for spam filtering An improved focused crawler: using web page classification and link priority evaluation Modified frequency-based term weighting schemes for text classification A comparison of event models for Naive Bayes text classification Classification based on decision tree algorithm for machine learning Support vector machine text classification system: using Ant Colony Optimization based feature subset selection Design of the piano score recommendation image analysis system based on the big data and convolutional neural network Dynamic embedding projection-gated convolutional neural networks for text classification Utilizing recurrent neural network for topic discovery in short text scenarios Supervised learning algorithm for multilayer spiking neural networks with long-term memory spike response model Scene text recognition based on bidirectional LSTM and deep neural network Cross-domain sentiment classification with bidirectional contextualized transformer language models A survey of text classification algorithms HDLTex: hierarchical deep learning for text classification Data augmentation for improving deep learning in image classification problem Construction of financial management early warning model based on improved ant colony neural network Construction and evaluation of intelligent medical diagnosis model based on integrated deep neural network Investigating COVID-19 news across four nations: a topic modeling and sentiment analysis approach Reducing performance bias for unbalanced text mining Screening PubMed abstracts: is class imbalance always a challenge to machine learning? Classification with LSTM networks in user behaviour analytics with unbalanced environment A joint back-translation and transfer learning method for low-resource neural machine translation SMOTE: synthetic minority over-sampling technique Exploratory undersampling for class-imbalance learning Fine-tuning BERT for multilabel sentiment analysis in unbalanced code-switching text Learning deep representation for imbalanced classification Learning to model the tail Large-scale long-tailed recognition in an open world Deep representation learning on long-tailed data: a learnable embedding augmentation perspective Feature space augmentation for long-tailed data ree-way decisions with probabilistic rough sets Two-stage three-way enhanced technique for ensemble learning in inclusive policy text classification Text classification of minimal risk with three-way decisions Chinese text classification algorithm based on ree-way Decisions BNS feature scaling: an improved representation over tf-idf for svm text classification Presenting a hybrid feature selection method using Chi2 and DMNB wrapper for E-mail spam filtering e numerical calculations in this paper have been done on the supercomputing system in the Supercomputing Center of Wuhan University.