Next Article in Journal
Triadic Structures in Interpersonal Communication
Previous Article in Journal
Alignment: A Hybrid, Interactive and Collaborative Ontology and Entity Matching Service
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Neighborhood Attribute Reduction: A Multicriterion Strategy Based on Sample Selection

1
School of Computer Science and Engineering, Jiangsu University of Science and Technology, Zhenjiang 212003, China
2
School of Science, Jiangsu University of Science and Technology, Zhenjiang 212003, China
*
Authors to whom correspondence should be addressed.
Information 2018, 9(11), 282; https://doi.org/10.3390/info9110282
Submission received: 16 September 2018 / Revised: 2 November 2018 / Accepted: 10 November 2018 / Published: 16 November 2018

Abstract

:
In the rough-set field, the objective of attribute reduction is to regulate the variations of measures by reducing redundant data attributes. However, most of the previous concepts of attribute reductions were designed by one and only one measure, which indicates that the obtained reduct may fail to meet the constraints given by other measures. In addition, the widely used heuristic algorithm for computing a reduct requires to scan all samples in data, and then time consumption may be too high to be accepted if the size of the data is too large. To alleviate these problems, a framework of attribute reduction based on multiple criteria with sample selection is proposed in this paper. Firstly, cluster centroids are derived from data, and then samples that are far away from the cluster centroids can be selected. This step completes the process of sample selection for reducing data size. Secondly, multiple criteria-based attribute reduction was designed, and the heuristic algorithm was used over the selected samples for computing reduct in terms of multiple criteria. Finally, the experimental results over 12 UCI datasets show that the reducts obtained by our framework not only satisfy the constraints given by multiple criteria, but also provide better classification performance and less time consumption.

1. Introduction

Rough sets [1,2], firstly proposed by Pawlak, have been demonstrated to be useful in data mining [3,4], artificial intelligence [5], decision analysis [6,7], and so on. As one of the important strategies of feature selection, attribute reduction in rough-set theory plays a key role, since it provides us with clear semantic explanations of the selected attributes. Those semantic explanations can be reflected by constraints in terms of the considered measures, such as approximation quality and conditional entropy. For example, Hu et al. [8] have studied uncertainty measures related to fuzzy rough sets, and then further explored approximation quality-based attribute reduction; Dai et al. [9,10,11] investigated attribute reduction with respect to several types conditional entropies; Wang et al. [12] not only proposed a conditional discrimination index for overcoming the limitations of conditional entropies, but also provided the corresponding approach to attribute reduction.
It must be noted that most of the previous results about attribute reduction are based on the consideration of a single measure. For example, if attribute reduction is designed to preserve the approximation quality, then intuitively it may not perform well in learning tasks. This is mainly because that approximation quality is only a measure of uncertainty, which is slightly related to, for example, classification, clustering, or regression. From this point of view, multiple criteria determined by different measures should be emphasized in attribute reduction.
Generally speaking, if the definition of attribute reduction is given, then the immediate problem is to find the reduct. In the field of rough sets, widely accepted strategies for finding reducts include the exhaustive and heuristic algorithms. Although the former can be used to derive all reducts, it is not suitable for practical applications because of huge computational complexity. Different from the exhaustive algorithm, the heuristic algorithm can be realized by a greedy searching strategy, which has been favored by most researchers thanks to its speed advantage.
Nevertheless, it is worth noting that the process of the heuristic algorithm is still based on scanning all samples in the data. For example, to generate the binary relation that characterizes the similarity between any two samples, all data samples must be taken into account. However, time consumption may fail to satisfy practical applications with the sharply increasing of the size of dataset. In order to tackle this problem, various attribute-reduction algorithms have been proposed that intended to utilize sample selection to reduce the dataset and simultaneously maintain the performance of the classification accuracy. Angiulli et al. [13] proposed the Fast Condensed Nearest Neighbor, which attempt to remove surplus samples and reduce the dataset. Li et al. [14] presented a critical pattern selection algorithm by considering local geometrical and statistical information. This algorithm selected both border and edge patterns from the dataset. Nicolia et al. [15] proposed a sample selection algorithm dealing with the class-imbalance problem. Lin et al. [16] proposed an approach for detecting the representative samples from large datasets. Zhai et al. [17,18] have analyzed the algorithms above and proposed the sample selection approaches that aim to achieve the same performance of a machine-learning algorithm as the whole dataset is used. Zhang et al. [19] proposed a fuzzy rough set-based information entropy for sample selection in a mixed dataset. Xu et al. [20] have applied the sample selection technique to multilabel feature selection for reducing time consumption. Following these results, sample selection may be feasible and effective.
From the above discussions, the motivation of this paper was to design a new framework of attribute reduction that considers both multiple criteria in the definition of attribute reduction and sample selection in the process of finding reducts. Firstly, sample selection was executed, which aims to decrease the number of samples. Immediately, lower time consumption may be achieved, mainly because the size of the data has been reduced. Secondly, the multicriteria strategy that considers at least two measures was designed. Finally, the voting mechanism was used to select the candidate attribute in each iteration, and then the multiple criteria reduct with sample selection was obtained. Our approach can not only guarantee that the derived reduct satisfies the constrains in terms of different measures, but also make the ensemble selection [5,21] of attributes possible.
The rest of this paper is organized as follows. First, we review some basic concepts in Section 2.1. In Section 3, following the limitations of a single measure, a multiple criteria-based attribute reduction strategy is studied. Then, sample selection and the multiple-criteria strategy are combined to find the corresponding reduct. In Section 4, the effectiveness of our approach over 12 UCI datasets is analyzed. We then conclude with some remarks and perspectives for future work in Section 5.

2. Preliminaries

2.1. Neighborhood Relation

Without loss of generality, a decision system can be described as D S = < U , A T { d } > , in which universe U is the set of samples, A T is the set of condition attributes, and d is a decision attribute. Furthermore, x U , d ( x ) indicates the label or decision value of sample x, and a ( x ) denotes the value of x over condition attribute a where a A T .
Given decision system D S , assume that all decision values in D S are discrete, and an indiscernibility relation [22,23,24] IND d can be defined as:
IND d = { ( x , y ) U × U : d ( x ) = d ( y ) } .
By IND d , we can obtain a partition over the universe, such that
U / IND d = { X 1 , X 2 , , X q } ,
i.e., universe U is partitioned into q different decision classes. Therefore, X i U / IND d , X i is referred to as the i-th decision class in rough-set theory. The decision class that contains the sample x is denoted by [ x ] d .
The rough-set objective is to approximate the decision classes by the information given by condition attributes. Such information can actually be represented by the form of information granules [25,26] from the viewpoint of granular computing. For instance, the equivalence class used in traditional rough sets is a typical example of an information granule.
Nevertheless, it should be emphasized that the equivalence classes are only suitable for dealing with categorical data, while numerical data [27,28,29] have been seen everywhere in real-world applications. To fill such a gap, many different types of information granules have been proposed. As an important information granule used in generalized rough sets, neighborhood has been widely accepted by researchers. This is mainly because: (1) the construction of neighborhood is based on the distance that can characterize the similarity between samples with numerical data; (2) different neighborhood scales can be easily obtained by using different radii, and then a multigranularity structure is naturally formed. To know what neighborhood is, the concept of neighborhood relation should be given as follows:
Given a decision system D S , A A T , suppose that A : U × U R + { 0 } is the Euclidean distance function in which R + is the set of positive real numbers, then A ( x , y ) represents the Euclidean distance between samples x and y by using the information over condition attributes in A. Immediately, the neighborhood relation is:
N A = { ( x , y ) U × U : A ( x , y ) δ } ,
in which δ is a given radius such that δ 0 .
Based on neighborhood relation, it is not difficult to obtain the neighborhood of x in terms of A, such that:
δ A ( x ) = { y U : A ( x , y ) δ } .

2.2. Neighborhood Rough Set and Neighborhood Classifier

Given a sample x, to avoid that only the sample x is in the neighborhood of x, which may bring us the difficulty for classification, Hu et al. [30] have modified the radius such that
δ = min y U { x } A ( x , y ) + δ · max y U { x } A ( x , y ) min y U { x } A ( x , y ) .
Following Equation (5), the modified neighborhood of sample x in terms of A is
δ A ( x ) = { y U : A ( x , y ) δ } .
Definition 1.
Given decision system D S , A A T and X i U / IND d , the neighborhood lower and upper approximations of X i in terms of A are defined as
X i ̲ A = { x U : δ A ( x ) X i } ;
X i ¯ A = { x U : δ A ( x ) X i } .
Pair X i ̲ A , X i ¯ A is referred to as a neighborhood rough set of X i in terms of A.
The concept of neighborhood can not only be used to construct a rough set, but can also be applied to design classifier [31]. Let us consider one of the simplest classifiers, i.e., the K Nearest Neighbors algorithm (KNN), which is effective in many cases. It is a lazy learning method: for a testing sample to be classified, its K nearest neighbors form a neighborhood of such testing sample, and the voting mechanism is used to determine the label of the testing sample based on the real labels of all samples in neighborhood. For more details about KNN algorithm, see references [32,33,34].
The main thinking of the neighborhood classifier [8] is similar to that of KNN; the difference lies in the fact that the number of neighbors used in a neighborhood classifier is determined by the radius, while the number of neighbors used in KNN is specified by experts. Therefore, different samples may have different numbers of neighbors if a neighborhood classifier is used. The detailed process of the neighborhood classifier [8] is shown in Algorithm 1, as follows.
Algorithm 1 Neighborhood Classifier (NEC)
Inputs: Decision system D S = < U , A T { d } > , a testing sample y U , radius δ ;
Outputs: Predicted label of y: Pre A T ( y ) .
1. x U , compute A T ( y , x ) ;
2.Compute δ by Equation (5), and then obtain δ A T ( y ) by Equation (6);
// Note that in NEC,  y δ A T ( y ) ;
3. X i U / IND d , compute the probability Pr ( X i | δ A T ( y ) ) = | δ A T ( y ) X i | | δ A T ( y ) | ;
4. X k = arg max { Pr ( X i | δ A T ( y ) ) : X i U / IND d } ;
5.Find the corresponding label in terms of X k and assign it to Pre A T ( y ) ;
6.Return Pre A T ( y ) .

2.3. Measures

Approximation quality is frequently used to evaluate the certainty of belongingness in rough-set theory. In a neighborhood rough set, the formal definition is shown as follows:
Definition 2.
Given decision system D S , A A T , the approximation quality of d related to A is defined as
γ ( A , d ) = i = 1 q X i ̲ A U .
This reflects the percentage of the samples that belong to one of the decision classes determinately by the semantic explanation of lower approximation. Obviously, 0 γ ( A , d ) 1 holds.
Remark 1.
Note that A A T , γ ( A , d ) γ ( A T , d ) does not always hold; the reason is that, if some condition attributes are eliminated from A T , then the value of δ obtained by Equation (5) changes.
Example 1.
As the decision system shown in Table 1, U = { x 1 , x 2 , x 10 } is the set of samples, A T = { a 1 , a 2 , , a 7 } is the set of condition attributes, and d is a decision attribute.
Suppose that δ = 0.10 , then by Equation (5), the values of δ for ten samples are 0.8553, 0.7208, 0.8463, 0.8583, 0.8651, 0.8183, 0.7941, 0.6669, 0.9005, and 0.7025, respectively, if A T is used. Consequently, it is obtained that γ ( A T , d ) = 0.1000 .
Suppose that condition attribute a 7 is eliminated from the above decision system and then A = { a 1 , a 2 , , a 6 } , by Equation (5), the values of δ for ten samples are 0.8212, 0.7059, 0.7633, 0.8349, 0.6992, 0.6860, 0.6738, 0.6289, 0.8439, and 0.6767, respectively. Immediately, γ ( A , d ) = 0.2000 is obtained.
The above results tell us that γ ( A , d ) γ ( A T , d ) does not always hold if A A T .
Conditional entropy is another widely accepted measure that is an effective tool for characterizing distinguishable ability in a decision system. The lower the value of conditional entropy is, the higher the ability of that condition attribute to distinguish samples that we will have. Such discrimination can be considered as a type of uncertainty. Presently, many definitions of conditional entropies have been proposed in terms of different requirements [9,10,11,35,36,37]. A typical representation of conditional entropy is shown in Definition 3.
Definition 3.
Reference[30] Given decision system D S , A A T , the conditional entropy of d related to A is defined as:
ENT ( A , d ) = 1 | U | x U log | δ A ( x ) [ x ] d | | δ A ( x ) | .
A A T , ENT ( A , d ) ENT ( A T , d ) does not always hold. This is because not only does the monotonic property of Equation (10) not hold [19], but also the value of δ obtained from Equation (5) is changed if some condition attributes are eliminated from A T .

3. Multiple-Criteria Reduct with Sample Selection

3.1. Attribute Reduction

Attribute reduction is one of the key topics in rough-set theory [38]. Generally speaking, the purpose of attribute reduction is to delete the redundant attributes by some given constraints [39]. These constraints can be constructed by well-known measures such as approximation quality and conditional entropy. Many different definitions of attribute reduction have been proposed with different measures or requirements. Dai et al [9] proposed extended conditional entropy in interval-valued decision systems and designed corresponding definitions and algorithms. Yao et al. [7] addressed different measures, such as confidence, coverage, generality, cost, and decision monotocity based on the decision-theoretic rough-set models. Jia et al. [40] compared most popular definitions and then proposed a generalized attribute reduction that not only considers the data but also users’ preferences. For more details about definitions of attribute reductions, see references [9,41,42,43].
Definition 4.
Given decision system D S , A A T ,
1.
A is the approximation quality reduct if and only if γ ( A , d ) γ ( A T , d ) and A A , γ ( A , d ) < γ ( A , d ) ;
2.
A is the conditional entropy reduct if and only if ENT ( A , d ) ENT ( A T , d ) and A A , ENT ( A , d ) > ENT ( A , d ) .
Different from Pawlak’s [1,2] traditional definition of attribute reduction for preserving approximation quality, the constraint used in Definition 4 indicates that the approximation quality will not be decreased at least. The reason is shown in Remark 1: the approximation quality is not strictly monotonic in terms of variations of condition attributes. The case of conditional entropy is similar to that of approximation quality.
If γ ( A { a } , d ) γ ( A , d ) , then a is redundant; in other words, attribute a has no contribution to the increase of approximation quality. If γ ( A { a } , d ) > γ ( A , d ) , then a can be considered as a member in the reduct set. Similarly, it is trivial to present the semantic explanation of redundant attributes in terms of conditional entropy reduct. Therefore, the significance of attributes in terms of two different reducts shown in Definition 4 can be defined as follows:
Sig γ ( a , A , d ) = γ ( A { a } , d ) γ ( A , d ) ;
Sig ENT ( a , A , d ) = ENT ( A , d ) ENT ( A { a } , d ) .
In a decision system, the above two significances both satisfy that the higher the value is, the more important the condition attribute a will be. Following the given significances, the algorithms of finding the reduct must be immediately designed. Up to now, many algorithms have been proposed to obtain reducts. Considering time efficiency, the forward greedy search strategy has become a common way to do this. This kind of algorithm starts from an empty set and gradually adds the attribute with the maximum significance into the candidate attribute subset in each iteration [44] until the constraint is satisfied. This kind of approach is frequently referred to as the heuristic algorithm.
Take the approximation quality reduct as an example; the reduct aims to derive a subset of condition attributes that do not decrease the value of approximation quality. The detailed process of heuristic algorithm for finding such reduct is shown in Algorithm 2.
Algorithm 2 Approximation Quality Reduct (AQR)
Inputs: Decision system D S = < U , A T { d } > , radius δ .
Outputs: An approximation quality reduct A.
1.Compute γ ( A T , d ) ;
2. A ;
3.Do
(1)
a i A T A , compute Sig γ ( a i , A , d ) , select a j such that Sig γ ( a j , A , d ) = max { Sig γ ( a i , A , d ) : a i A T A } ;
(2)
A = A { a j } ;
(3)
Compute γ ( A , d ) ;
Until γ ( A , d ) γ ( A T , d ) ;
4.Return A.
Similarly, if it is required to compute Conditional Entropy Reduct (CER), as the conditional entropy is a measure that characterizes the distinguishing information of a subset of attributes, and the lower of the value of the conditional entropy is, the greater the distinguishment ability of the attribute set is. Then, the termination in Step 3 of Algorithm 2 is replaced by “ENT ( A , d ) ENT ( A T , d ) ”, and the significance of attribute in Step 3(1) is replaced by Equation (12), i.e., S i g ENT ( a i , A , d ) ; then, we select the attribute that has maximum significance.
The time complexity of a computing neighborhood relation is O ( | U | 2 ) , in which | U | is the number of samples in a dataset. In the worst case, there are | A T | attributes should be added into the reduct, i.e., no attribute is redundant; then, Step 3 in Algorithm 2 is executed | A T | times. In the i-th iteration, Step 3 is executed | A T | i + 1 times. Finally, the time complexity of AQR is O ( | U | 2 × | A T | 2 ) . Similarly, the time complexity of CER is also O ( | U | 2 × | A T | 2 ) .

3.2. Limitations of Single Measure

The above algorithm shows us a complete process of computing the reduct that is determined by a single measure, i.e., either approximation quality or conditional entropy. However, the derived reduct may fail to meet the constraints with multiple criteria. We used the following example to explain it:
Example 2.
In the decision system shown in Table 1, suppose that δ = 0.15 , then γ ( A T , d ) = 0.1000 and ENT ( A T , d ) = 0.6879 ; both of them are obtained by raw attributes.
Furthermore, by Definition 4 and the heuristic process, the obtained approximation-quality reduct is A 1 = { a 1 , a 2 , a 3 } , and the obtained conditional-entropy reduct is A 2 = { a 1 , a 2 , a 3 , a 4 , a 5 } .
If the approximation-quality reduct is selected, then γ ( A 1 , d ) = 0.1000 , and ENT ( A 1 , d ) = 0.7860 . It is observed that the value of approximation quality is maintained, while the value of conditional entropy is increased. In other words, though the constraint based on approximation quality is satisfied by A 1 , such subset of attributes does not meet the constraint in terms of conditional entropy.
If the conditional-entropy reduct is selected, then γ ( A 2 , d ) = 0.0000 , and ENT ( A 2 , d ) = 0.6762 . It is observed that the value of conditional entropy has been significantly decreased, while the value of approximation quality is also decreased. In other words, the constraint defined by approximation quality cannot be guaranteed if the conditional-entropy reduct is used.
The above example tells us that a reduct in terms of a single measure does not meet the constraints in terms of multiple criteria. Therefore, to alleviate such a problem, we propose a multiple-criteria attribute-reduction strategy that considers both the evaluations of approximation quality and conditional entropy.

3.3. Multiple-Criteria Reduct

Since the single-criterion reduct cannot meet the constraints, then the multiple-criteria framework [45] can be a solution. The definition of a multiple-criteria reduct presented as follows:
Definition 5.
Given decision system D S , if A A T , A is the multiple-criteria reduct if and only if:
1.
γ ( A , d ) γ ( A T , d ) and ENT ( A , d ) ENT ( A T , d ) ;
2.
A A , γ ( A , d ) < γ ( A , d ) or ENT ( A , d ) > ENT ( A , d ) .
Different from the approximation-quality and conditional-entropy reducts shown in Definition 4, the multiple-criteria reduct shown in Definition 5 is defined by considering constraints given by both approximation quality and conditional entropy.
Algorithm 3 presents a heuristic process to compute our multiple-criteria reduct. It should be emphasized that, to derive attribute significance, Equations (11) and (12) should be used.
In Step 3, m and n represent the attribute locations that have maximal significances in terms of approximation quality and conditional entropy, respectively. In Step 3(2), if m and n are the same, then there is no conflict for voting. Otherwise, two attributes have conflict, which means that the max values of significances computed by the measures of approximation quality and conditional entropy are derived from different attributes. Then, a mechanism for selection is required. In this case, we select one attribute without considering the measures, mainly because approximation quality and conditional entropy take the same weight in our algorithm. To make the algorithm more stable, the attribute ranks lower in the order of the raw attributes that are selected instead of a random one. Such thinking is similar to what has been addressed in reference [21].
Similar to AQR, the time complexity of MCR is O ( | U | 2 × | A T | 2 ) , where | U | represents the number of samples in decision system ( D S ), and | A T | represents the number of condition attributes. However, MCR may spend more time on computing reduct, because MCR should compute two different types of significances in each iteration.
Algorithm 3 Multiple-Criteria Reduct (MCR)
Inputs: Decision system D S = < U , A T { d } > , radius δ .
Outputs: A multiple criteria reduct A.
1.Compute γ ( A T , d ) and ENT ( A T , d ) ;
2. A ;
1.Do
(1)
a i A T A , compute Sig γ ( a i , A , d ) and Sig ENT ( a i , A , d ) , select a m and a n such that Sig γ ( a m , A , d ) = max { Sig γ ( a i , A , d ) : a i A T A } , Sig ENT ( a n , A , d ) = max { Sig ENT ( a i , A , d ) : a i A T A } ;
(2)
Select a j , where j = min ( m , n ) ;
(3)
A = A { a j } ;
(4)
Compute γ ( A , d ) and ENT ( A , d ) ;
Until γ ( A , d ) γ ( A T , d ) and ENT ( A , d ) ENT ( A T , d ) ;
4.Return A.

3.4. Multiple-Criteria Reduct with Sample Selection

Obviously, the process of the heuristic algorithm for computing the reduct is still based on scanning all samples in the data. To further improve the time efficiency of the algorithms shown in Section 3.1 and Section 3.3, reducing the size of samples may be a feasible solution.
In the field of machine learning and feature selection [46,47], the technique of sample selection has been widely used. For instance, following many previous results [20,48,49,50], it has been pointed out that sample selection is a useful method. Wilson et al. [48] provided a survey of previous algorithms, and proposed six additional reduction algorithms that can be used to remove samples from the concept description. Brighton et al. [49] proposed that internal samples positioned away from class boundaries have little or no effect on classification accuracy; on the contrary, samples that lie close to class boundaries hold more information to accurately describe the decision surface. Nikolaidis et al. [50] proposed the Class Boundary Preserving Algorithm (CBP). CBP divided all data into two sets that are referred to as the internal-sample set and boundary-sample set, and focused more on the boundary samples. Xu et al. [20] further expanded the sample selection of boundary samples and introduced it into multilabel datasets. From the above analyses, we can find that the samples in the boundary region are more important than other samples. We propose an algorithm to compute a multiple-criteria reduct by using boundary samples instead of whole samples in the data.
First of all, we used a K-means clustering algorithm to choose K cluster centroids [51,52,53,54]. This process is executed M times because the result of K-means clustering is not stable. Secondly, we compute average cluster centroids by those results. Finally, we select those samples that are far away from the average cluster centroids, and construct a new decision system. We select those samples that are far away from the average cluster centroids, which is mainly because: (1) these samples are more difficult to be correctly classified, and samples nearer to the average cluster centroids tend to be closer to each other, making it easy for them to be classified correctly; (2) these samples sometimes fail to be assigned to the lower approximation set, while the samples closer to the average cluster centroids tend to be in the lower approximation set. Therefore, in order to improve classification performance and reduce the time consumption with the neighborhood rough-set model, we apply boundary samples instead of all samples. To judge whether a sample is far away from the cluster centroid, we used Definition 6, as follows:
Definition 6.
Given a cluster C j , C j * is the cluster centroid of C j , dist ( x , C j * ) denotes the distance between x C j and C j * , and the average distance between all samples in C j and C j * is
dist ¯ ( C j * ) = 1 | C j | x C j dist ( x , C j * ) .
Remark 2.
x C j , if dist ( x , C j * ) dist ¯ ( C j * ) , then x is referred to as a sample which is far away from the cluster centroid C j * , such sample is selected for constructing new decision system.
With all boundary samples selected, new decision system D S can be constructed. Obviously, the size of the data in D S is smaller than that in decision system D S . From this point of view, the time consumption of computing the reduct may be reduced. Algorithm 4 shows us the heuristic process to compute a multiple-criteria reduct by using sample selection.
Algorithm 4 Multiple-Criteria Reduct with Sample Selection (MCRSS)
Inputs: Decision system D S = < U , A T { d } > , radius δ , M.
Outputs: A multiple criteria reduct A.
1. U = ;
// Initialize the universe of new decision system;
2.For r = 1 to M
 Execute K-means clustering algorithm over D S , obtain clusters C r = { C 1 r , C 2 r , , C K r } ;
// In K-means clustering, K is the number of decision classes;
End For
13.For j = 1 to K
 Obtain the j-th average cluster centroid
C j * = r = 1 M C j r M ;
End for
4.For j = 1 to K
x C j , if dist ( x , C j * ) dist ¯ ( C j * ) , then U = U { x } ;
End For
// The new decision system D S = < U , A T { d } > is constructed;
5.Compute γ ( A T , d ) and ENT ( A T , d ) over D S ;
6. A ;
7.Do
(1)
a i A T A , compute Sig γ ( a i , A , d ) and Sig ENT ( a i , A , d ) , select a m and a n such that Sig γ ( a m , A , d ) = max { Sig γ ( a i , A , d ) : a i A T A } , Sig ENT ( a n , A , d ) = max { Sig ENT ( a i , A , d ) : a i A T A } ;
(2)
Select a j , where j = min ( m , n ) ;
(3)
A = A { a j } ;
(4)
Compute γ ( A , d ) and ENT ( A , d ) ;
Until γ ( A , d ) γ ( A T , d ) and ENT ( A , d ) ENT ( A T , d ) ;
8.Return A.
The first four steps show us the process of sample selection, i.e., the process of constructing a new decision system. In Step 1, the universe of the new decision system is initialized. In the following two steps, a K-means clustering algorithm is executed M times, and the average cluster centroids are obtained. In Step 4, samples that are far away from the average cluster centroids are immediately selected, and the new decision system is constructed. The last three steps are used to compute a multiple-criteria reduct over the new decision system.
The time complexity of MCRSS is O ( | U | 2 × | A T | 2 + K × | U | × M ) , where | U | represents the number of samples in the new decision system ( D S ); both K and M are constants. It must be noted that | U | < | U | .
The time complexity of MCR is O ( | U | 2 × | A T | 2 ) , so we compare the time complexity between MCR and MCRSS. Generally speaking, K × M ( | U | | U | 2 / | U | ) × | A T | 2 holds for most of the data because K and M are constants that are much less than the number of samples.
Therefore, it is a trivial to show that O ( | U | 2 × | A T | 2 ) O ( | U | 2 × | A T | 2 + K × | U | × M ) holds, in other words, the time consumption of MCRSS is less than that of MCR.
The sample-selection strategy shown in Algorithm 4 can also be used in computing approximation-quality and conditional-entropy reducts. It is immediately trivial to design two algorithms: Approximation-Quality Reduct with Sample Selection (AQRSS) and Conditional-Entropy Reduct with Sample Selection (CERSS). The time complexities of AQRSS and CERSS are also O ( | U | 2 × | A T | 2 + K × | U | × M ) .

4. Experimental Analysis

To validate the effectiveness of MCRSS proposed in this paper, 12 UCI datasets were collected to conduct the experiments. The basic descriptions of the datasets are shown in Table 2. All experiments were carried out on a personal computer with Windows 7, dual-core 1.50 GHz CPU, and 8 GB memory. The programming language was MATLAB R2016a.
In the following, five-folder Cross-Validation (5-CV) was adopted. In other words, we divided each set of data into five parts of the same size, which are denoted by U 1 U 5 : for each round of computation, 80% of the samples in the data were regarded as the training samples for computing reducts, and the rest were considered as the test samples for computing measures by the attributes in reducts. Furthermore, in this experiment, M = 5 , i.e., the K-means clustering, was executed five times to generate average cluster centroids. Ten different values of δ , such that 0.03 , 0.06 , , 0.30 , were also selected.

4.1. Comparisons of Approximation Qualities

Figure 1 shows us the detailed results of approximation qualities with respect to three different algorithms for computing reducts, AQRSS, MCRSS, and CERSS. AQRSS and CERSS are an approximation-quality reduct and a conditional-entropy reduct with sample selection technique, while MCRSS is a multiple-criteria reduct with sample selection. AQRSS, MCRSS, and CERSS have the same meaning in other comparisons in Section 4.
In Figure 1, we can observe the following:
  • If the value of δ increases, then the decreasing trends have been obtained for approximation qualities with respect to three different reducts, though those decreasing trends are not necessarily monotonic.
  • By comparing it with AQRSS, MCRSS can preserve or slightly increase approximation qualities. This is mainly because the constraint designed by the measure of approximation quality is also considered in MCRSS. Take, for instance, the “Ionosphere” dataset; if δ = 0.12 , then the approximation qualities derived by MCRSS and AQRSS are 0.6049 and 0.4644, respectively.
  • An interesting observation is that the approximation qualities obtained by CERSS may be greater than those obtained by AQRSS in some datasets. Take, for instance, the “Dermatology” dataset; if δ = 0.06 , then approximation qualities derived by MCRSS, AQRSS, and CERSS are 0.9333, 0.8606, and 0.9151, respectively. Such results tell us that AQRSS is not always good in deriving higher approximation qualities.

4.2. Comparisons of Conditional Entropies

Figure 2 shows us the detailed results of conditional entropies with respect to three different algorithms for computing reducts.
In Figure 2, we can observe the following:
  • If the value of δ increases, then the increasing trends have been obtained for conditional entropies with respect to three different reducts, though those increasing trends are not strictly monotonic.
  • In most cases, there are slight differences between conditional entropies generated by MCRSS and CERSS, which can be attributed to the constraint designed by the measure of conditional entropy that has also been considered in MCRSS. Take, for instance, the “Breast Tissue" dataset; if δ = 0.15 , then the conditional entropies derived by MCRSS and CERSS are 0.6127 and 0.6599, respectively.
  • In most cases, the conditional entropies obtained by AQRSS are greater than those derived by both CERSS and MCRSS. This observation demonstrates that, if we only pay attention to the single measure of approximation quality, the obtained reduct may not be effective in terms of conditional entropy. Take, for instance, the “Forest-Type Mapping” dataset; if δ = 0.21 , then the conditional entropies derived by MCRSS, CERSS, and AQRSS are 0.4744, 0.5507, and 0.8951, respectively.

4.3. Comparisons of Classification Accuracies

In the following, the neighborhood classifier was used to measure the classification performances of the reducts derived from three different algorithms. The detailed results are shown in Figure 3.
In Figure 3, we can observe that the classification accuracies obtained by MCRSS are greater than those obtained by AQRSS and CERSS. Take the “Statlog (Vehicle Silhouettes)” dataset as an example; if δ = 0.06 , then the classification accuracies derived by MCRSS, AQRSS, and CERSS are 0.6205, 0.4352, and 0.6047, respectively. Such results tell us that the MCRSS algorithm provides better classification performance with the use of a neighborhood classifier.

4.4. Comparisons of Reduct Lengths

Figure 4 shows us the reduct lengths derived from three different algorithms.
InFigure 4, we can observe that the reduct lengths obtained by considering multiple criteria are greater than the lengths of reducts obtained by a single measure (approximation quality or conditional entropy). This is mainly because the multiple-criteria reduct in this experiment considers two measures, and then the constraint is stricter than the constraint defined by only one measure.

4.5. Comparisons of Time Consumptions

In the following, we compare the time consumption of several algorithms, AQRSS, CERSS, MCRSS, and MCR, in generating reducts. AQRSS and CERSS are used to find the approximation-quality reduct and conditional-entropy reduct with sample selection, respectively; MCRSS and MCR are used to find multiple-criteria reducts with and without sample selection, respectively. The detailed results are shown in Figure 5.
The following conclusions can be obtained from Figure 5.
  • The time consumption of MCRSS is higher than that of AQRSS and CERSS, though the time complexities of these three algorithms are the same. The reasons include two aspects: (1) MCRSS computes two attribute significances instead of one in each iteration; (2) the length of the reduct derived by MCRSS is frequently greater than those derived by AQRSS and CERSS, i.e., more iterations should be used.
  • By comparing it with the time consumption of MCR, MCRSS time consumption was significantly reduced. From this point of view, sample selection is effective in the process of finding a reduct from the viewpoint of saving time.

4.6. Comparisons of Core Attributes

In the following, we compare the three algorithms, AQRSS, CERSS and MCRSS, in the view of core attributes. For readers’ convenience, we only display the core attributes with one fixed radius; given δ = 0.15 , we use boundary samples to compute the core attributes [55,56], and the thinking of the process is similar to the algorithm proposed by Wang et al. [56]. We removed only one attribute from the raw attributes ( A T ) to make the subset that is made up of the remaining attributes that cannot satisfy the constraints in definition. Take the measure of “approximation quality” (AQRSS) as an example; A T = { a 1 , a 2 a n } : (1) remove a 1 in the first time, then the remaining attributes construct the subset A = { a 2 , a 3 a n } ; (2) compute γ ( A T , d ) and γ ( A , d ) in the new decision system D S ; (3) if γ ( A , d ) < γ ( A T , d ) , then a 1 can be a member of the core set. a 2 is removed in the second time and a n is removed in the n-th time. Similar algorithms are used to compute core sets for conditional entropy (CERSS) and multicriterion (MCRSS).
For readers’ convenience, to compare the results of these three algorithms (AQRSS, CERSS, MCRSS), the order of attributes was applied. When several consecutive attributes are core attributes, the order are listed with the symbol “-”.Take the dataset of “Hayes Roth”(ID: 6) as an example; since core attributes in terms of three algorithms are all { a 1 , a 2 , a 3 , a 4 } , which present “1-4” in Table 3. With a careful investigation of Table 3, the core of MCRSS is the union set of the core of AQRSS and CERSS in general. Take the data set of “Forest Type Mapping”(ID: 5) as an example, the core attributes of AQRSS and CERSS are { a 1 , a 6 , a 7 , a 16 , a 22 } and { a 1 , a 2 , a 4 , a 6 , a 15 , a 16 , a 22 , a 23 , a 25 } , respectively. And the core attributes are { a 1 , a 2 , a 4 , a 6 , a 7 , a 15 , a 16 , a 22 , a 23 , a 25 } , which is the union set of the core attributes of AQRSS and CERSS. In the “Dermatology” (ID: 4) dataset, this union-set relation is not correct since the core attributes of the three algorithms are { a 5 , a 15 } , { a 5 , a 15 , a 21 } and { a 5 , a 9 , a 15 , a 21 , a 22 } , respectively. More information is shown as follows.
Given δ = 0.15 , the results in Table 4 were obtained from all the data without using sample selection, and 5-CV was also applied to compute the mean values of these three measures (approximation quality, conditional entropy, and classification accuracy).
After a careful investigation of Table 4, it can be seen that MCRSS improves approximation quality and classification accuracy, and it also reduces the conditional entropy.
  • In the comparisons of these three measures, the largest values of approximation quality (classification accuracy) are in bold, and the smallest values of conditional entropy are underlined. It should be emphasized that in Datasets 6 and 10, the values of these three measures are the same. This is mainly because the cores of the three algorithms are the same, which can be seen from Table 3.
  • The results shown in Table 4 can generally stay consistent with the results shown above (Figure 1, Figure 2 and Figure 3), which are obtained from the datasets with sample selection. The reducts obtained by MCRSS can not only preserve approximation quality (Figure 1) and reduce conditional entropy (Figure 2), but also improve classification accuracy performance (Figure 3).
  • We can find that conditional entropy and approximation quality both have an important role in improving performance. The measure of conditional entropy may contribute a little more in improving classification accuracy values. The phenomenon that most of the values derived from the CERSS and MCRSS are the same may illustrate that the constraint of conditional entropy is more helpful in improving classification accuracy.

5. Conclusions and Future Perspectives

In this paper, a framework of a multiple-criteria reduct with sample selection has been proposed. Different from the traditional attribute-reduction algorithm that only uses one measure, our algorithm is executed based on the multiple criteria, which include approximation quality and conditional entropy. Experimental results show that the reduct computed by our algorithm can not only increase approximation quality and preserve conditional entropy, but also provide better classification performance. Since we also applied boundary samples instead of the whole samples in the data, our algorithm needed to spend less time in finding reducts.
The following topics merit further investigations:
  • Only two measures have been used to design multiple criteria; some other measures, such as classification accuracy [57] and neighborhood discrimination index [12], will be further added into the construction of multiple criteria.
  • Multiple-criteria attribute reduction is realized by a neighborhood rough set; it can also be introduced into other rough-set models, such as a fuzzy rough set [19] and decision-theoretic rough set [58].
  • Attribute reduction can be considered as the first step of data processing, and classification performances in terms of different classifiers [59,60] based on our reducts will be further explored.

Author Contributions

Project administration, X.Y.; software, P.W.; supervision, X.C.; writing—review and editing, Y.G.

Funding

This research was funded by the Natural Science Foundation of China (Nos. 61572242, 61502211, 61503160).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pawlak, Z. Rough Sets: Theoretical Aspects of Reasoning about Data; Kluwer Academic Publishers: Dordrecht, The Netherlands, 1991; ISBN 978-0792314721. [Google Scholar]
  2. Pawlak, Z.; Skowron, A. Rough sets: Some extensions. Inf. Sci. 2007, 177, 28–40. [Google Scholar] [CrossRef] [Green Version]
  3. Chen, H.M.; Li, T.R.; Luo, C.; Horng, S.J.; Wang, G. A decision-theoretic rough set approach for dynamic data mining. IEEE Trans. Fuzzy Syst. 2015, 23, 1–14. [Google Scholar] [CrossRef]
  4. Kaneiwa, K.; Kudo, Y. A sequential pattern mining algorithm using rough set theory. Int. J. Approx. Reason. 2011, 52, 881–893. [Google Scholar] [CrossRef]
  5. Hu, Q.H.; Yu, D.R.; Xie, Z.X.; Li, X. EROS: Ensemble rough subspaces. Pattern Recognit. 2007, 40, 3728–3739. [Google Scholar] [CrossRef]
  6. Dowlatshahi, M.B.; Derhami, V.; Nezamabadi, P.H. Ensemble of filter-based rankers to guide an epsilon-greedy swarm optimizer for high-dimensional feature subset selection. Information 2017, 8, 152. [Google Scholar] [CrossRef]
  7. Yao, Y.Y.; Zhao, Y. Attribute reduction in decision-theoretic rough set models. Inf. Sci. 2008, 178, 3356–3373. [Google Scholar] [CrossRef] [Green Version]
  8. Hu, Q.H.; Yu, D.R.; Xie, Z.X. Neighborhood classifiers. Expert Syst. Appl. 2008, 34, 866–876. [Google Scholar] [CrossRef]
  9. Dai, J.H.; Wang, W.T.; Xu, Q.; Tian, H. Uncertainty measurement for interval-valued decision systems based on extended conditional entropy. Knowl.-Based Syst. 2012, 27, 443–450. [Google Scholar] [CrossRef]
  10. Dai, J.H.; Xu, Q.; Wang, W.T.; Tian, H. Conditional entropy for incomplete decision systems and its application in data mining. Int. J. Gen. Syst. 2012, 41, 713–728. [Google Scholar] [CrossRef]
  11. Dai, J.H.; Wang, W.T.; Tian, H.W.; Liu, L. Attribute selection based on a new conditional entropy for incomplete decision systems. Knowl.-Based Syst. 2013, 39, 207–213. [Google Scholar] [CrossRef]
  12. Wang, C.Z.; Hu, Q.H.; Wang, X.Z.; Chen, D.; Qian, Y.; Dong, Z. Feature selection based on neighborhood discrimination index. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 2986–2999. [Google Scholar] [CrossRef] [PubMed]
  13. Angiulli, F. Fast nearest neighbor condensation for large data sets classification. IEEE Trans. Knowl. Data Eng. 2007, 19, 1450–1464. [Google Scholar] [CrossRef]
  14. Li, Y.H.; Maguire, L. Selecting critical patterns based on local geometrical and statistical information. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 1189–1201. [Google Scholar] [CrossRef] [PubMed]
  15. Nicolia, G.P.; Javier, P.R.; De, H.G.A. Oligois: Scalable instance selection for class-imbalanced data sets. IEEE Trans. Cybern. 2013, 43, 332–346. [Google Scholar] [CrossRef]
  16. Lin, W.C.; Tsai, C.F.; Ke, S.W.; Hung, C.W. Learning to detect representative data for large scale instance selection. J. Syst. Softw. 2015, 106, 1–8. [Google Scholar] [CrossRef]
  17. Zhai, J.H.; Wang, X.Z.; Pang, X.H. Voting-based instance selection from large data sets with mapreduce and random weight networks. Inf. Sci. 2016, 23, 1066–1077. [Google Scholar] [CrossRef]
  18. Zhai, J.H.; Li, T.; Wang, X.Z. A cross-selection instance algorithm. J. Intell. Fuzzy Syst. 2016, 3, 717–728. [Google Scholar] [CrossRef]
  19. Zhang, X.; Mei, C.L.; Chen, D.G.; Li, J. Feature selection in mixed data: a method using a novel fuzzy rough set-based information entropy. Pattern Recognit. 2016, 56, 1–15. [Google Scholar] [CrossRef]
  20. Xu, S.P.; Yang, X.B.; Yu, H.L.; Yang, J.; Tsang, E.C. Multi-label learning with label-specific feature reduction. Knowl.-Based Syst. 2016, 104, 52–61. [Google Scholar] [CrossRef]
  21. Yang, X.B.; Yao, Y.Y. Ensemble selector for attribute reduction. Appl. Soft Comput. 2018, 70, 1–11. [Google Scholar] [CrossRef]
  22. Ju, H.R.; Yang, X.B.; Song, X.N.; Qi, Y. Dynamic updating multigranulation fuzzy rough set: Approximations and reducts. Int. J. Mach. Learn. Cybern. 2014, 5, 981–990. [Google Scholar] [CrossRef]
  23. Yang, X.B.; Yu, D.J.; Yang, J.Y.; Wei, L. Dominance-based rough set approach to incomplete interval-valued information system. Data Knowl. Eng. 2009, 68, 1331–1347. [Google Scholar] [CrossRef]
  24. Yao, Y.Y. Relational interpretations of neighborhood operators and rough set approximation operators. Inf. Sci. 1998, 111, 239–259. [Google Scholar] [CrossRef] [Green Version]
  25. Yang, X.B.; Qian, Y.H.; Yang, J.Y. Hierarchical structures on multigranulation spaces. J. Comput. Sci. Technol. 2012, 27, 1169–1183. [Google Scholar] [CrossRef]
  26. Yang, X.B.; Qi, Y.S.; Song, X.N.; Yang, J. Test cost sensitive multigranulation rough set: Model and minimal cost selection. Inf. Sci. 2013, 250, 184–199. [Google Scholar] [CrossRef]
  27. Chen, D.G.; Wang, C.Z.; Hu, Q.H. A new approach to attribute reduction of consistent and inconsistent covering decision systems with covering rough sets. Inf. Sci. 2007, 177, 3500–3518. [Google Scholar] [CrossRef]
  28. Hu, Q.H.; Pedrycz, W.; Yu, D.R.; Lang, J. Selecting discrete and continuous features based on neighborhood decision error minimization. IEEE Trans. Syst. Man Cybern. B 2010, 40, 137–150. [Google Scholar] [CrossRef]
  29. Zhang, X.; Mei, C.L.; Chen, D.G.; Li, J. Multi-confidence rule acquisition and confidence-preserved attribute reduction in interval-valued decision systems. Int. J. Approx. Reason. 2014, 55, 1787–1804. [Google Scholar] [CrossRef]
  30. Hu, Q.H.; Che, X.J.; Zhang, L.; Guo, M.; Yu, D. Rank entropy based decision trees for monotonic classification. IEEE Trans. Knowl. Data Eng. 2012, 24, 2052–2064. [Google Scholar] [CrossRef]
  31. Liu, J.F.; Hu, Q.H.; Yu, D.R. A weighted rough set based method developed for class imbalance learning. Inf. Sci. 2008, 178, 1235–1256. [Google Scholar] [CrossRef]
  32. Guo, G.D.; Wang, H.; Bell, D.; Bi, Y.; Greer, K. KNN model-based approach in classification. Lect. Notes Comput. Sci. 2003, 2888, 986–996. [Google Scholar] [CrossRef]
  33. Li, S.Q.; Harner, E.J.; Adjeroh, D.A. Random knn feature selection—A fast and stable alternative to random forests. BMC Bioinform. 2011, 12, 450. [Google Scholar] [CrossRef] [PubMed]
  34. Sahigara, F.; Ballabio, D.; Todeschini, R.; Consonni, V. Defining a novel k-nearest neighbours approach to assess the applicability domain of a QSAR model for reliable predictions. J. Chem. 2013, 5, 27–36. [Google Scholar] [CrossRef] [PubMed]
  35. Lin, G.P.; Liang, J.Y.; Qian, Y.H. Uncertainty measures for multigranulation approximation space. Knowl.-Based Syst. 2015, 23, 443–457. [Google Scholar] [CrossRef]
  36. Li, M.M.; Zhang, X.Y. Information fusion in a multi-source incomplete information system based on information entropy. Entropy 2017, 19, 570. [Google Scholar] [CrossRef]
  37. Karevan, Z.; Suykens, J.A.K. Transductive feature selection using clustering-based sample etropy for temperature prediction in weather forecasting. Entropy 2018, 20, 264. [Google Scholar] [CrossRef]
  38. Ju, H.R.; Li, H.X.; Yang, X.B.; Zhou, X.; Huang, B. Cost-sensitive rough set: A multi-granulation approach. Knowl.-Based Syst. 2017, 123, 137–153. [Google Scholar] [CrossRef]
  39. Dou, H.L.; Yang, X.B.; Song, X.N.; Yu, H.; Wu, W.Z.; Yang, J. Decision-theoretic rough set: A multicost strategy. Knowl.-Based Syst. 2016, 91, 71–83. [Google Scholar] [CrossRef]
  40. Jia, X.Y.; Shang, L.; Zhou, B.; Yao, Y. Generalized attribute reduct in rough set theory. Knowl.-Based Syst. 2016, 91, 204–218. [Google Scholar] [CrossRef]
  41. Li, H.X.; Zhou, X.Z. Risk decision making based on decision-theoretic rough set: A three-way view decision model. Int. J. Comput. Intell. Syst. 2011, 4, 1–11. [Google Scholar] [CrossRef]
  42. Qian, Y.H.; Liang, J.Y.; Pedrycz, W.; Dang, C. Positive approximation: An accelerator for attribute reduction in rough set theory. Artif. Intell. 2010, 174, 597–618. [Google Scholar] [CrossRef]
  43. Qian, Y.H.; Liang, J.Y.; Pedrycz, W.; Dang, C. An efficient accelerator for attribute reduction from incomplete data in rough set framework. Pattern Recognit. 2011, 44, 1658–1670. [Google Scholar] [CrossRef]
  44. Jensen, R.; Shen, Q. Fuzzy-rough sets assisted attribute selection. IEEE Trans. Fuzzy Syst. 2007, 15, 73–89. [Google Scholar] [CrossRef] [Green Version]
  45. Li, J.Z.; Yang, X.B.; Song, X.N.; Li, J.; Wang, P.; Yu, D.J. Neighborhood attribute reduction: A multi-criterion approach. Int. J. Mach. Learn. Cybern. 2017, 1–12. [Google Scholar] [CrossRef]
  46. Dash, M.; Liu, H. Consistency-based search in feature selection. Artif. Intell. 2003, 151, 155–176. [Google Scholar] [CrossRef]
  47. Hu, Q.H.; Pan, W.W.; Zhang, L.; Zhang, D.; Song, Y.; Guo, M.; Yu, D. Feature selection for monotonic classification. IEEE Trans. Fuzzy Syst. 2012, 20, 69–81. [Google Scholar] [CrossRef]
  48. Wilson, D.R.; Martinez, T.R. Reduction techniques for instance-based learning algorithms. Mach. Learn. 2000, 38, 257–286. [Google Scholar] [CrossRef]
  49. Brighton, H.; Mellish, C. Advances in instance selection for instance-based learning algorithms. Data Min. Knowl. Discov. 2002, 6, 153–172. [Google Scholar] [CrossRef]
  50. Nikolaidis, K.; Goulermas, J.Y.; Wu, Q.H. A class boundary preserving algorithm for data condensation. Pattern Recognit. 2011, 44, 704–715. [Google Scholar] [CrossRef]
  51. Aldahdooh, R.T.; Ashour, W. DIMK-means distance-based initialization method for k-means clustering algorithm. Int. J. Intell. Syst. Appl. 2013, 5, 41–51. [Google Scholar] [CrossRef]
  52. Huang, K.Y. An enhanced classification method comprising a genetic algorithm, rough set theory and a modified PBMF-index function. Appl. Soft. Comput. 2012, 12, 46–63. [Google Scholar] [CrossRef]
  53. Lingras, P.; Chen, M.; Miao, D. Qualitative and quantitative combinations of crisp and rough clustering schemes using dominance relations. Int. J. Approx. Reason. 2014, 55, 238–258. [Google Scholar] [CrossRef]
  54. Yang, J.; Ma, Y.; Zhang, X.F.; Li, S.; Zhang, Y. An initialization method based on hybrid distance for k-means algorithm. Neural Comput. 2017, 29, 3094–3117. [Google Scholar] [CrossRef] [PubMed]
  55. Vashist, R.; Garg, M.L. Rule generation based on reduct and core: A rough set approach. Int. J. Comput. Appl. 2011, 29, 1–5. [Google Scholar] [CrossRef]
  56. Wang, G.Y.; Ma, X.A.; Yu, H. Monotonic uncertainty measures for attribute reduction in probabilistic rough set model. Int. J. Approx. Reason. 2015, 59, 41–67. [Google Scholar] [CrossRef]
  57. Peng, H.C.; Long, F.H.; Ding, C. Feature selection based on mutual information: Criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1226–1238. [Google Scholar] [CrossRef] [PubMed]
  58. Azam, N.; Yao, J.T. Game-theoretic rough sets for recommender systems. Knowl.-Based Syst. 2014, 72, 96–107. [Google Scholar] [CrossRef]
  59. Korytkowski, M.; Rutkowski, L.; Scherer, R. Fast image classification by boosting fuzzy classifiers. Inf. Sci. 2015, 327, 175–182. [Google Scholar] [CrossRef]
  60. Tsang, E.C.C.; Hu, Q.H.; Chen, D.G. Feature and instance reduction for PNN classfiers based on fuzzy rough sets. Int. J. Mach. Learn. Cybern. 2016, 7, 1–11. [Google Scholar] [CrossRef]
Figure 1. Comparisons of approximation qualities.
Figure 1. Comparisons of approximation qualities.
Information 09 00282 g001
Figure 2. Comparisons of conditional entropies.
Figure 2. Comparisons of conditional entropies.
Information 09 00282 g002
Figure 3. Comparisons of classification accuracies.
Figure 3. Comparisons of classification accuracies.
Information 09 00282 g003
Figure 4. Comparisons of reduct lengths.
Figure 4. Comparisons of reduct lengths.
Information 09 00282 g004
Figure 5. Comparisons of time consumption.
Figure 5. Comparisons of time consumption.
Information 09 00282 g005
Table 1. Decision-system example.
Table 1. Decision-system example.
a 1 a 2 a 3 a 4 a 5 a 6 a 7 d
x 1 0.81470.15760.65570.70600.43870.27600.75131
x 2 0.90580.97060.03570.03180.38160.67970.25511
x 3 0.12700.95720.84910.27690.76550.65510.50602
x 4 0.91340.48540.93400.04620.79520.16260.69912
x 5 0.63240.80030.67870.09710.18690.11900.89093
x 6 0.09750.14190.75770.82350.48980.49840.95933
x 7 0.27850.42180.74310.69480.44560.95970.54721
x 8 0.54690.91570.39220.31710.64630.34040.13862
x 9 0.95750.79220.65550.95020.70940.58530.14933
x 10 0.96490.95950.17120.03440.75470.22380.25752
Table 2. Descriptions of datasets.
Table 2. Descriptions of datasets.
IDData SetsSamplesAttributesDecision Classes
1Breast Cancer Wisconsin (Diagnostic)569302
2Breast Tissue10696
3Cardiotocography21262110
4Dermatology365346
5Forest-Type Mapping523274
6Hayes Roth13243
7Ionosphere351342
8Molecular Biology106572
9Statlog (Vehicle Silhouettes)846184
10Vertebral Column31062
11Wine178133
12Yeast1484810
Table 3. Core attributes.
Table 3. Core attributes.
IDDatasetsAQRSSCERSSMCRSS
1Breast Cancer Wisconsin (Diagnostic)2,8,10,282,8,20,28,292,8,20,28,29
2Breast Tissue2-32-3,92-3,9
3Cardiotocography1-2,4-5,8,17,212,4-5,8,10-12,181-2,4-5,8,10-12,18
4Dermatology5,155,15,215,9,15,21,22
5Forest-Type Mapping1,6-7,16,221-2,4,6,15-16,22-23,251-2,4,6-7,15-16,22-23,25
6Hayes Roth1-41-41-4
7Ionosphere1,3,12-14,26,30,32-341,3,11,15,17,19,21,321,3,11,15,17-22,30,32-34
8Molecular Biology13,38-39,48-49,5613,38-40,48-4913,38-40,48-49,56
9Statlog (Vehicle Silhouettes)4,10,14,16-181,4-6,8,10,14,16-181,4-6,8,10,14,16-18
10Vertebral Column2-3,52-3,52-3,5
11Wine2,5,7,10-131-2,4-5,7-8,10-131-2,4-5,7-8,10-13
12Yeast1-4,6-81-71-8
Table 4. Results obtained by using core attributes.
Table 4. Results obtained by using core attributes.
Approximation QualityConditional EntropyClassification Accuracy
IDAQRSSCERSSMCRSSAQRSSCERSSMCRSSAQRSSCERSSMCRSS
10.61730.63030.62600.16600.16610.16500.93320.95600.9578
20.08000.32000.32000.89630.50380.50380.12290.46280.4628
30.25400.24080.23520.58100.57180.57840.82640.88100.9045
40.99450.99180.97820.14500.08960.07580.52190.77600.8387
50.27830.31300.31300.47160.40950.42490.73420.85470.8910
60.45050.45050.45050.39130.39130.39130.80330.80330.8033
70.64410.56460.66670.43740.34420.43840.77490.83190.8319
80.40300.36970.40300.39190.44730.42850.67100.71820.7182
90.12980.14030.14030.76050.70130.70130.22110.57570.5757
100.48820.48820.48820.27990.27990.27990.87190.87190.8719
110.81420.84080.84080.10850.09660.09660.96650.97750.9775
120.05320.04430.04430.92160.92030.92030.52660.52130.5213
average0.43400.44950.45460.46250.41010.41700.66450.77130.7756

Share and Cite

MDPI and ACS Style

Gao, Y.; Chen, X.; Yang, X.; Wang, P. Neighborhood Attribute Reduction: A Multicriterion Strategy Based on Sample Selection. Information 2018, 9, 282. https://doi.org/10.3390/info9110282

AMA Style

Gao Y, Chen X, Yang X, Wang P. Neighborhood Attribute Reduction: A Multicriterion Strategy Based on Sample Selection. Information. 2018; 9(11):282. https://doi.org/10.3390/info9110282

Chicago/Turabian Style

Gao, Yuan, Xiangjian Chen, Xibei Yang, and Pingxin Wang. 2018. "Neighborhood Attribute Reduction: A Multicriterion Strategy Based on Sample Selection" Information 9, no. 11: 282. https://doi.org/10.3390/info9110282

APA Style

Gao, Y., Chen, X., Yang, X., & Wang, P. (2018). Neighborhood Attribute Reduction: A Multicriterion Strategy Based on Sample Selection. Information, 9(11), 282. https://doi.org/10.3390/info9110282

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop