Next Article in Journal
Information-Theoretic Analysis of Cardio-Respiratory Interactions in Heart Failure Patients: Effects of Arrhythmias and Cardiac Resynchronization Therapy
Previous Article in Journal
Comparison of Bootstrap Methods for Estimating Causality in Linear Dynamic Systems: A Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Online Multi-Label Streaming Feature Selection Based on Label Group Correlation and Feature Interaction

1
Department of Computer Science and Technology, Huaqiao University, Xiamen 361021, China
2
Xiamen Key Laboratory of Computer Vision and Pattern Recognition, Huaqiao University, Xiamen 361021, China
3
Fujian Key Laboratory of Big Data Intelligence and Security, Huaqiao University, Xiamen 361021, China
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(7), 1071; https://doi.org/10.3390/e25071071
Submission received: 10 June 2023 / Revised: 10 July 2023 / Accepted: 14 July 2023 / Published: 17 July 2023
(This article belongs to the Section Information Theory, Probability and Statistics)

Abstract

:
Multi-label streaming feature selection has received widespread attention in recent years because the dynamic acquisition of features is more in line with the needs of practical application scenarios. Most previous methods either assume that the labels are independent of each other, or, although label correlation is explored, the relationship between related labels and features is difficult to understand or specify. In real applications, both situations may occur where the labels are correlated and the features may belong specifically to some labels. Moreover, these methods treat features individually without considering the interaction between features. Based on this, we present a novel online streaming feature selection method based on label group correlation and feature interaction (OSLGC). In our design, we first divide labels into multiple groups with the help of graph theory. Then, we integrate label weight and mutual information to accurately quantify the relationships between features under different label groups. Subsequently, a novel feature selection framework using sliding windows is designed, including online feature relevance analysis and online feature interaction analysis. Experiments on ten datasets show that the proposed method outperforms some mature MFS algorithms in terms of predictive performance, statistical analysis, stability analysis, and ablation experiments.

1. Introduction

Multi-label feature selection (MFS) plays a crucial role in addressing the preprocessing of high-dimensional multi-label data. Numerous methods have been proposed and proven to be effective in improving prediction performance and model interpretability. However, traditional MFS methods assume that all features are collected and presented to the learning model beforehand [1,2,3,4]. This assumption does not align with many practical application scenarios where not all features are available in advance. In video recognition, for example, each frame may possess important features that become available over time. Hence, achieving real-time feature processing has emerged as a significant concern [5,6,7,8].
Online multi-label feature selection with streaming features is an essential branch of MFS that facilitates the efficient real-time management of streaming features. It provides significant advantages, such as low time and space consumption, particularly when dealing with extremely high-dimensional datasets. Some notable works in this have attracted attention, including online multi-label streaming feature selection based on a neighborhood rough set (OM-NRS) [9], multi-label streaming feature selection (MSFS) [10], and novel streaming feature selection(ASFS) [11]. However, these methods primarily focus on eliminating irrelevant and/or redundant features. In addition to identifying irrelevant and/or redundant features, feature interaction is crucial but often overlooked. Feature interaction refers to features that have weak or independent correlations with the label, but when combined with other features, they may exhibit a strong association with the predicted label. Streaming feature selection with feature interaction (SFS-FI) [12] is a representative approach that considers feature interaction dynamically. SFS-FI has successfully identified the impact of feature interaction; however, it lacks the capability to tackle the learning challenge in multi-label scenarios.
Another difficulty with online MFS is that labels are universally correlated, which is a distinctive property of multi-label data [13,14,15,16]. Intuitively, a known label can aid in learning an unknown one, and the co-occurrence of two labels may provide additional information to the model. For example, an image with ‘grassland’ and ‘lion’ labels is likely also to be marked as ‘African’; similarly, ‘sea’ and ‘ship’ labels would tend to appear together in a short video, while ‘train’ and ‘sea’ labels tend not to appear together. Some research work has been carried out around label correlation. Representative work includes multi-label streaming feature selection (MSFS) and online multi-label streaming feature selection with label correlation ( O M S F S L C ). MSFS captures label correlation by constructing a new data representation pattern for label space and utilizes the constructed label relationship matrix to examine the merits of features. O M S F S L C constructs label weights by calculating label correlation, and, on this basis, integrates label weights into significance analysis and relevance analysis of streaming features. The methods mentioned above select features by evaluating the relationship between features and the global label space. This strategy may not be optimal as it is challenging to comprehend and specify the relationship between relevant labels and features. Based on research by Li et al. [17], it has been found that strongly correlated labels tend to share similar specific features, while weakly related labels typically have distinct features. In line with this observation, this paper will categorize related labels into two groups: strongly related labels will be grouped together, while weakly related labels will be separated into different groups.
Accordingly, a novel online multi-label streaming feature selection based on label group correlation and feature interaction, namely OSLGC, is proposed to select pertinent and interactive features from streaming features. Firstly, our method involves calculating the correlation matrix of labels and using graph theory to group related labels. Labels within the same group have a strong correlation, while labels from different groups have a weak correlation. Then, we define the feature relevance item and integrate the label weight and feature interaction weight into the feature relevance item. Subsequently, a framework based on sliding windows is established, which iteratively processes streaming features through two steps: online feature correlation analysis and online feature interaction analysis. Finally, extensive experiments demonstrate that OSLGC can yield significant performance improvements compared to other mature MFS methods. The uniqueness of OSLGC is established as follows:
  • By utilizing graph theory, label groups are constructed to ensure that closely associated labels are grouped together. This method provides an effective means of visualizing the relationships among labels.
  • We provide a formal definition of feature interaction and quantify the impact of feature interaction under different label groups. Based on this, OSLGC is capable of selecting features with interactivity.
  • A novel streaming feature selection framework using sliding windows is proposed, which resolves the online MFS problem by simultaneously taking feature interaction, label importance, as well as label group correlation, into account.
  • Experiments on ten datasets demonstrate that the proposed method is competitive with existing mature MFS algorithms in terms of predictive performance, statistical analysis, stability analysis, and ablation experiments.
The rest of this article is arranged as follows: In Section 2, we review previous research. Section 3 provides the relevant preparatory information. In Section 4, we present the detailed procedure for OSLGC. In Section 5, we report the empirical study. Finally, Section 6 sums up the work of this paper and considers the prospects, priorities and direction of future research.

2. Related Work

Multi-label feature selection (MFS), as a widely known data preprocessing method, has achieved promising results in different application fields, such as emotion classification [18], text classification [19], and gene detection [20]. Depending on whether the features are sufficiently captured in advance, existing MFS methods can be divided into batch and online methods.
The batch method assumes that the features presented to learning are pre-available. Generally speaking, it can be further subdivided into several types according to the characteristics provided by the complex label space, including missing labels [21,22], label distribution [23,24], label selection [25], label imbalance [26,27], streaming labels [28,29], partial labels [30,31], label-specific features [32,33,34], and label correlation [35,36,37]. Among them, investigating label correlation is considered to be a favorable strategy to promote the performance of learning. To date, many works have focused on this. For instance, label supplementation for multi-label feature selection (LSMFS) [38] evaluates the relationship between labels using mutual information provided by the features. Quadratically constrained linear programming (QCLP) [39] introduces a matrix variable normal prior distribution to model label correlation. By minimizing the label ranking loss of label correlation regularization, QCLP is able to identify a feature subset. On the other hand, label-specific features emphasize that different labels may possess their own specific features. One of the most representative studies, Label specIfic FeaTures (LIFT) [32], has shown that using label-specific features to guide the MFS process can elevate the performance and interpretability of learning tasks. Recently, group-preserving label-specific feature selection (GLFS) [33] has been used to exploit label-specific features and common features with l2,1-norm regularization to support the interpretability of the selected features.
The online method differs from the batch method in that features are generated on-the-fly and feature selection takes place in real-time as the features arrive. Based on the characteristics of the label space, it can be categorized into two groups: label independence and label correlation. For label independence, several methods have been proposed, such as the streaming feature selection algorithm with dynamic sliding Windows and feature repulsion loss (SF-DSW-FRL) [40], multi-objective online streaming multi-label feature selection using mutual information (MOML) [41], streaming feature selection via class-imbalance aware rough set (SFSCI) [42], online multi-label group feature selection (OMGFS) [43], and multi-objective multi-label-based online feature selection (MMOFS) [44]. Similar to the static MFS methods, the online MFS approach also focuses on exploring label correlation. For instance, MSFS [10] uses the relationship between samples and labels to construct a new data representation model to measure label correlation, and implements feature selection by designing feature correlation and redundancy analysis. Multi-label online streaming feature selection with mutual information (ML-OSMI) [45] uses high-order methods to determine label correlation, and combines spectral granulation and mutual information to evaluate streaming features. Unfortunately, existing methods cannot exactly capture the impact of label relationships on the evaluation of streaming features and are hindered by a time-consuming calculation procedure. Thus, online multi-source streaming features selection (OMSFS) [7] investigates label correlation by calculating mutual information, and, on this basis, constructs the weight for each label and designs a significance analysis to accelerate the computational efficiency.
Based on our review of previous studies, we find that with the arrival of each new feature, existing methods can be effective in dealing with streaming features. However, these methods pay more attention to the contribution of features to all labels, and do not explore the specific relationship between features and labels. To put it simply, they fail to consider that highly correlated labels may have common features, while weakly correlated labels may have distinct features. Additionally, it is important to mention that most previous works have focused on selecting the most relevant features to labels, but have ignored the potential contribution of feature interactions to labels. In contrast, our framework pays close attention to feature interactions and label group correlation, and seeks to explore the specific features and label group weights corresponding to the label group.

3. Preliminaries

3.1. Multi-Label Learning

Given a multi-label information table M L S = < U , F , L > , where U = { x 1 , x 2 , , x n } is a non-empty instance set, F = { f 1 , f 2 , , f d } and L = { l 1 , l 2 , , l m } are a feature set and label set used to describe instances, respectively. l i ( x k ) represents the value of label l i on instance x k U , where l i ( x k ) = 1 , only if x k possesses l i , and 0 otherwise. The task of multi-label learning is to learn a function h : U 2 L .

3.2. Basic Information-Theoretic Notions

This section introduces some basic information theory concepts which are commonly used in the evaluation of feature quality.
Definition 1.
Let X = { x 1 , x 2 , , x n } be a discrete random variable and P ( x i ) be the probability of x i , then the entropy of X is
H ( X ) = x i X P ( x i ) l o g P ( x i ) .
H ( X ) is a measure of randomness or uncertainty in the distribution of X. It is at a maximum when all the possible values of X are equal, and at a minimum when X takes only one value with probability 1.
Definition 2.
Let Y = { y 1 , y 2 , , y m } be another random variable. Then, the joint entropy H ( X , Y ) of X and Y is:
H ( X , Y ) = x i X y j Y P ( x i , y j ) l o g P ( x i , y j ) ,
where P ( x i , y j ) denotes the joint probability of x i and y j .
Definition 3.
Given X and Y, when the variable Y is known, the residual uncertainty of X can be determined by the conditional entropy H ( X | Y ) :
H ( X | Y ) = x i X y j Y P ( x i , y j ) l o g P ( x i | y j ) ,
where P ( x i | y j ) is the conditional probability of x i given y j .
Definition 4.
Given X and Y, then the amount of information shared by two variables can be determined by the mutual information I ( X ; Y ) :
I ( X ; Y ) = x i X y j Y P ( x i , y j ) l o g P ( y j | x i ) P ( y j ) .
The larger the I ( X ; Y ) value, the stronger the correlation between the two variables. Inversely, the two variables are independent if I ( X ; Y ) = 0 .
Definition 5.
Given variables X, Y and Z, when Z is given, the uncertainty of X due to the known Y can be measured by the conditional mutual information I ( X ; Y Z ) :
I ( X ; Y Z ) = x i X y j Y z k Z P ( x i , y j , z k ) log P ( x i y j , z k ) P ( x i y j ) .

4. The OSLGC Method

4.1. Exploiting Label Group Correlation

To investigate label group correlation, in this subsection, we introduce a graph-based method to further distinguish the relevant labels, which can effectively differentiate relevant labels by grouping strongly related labels together and separating weakly related ones into different groups. The process involves two fundamental steps: (1) constructing an undirected graph of the labels, and (2) partitioning the graph to create distinct label groups.
In the first step, OSLGC aims to construct an undirected graph that effectively captures the label correlation among all labels, thus providing additional information for streaming feature evaluation. For this purpose, it is necessary to investigate the correlation between labels.
Definition 6.
Given < U , F , L > , x k U , l i , l j L , l i ( x k ) represents the value of label l i with respect to instance x k , the correlation r i j between the labels is defined as:
r i j = x k U P ( l i ( x k ) , l j ( x k ) ) log P ( l i ( x k ) , l j ( x k ) ) P ( l i ( x k ) ) P ( l j ( x k ) ) .
Obviously, if  l i and l j are independent, then r i j = 0 ; otherwise, r i j > 0 .
Using Equation (6), the label correlation matrix M ( R L C ) is obtained, and the form of M ( R L C ) is shown below.
M ( R L C ) = r 11 r 12 r 1 m r 21 r 22 r 2 m r m 1 r m 2 r m m .
Based on the matrix, the weighted undirected graph of the label correlation can be structured G r a p h = ( V , E ) , where V = { l i | l i L i [ 1 , m ] } and E = { ( l i , l j ) | l i , l j L , i , j [ 1 , m ] , i j } mean the vertex and edge of G r a p h , respectively. As  M ( R L C ) is symmetric, G r a p h is an undirected graph that reflects the correlation among all labels. But, regrettably, G r a p h has m vertices and m ( m 1 ) / 2 edges. For ultra-high-dimensional data, the density of the graph will be considerable, which often leads to strong edge interweaving of different weights. Moreover, the resolution of complete graphs is an NP-hard problem. Therefore, for  G r a p h , it is necessary to reduce the edges of G r a p h .
In the second step, OSLGC aims to divide the graph and create label groups. With this intent, we first generate a minimum spanning tree ( M S T ) through the Prim algorithm. M S T has the same vertices as G r a p h and  partial edges of G r a p h . The weight of the link edge in the M S T is expressed as W ( l i , l j ) , which is essentially different for different edges. To divide strongly correlated labels into groups, we set the threshold to break the edges below the threshold in M S T .
Definition 7.
Given W ( l i , l j ) represents the weight of the edges, and the threshold for weak label correlation is defined as:
δ = ( l i , l j ) M S T W ( l i , l j ) m 1 .
δ is the average of the edge weights, which is used to divide the label groups, thereby putting the strongly related labels in the same group.
Concretely, if  W ( l i , l j ) δ , which means that the relationship between labels l i and l j is a strength label correlation, we will reserve the edge that connects l i with l j . If W ( l i , l j ) < δ , which explains that the relationship between labels l i and l j is a weakness label correlation, we can remove the edge that connects l i with l j from MST. Hence, the MST can be segmented into forests by threshold segmentation. In the forest, the label nodes within each subtree are strongly correlated, while the label nodes between subtrees are weakly correlated. Based on this, we can treat each subtree as a label group, denoted as L = { L G 1 L G 2 L G p } .
Example 1.
A multi-label dataset is presented in Table 1. First, the label correlation matrix is calculated using Equation (6), as follows:
M ( R L C ) = 1.00 0.13 0.05 0.61 0.02 0.26 0.13 1.00 0.13 0.28 0.00 0.13 0.05 0.13 1.00 0.00 0.02 0.02 0.61 0.28 0.00 1.00 0.00 0.13 0.02 0.00 0.02 0.00 1.00 0.26 0.26 0.13 0.02 0.13 0.26 1.00 .
Then, we can create the label undirected graph by using the label correlation matrix, as shown in Figure 1a. Immediately afterwards, the minimum spanning tree is generated by the Prim algorithm, as shown in Figure 1b. Finally, the threshold δ of MST is calculated using Equation (7), and the edges that meet condition W ( l i , l j ) < δ are removed, as shown in Figure 1c.

4.2. Analysis Feature Interaction under Label Group

As a rule, the related labels generally share some label-specific features [17,33], i.e., labels within the same label group may share the same label-specific features. Thus, to generate label-specific features for different label groups, in this subsection, we will further analyze feature relationships under different label groups, including feature independence, feature redundancy, and feature interaction. We also give the interaction weight factor to quantify the influence degree of the feature relationship under different label groups.
Definition 8
(Feature independence). Given a set of label groups L = { L G 1 L G 2 L G p } , L G h L , S t = { f 1 , f 2 , , f d * } denotes the selected features, and  f t is a new incoming feature at time t. For  f i S t , f i and f t are referred to as feature independence under L G h if, and only if:
I ( f i ; L G h ) + I ( f t ; L G h ) = I ( f i , f t ; L G h ) .
According to Definition 8, I ( f i ; L G h ) + I ( f t ; L G h ) = I ( f i , f t ; L G h ) suggests that the information provided by feature f i and f t for the label group L G h are non-interfering, i.e., the features are independent of each other under label group L G h .
Theorem 1.
If I ( f t ; L G h | f i ) = I ( f t ; L G h ) or I ( f i ; L G h | f t ) = I ( f i ; L G h ) , then f i and f t are independent under label group L G h .
Proof. 
I ( f i , f t ; L G h ) = I ( f i ; L G h ) + I ( f t ; L G h | f i ) = I ( f t ; L G h ) + I ( f i ; L G h | f t ) . If  I ( f i ; L G h | f t ) = I ( f i ; L G h ) or I ( f t ; L G h | f i ) = I ( f t ; L G h ) , I ( f i , f t ; L G h ) = I ( f i ; L G h ) + I ( f t ; L G h ) . Thus, f i and f t are independent under label group L G h .    □
Theorem 2.
If f i and f t are independent, under the condition that label group L G h is known, then I ( f i ; f t | L G h ) = 0 .
Proof. 
If f i and f t are independent, i.e.,  I ( f i ; f t ) = 0 , according to Definition 5, it can be proven that I ( f i ; f t | L G h ) = 0 .
   □
Definition 9
(Feature redundancy). Given a set of label groups L = { L G 1 L G 2 L G p } , L G h L , S t = { f 1 , f 2 , , f d * } denotes the selected features, and  f t is a new incoming feature. For  f i S t , f i and f t are referred to as feature redundancy under L G h if, and only if:
I ( f i ; L G h ) + I ( f t ; L G h ) > I ( f i , f t ; L G h ) .
Equation (9) suggests that there is partial duplication of information provided by two features; that is, the amount of information brought by two features f i and f t together for label group L G h is less than the sum of the information brought by the two features for L G h separately.
Theorem 3.
If I ( f i ; L G h | f t ) < I ( f i ; L G h ) or I ( f t ; L G h | f i ) < I ( f t ; L G h ) , then the relationship between f i and f t is a pair of feature redundancy under label group L G h .
Proof. 
I ( f i , f t ; L G h ) = I ( f i ; L G h ) + I ( f t ; L G h | f i ) = I ( f t ; L G h ) + I ( f i ; L G h | f t ) . If  I ( f i ; L G h | f t ) < I ( f i ; L G h ) or I ( f t ; L G h | f i ) < I ( f t ; L G h ) , I ( f i , f t ; L G h ) < I ( f i ; L G h ) + I ( f t ; L G h ) . Thus, the relationship between f i and f t is a pair of feature redundancy under label group L G h .    □
Definition 10
(Feature interaction). Given a set of label groups L = { L G 1 L G 2 L G p } , L G h L , S t = { f 1 , f 2 , , f d * } denotes the selected features, and  f t is a new incoming feature. For  f i S t , f i and f t are referred to as a feature interaction under L G h if, and only if:
I ( f i ; L G h ) + I ( f t ; L G h ) < I ( f i , f t ; L G h ) .
Equation (10) suggests that there is a synergy between features f i and f t together for label group L G h ; that is, they yield more information together for label group L G h than what could be expected from the sum of I ( f i ; L G h ) and I ( f t ; L G h ) .
Theorem 4.
If I ( f i ; L G h | f t ) > I ( f i ; L G h ) or I ( f t ; L G h | f i ) > I ( f t ; L G h ) , then f i and f t is a pair of feature interaction under label group L G h .
Proof. 
I ( f i , f t ; L G h ) = I ( f i ; L G h ) + I ( f t ; L G h | f i ) = I ( f t ; L G h ) + I ( f i ; L G h | f i ) . If I ( f i ; L G h | f t ) > I ( f i ; L G h ) or I ( f t ; L G h | f i ) > I ( f t ; L G h ) , I ( f i , f t ; L G h ) > I ( f i ; L G h ) + I ( f t ; L G h ) . Thus, f i and f t are a pair of feature positive interaction under label group L G h .    □
Property 1.
If two features f i and f t are not independent, the correlations between f i and f t under a different label group L G h are distinct. It is easy to show with Example 2.
Example 2.
Continue Table 1. As shown in Table 2, we can see that I ( f 1 , f 2 ; L G 1 ) = 0.997 is less than I ( f 1 ; L G 1 ) + I ( f 2 ; L G 1 ) = 1.227 , and, according to Definition 9, f 1 and f 2 is a feature redundancy under label group L G 1 . However, for label group L G 3 , it satisfies that I ( f 1 , f 2 ; L G 3 ) > I ( f 1 ; L G 3 ) + I ( f 2 ; L G 3 ) ; that is, f 1 and f 2 is a feature interaction under the label group L G 3 . This finding suggests that the relationship between f 1 and f 2 changes dynamically under different label groups.
Consequently, to evaluate features accurately, it is imperative to quantify the influence of the feature relationships on feature relevance. That is, the inflow of a new feature f t has a positive effect in predicting labels, and we should enlarge the weight of feature f t ; otherwise, the weight of feature f t should be reduced. The feature interaction weight factor is defined to quantize the impact of the feature relationships as follows:
Table 2. The relationship between features.
Table 2. The relationship between features.
Mutual InformationCombinationFeature Relationship
I ( f 1 ; L G 1 ) = 0.771 I ( f 1 , f 2 ; L G 1 ) = 0.997 Feature redundancy
I ( f 2 ; L G 1 ) = 0.446
I ( f 1 ; L G 3 ) = 0.020 I ( f 1 , f 2 ; L G 3 ) = 0.171 Feature interaction
I ( f 2 ; L G 3 ) = 0.020
Definition 11
(Feature Interaction Weight). Given a set of label groups L = { L G 1 L G 2 L G p } , L G h L , S t = { f 1 , f 2 , , f d * } denotes the selected features, and  f t is a new incoming feature. For  f i S t , the feature interaction weight between f i and f t is defined as:
F W ( f i , f t , L G h ) = I ( f i , f t ; L G h ) I ( f i ; L G h ) + I ( f t ; L G h ) .
F W ( f i , f t , L G h ) offers additional information for evaluating feature f t . If feature f t and the selected feature f i S t is independent or redundant, it holds that F W ( f i , f t , L G h ) 1 . However, if the feature relationship is interactive, it holds that F W ( f , f t , L G h ) > 1 .

4.3. Streaming Feature Selection with Label Group Correlation and Feature Interaction

Streaming features refer to features acquired over time; however, in fact, not all features obtained dynamically are helpful for prediction. Therefore, it is necessary to extract valuable features from the streaming features’ environment. To achieve this purpose, in this paper, we implement the analysis of streaming features in two stages: online feature relevance analysis and online feature interaction analysis.

4.3.1. Online Feature Relevance Analysis

The purpose of feature relevance analysis is to select features that are important to the label groups. Correspondingly, the feature relevance is defined as follows:
Definition 12
(Feature Relevance). Given label groups L = { L G 1 L G 2 L G p } , f t is a new incoming feature, the feature relevance item is defined as:
γ ( f t ) = h = 1 p I ( f t ; L G h ) × W ( L G h ) .
In which, W ( L G h ) denotes the weight assigned to each label group, and W ( L G h ) = H ( L G h ) j = 1 p H ( L G j ) where H ( L G h ) is the information entropy of label group L G h . The higher the weight of the label group, the more important the label group is to other label groups. In other words, the corresponding label-specific features of the label group should have higher feature importance.
Definition 13.
Given label groups L = { L G 1 L G 2 L G p } , f t is a new incoming feature, and  γ ( f t ) is the feature relevance. With a pair of thresholds α and β ( 0 < α < β ), we define:
(1) f t is strongly relevant, if  β γ ( f t ) ;
(2) f t is weakly relevant, if  α < γ ( f t ) < β ;
(3) f t is irrelevant, if  0 γ ( f t ) α .
In general, for a new incoming feature f t , if  f t is powerfully relevant, we will select it; if f t is irrelevant, we will directly abandon it and no longer consider it later; if f t is weakly relevant, there is a risk of greater misjudgment in making a decision immediately, including selecting or abandoning, and the best approach is to obtain more information to make a decision.

4.3.2. Online Feature Interaction Analysis

Definition 13 can be used to make intuitive judgments about features that are weakly correlated. However, Definition 13 does not provide a basis for selecting or abandoning weakly relevant features. Therefore, it is necessary to determine whether to remove or retain the weakly relevant features.
Definition 14.
Given label groups L = { L G 1 L G 2 L G p } , S t = { f 1 , f 2 , , f d * } denotes the selected features, and  f t is a new incoming feature. The feature relevance when considering feature interaction, called the enhanced feature relevance, is defined as:
𝟊 ( f t ) = 1 | S t | i = 1 d * h = 1 p I ( f t ; L G h ) × W ( L G h ) × F W ( f i , f t ; L G h ) .
In which, F W ( f i , f t ; L G h ) is the feature interaction weight between f t and f i S t . Furthermore, to determine whether to retain the weakly relevant feature, we set the mean value of feature relevance about the selected features as the relevance threshold, as follows:
Definition 15.
Given S t = { f 1 , f 2 , , f d * } denotes the selected features, f i S t , at time t, the mean value of the feature relevance about the selected features is:
M e a n t = i = 1 d * γ ( f i ) | S t | .
Obviously, when 𝟊 ( f t ) > M e a n t , it shows that the weak relevant feature f t interacts with the selected features. In this case, f t can enhance the prediction ability and be selected as an effective feature. Otherwise, when 𝟊 ( f t ) M e a n t , it denotes that adding the weakly relevant feature f t does not promote the prediction ability for labels, and, in this case, we can discard the feature f t .

4.3.3. Streaming Feature Selection Strategy Using Sliding Windows

According to Definition 13, two main issues need to be addressed: (1) how to design a streaming feature selection mechanism to discriminate the newly arrived features; (2) how to set proper thresholds for α and β .
(1) Streaming feature selection with sliding windows: To solve the first challenge, a sliding window mechanism is proposed to receive the arrived features in a timed sequence, which is consistent with the dynamic nature of the streaming features. The specific process can be illustrated using the example in Figure 2.
  • First, the sliding window (SW) continuously receives and saves the arrived features. When the number of features in the sliding window reaches the preset size, the features in the window are discriminated, which includes decision-making with regard to selection, abandonment, or delay.
  • According to the feature relevance γ ( f t ) (Definition 12), we select the strongly relevant features, as shown in Figure 2. We can straightforwardly select strongly relevant features, e.g.,  f 15 and f 18 . Similarly, for the irrelevant features, they are discarded from the sliding window, e.g.,  f 16 and f 17 .
  • For weakly relevant features, we need to further analyze the enhanced feature relevance by considering the feature interaction. If the weakly relevant features satisfy the condition 𝟊 ( f t ) > M e a n t , they can be selected, e.g.,  f 19 ; otherwise, the weakly relevant features are retained in the sliding window, for example, f 14 , and new features are awaited to flow into the sliding window.
This process is performed repeatedly. That is, when the features in the sliding window reach saturation or no new features appear, the next round of feature analysis is performed.
Figure 2. Streaming feature selection with sliding window.
Figure 2. Streaming feature selection with sliding window.
Entropy 25 01071 g002
(2) Thresholds setting of α and β : To solve the second challenge, we assume that the experimental data follow a normal distribution and the streaming features arrive randomly. Inspired by the 3 σ principle of normal distribution, we set α and β as the mean and standard deviation of features in the sliding window.
Definition 16.
Given a sliding window S W , f t S W , and  γ ( f t ) is the feature relevance, then, at time t, the mean value μ t of the sliding window is:
μ t = f t S W γ ( f t ) | S W | .
Definition 17.
Given a sliding window S W , f t S W , and  γ ( f t ) is the feature relevance, then, at time t, the standard deviation σ t of the sliding window is:
σ t = f t S W ( γ ( f t ) μ t ) 2 | S W | .
Therefore, we combine the 3 σ principle of normally distributed data to redefine the three feature relationships.
Definition 18.
Given γ ( f t ) is the feature relevance, at time t, μ t and σ t are the mean and standard deviation of the features in the sliding window. Then, we define three feature relationships as:
(1) f t is strongly relevant, if  μ t + σ t γ ( f t ) ;
(2) f t is weakly relevant, if  μ t σ t < γ ( f t ) < μ t + σ t ;
(3) f t is irrelevant, if  0 γ ( f t ) μ t σ t .
Through the above analysis, we propose a novel algorithm, named OSLGC, as shown in Algorithm 1.
Algorithm 1 The OSLGC algorithm
  • Input: S W : sliding window, f i : predictive features, L: label set.
  • Output: S t : the feature subset at time t.
    1:
    Generate label groups L = { L G 1 L G 2 L G p } by Section 4.1;
    2:
    repeat
    3:
       Get a new feature f t at time t;
    4:
       Add feature f t to the sliding window S W ;
    5:
       while  S W is full or no features are available do
    6:
         Compute μ t , γ t , and M e a n t ;
    7:
         for each f t S W  do
    8:
            if  γ ( f t ) μ t + σ t  then
    9:
              if  𝟊 ( f t ) > M e a n t  then
    10:
                  S t = S t f t ;
    11:
              end if
    12:
            else
    13:
              Discard f t ;
    14:
            end if
    15:
         end for
    16:
       end while
    17:
    until No features are available;
    18:
    Return S t ;
The major computation in OSLGC is feature analysis in sliding windows (Steps 5–16). Assuming | F t | is the number of currently arrived features, and | L | is the number of labels, in the best-case scenario, OSLGC obtains a feature subset after running online feature relevance analysis, and the time complexity is O ( | F t | · | L | ) . However, in many cases, the features are not simply strongly relevant or irrelevant, but include weakly relevant instances. Therefore, online feature interaction analysis needs to be further performed. The final time complexity is O ( | F t | · | F t | · | L | ) .

5. Experiments

5.1. Data Sets

We conducted experiments on ten multi-label datasets, which were mainly from three different domains, that is, text, audio, and images, respectively. Among them, the first eight datasets (i.e., Business, Computer, Education, Entertainment, Health, Entertainment, Reference, and Society) were taken from Yahoo, and were derived from the actual web text classification. For audio, Birds is an audio dataset that identifies 19 species of birds. For images, Scene includes 2407 images with up to six labels per image. These datasets are freely available for public download and have been widely used in research on multi-label learning.
Detailed information is provided in Table 3. For each dataset S, we use | S | , F ( S ) , and L ( S ) to represent the number of instances, number of features, and number of labels, respectively. L C a r d ( S ) denotes the average number of labels per example, and L D e n ( S ) standardizes L C a r d ( S ) according to the number of possible labels. In addition, it is worth noting that the number of instances and the number of labels in different datasets vary from 645 to 5000 and from 6 to 33, respectively. These datasets with varied properties provide a solid foundation for algorithm testing.

5.2. Experimental Setting

To visualize the performance of OSLGC, we compared OSLGC with several recent MFS algorithms. For a reasonable comparison, two different types of algorithms were selected as comparison algorithms, including (1) two online multi-label streaming feature selection algorithms, and (2) five MFS methods based on information theory. Specifically, the two online multi-label streaming feature selection methods included multi-label streaming feature selection (MSFS) [10] and online multi-label feature selection based on neighborhood rough set (OMNRS) [9]. On the other hand, the five MFS methods based on information theory were multi-label feature selection with label dependency and streaming labels (MSDS) [16], multi-label feature selection with streaming labels (MLFSL) [28], label supplementation for multi-label feature selection (LSMFS) [38], maximum label supplementation for multi-label feature selection (MLSMFS) [38], and constraint regression and adaptive spectral graph (CSASG [46]), respectively. Details of these algorithms are provided below.
  • MSDS: It acquires features by exploring the feature significance, label significance, and label specific features, simultaneously.
  • LSMFS: It leverages label relationships to extract all feature supplementary information for each label from other labels.
  • MLSMFS: It is similar to LSMFS, but it maximizes the feature supplementary information of each label from other labels.
  • MSFS: It realizes streaming feature selection by conducting online relevance and redundancy analysis.
  • OMNRS: It sets the bounds of pairwise correlation between features to discard redundant features.
  • MLFSL: It is an MFS algorithm based on streaming labels, which fuses the feature rankings by minimizing the overall weighted deviation.
  • CSASG: It proposes a multi-label feature selection framework, which incorporates a spectral graph term based on information entropy into the manifold framework.
For the proposed method, the size of the sliding window | S W | is set to 15 in this paper. For the algorithms that obtain the feature subset, e.g., MSDS, MSFS, and OMNRS, we use the feature subset obtained by these algorithms to construct new data for prediction. For the algorithms that obtain feature ranking, e.g., MLFSL, LSMFS, MLSMFS, and CSASG, the first p features are selected, which depends on the dimension of the feature subset obtained by the OSLGC algorithm. Furthermore, we select the average precision (AP), Hamming loss (HL), one error (OE), and macro-F1 (F1), as the evaluation metrics. Due to space limitations, information on these metrics will not be provided in detail. The formulas and descriptions of all the evaluation metrics are provided in [47,48]. Finally, MLkNN (k = 10) is selected as the basic classifier.

5.3. Experimental Results

Table 4, Table 5, Table 6 and Table 7 display the results for the different evaluation metrics, where the symbol “↓ (↑)” indicates “the smaller (larger), the better”. Boldface highlights the best prediction performance for a specific dataset, and the penultimate row in each table shows the average value of the algorithm on all datasets. Furthermore, the Win/Draw/Loss record provides the number of datasets where OSLGC outperforms, performs equally to, and underperforms compared to the other algorithms, respectively. The experimental results indicate that OSLGC exhibits strong competitiveness compared with other algorithms; the experimental results also provide some interesting insights.
  • For web text data, OSLGC is capable of achieving the best predictive performance on at least 7 out of the 8 datasets on all the evaluation metrics. This suggests that the proposed method is suitable for selecting features for web text data.
  • For the Birds and Scene data, OSLGC achieves the best result on 3 out of 4 evaluation metrics. For the remaining evaluation metric, OSLGC ranks second with a disadvantage of 0.96 % and 1.51 %, respectively. This result indicates that OSLGC can also be applied to the classification problem of other data types, such as images and audio.
  • The average prediction results of all datasets were comprehensively investigated, with the results showing that the performance of OSLGC has obvious advantages. Furthermore, the Win/Draw/Loss records clearly demonstrate that OSLGC can outperform the other algorithms.
  • Although MSFS, OMNRS, and OSLGC are proposed to manage streaming features, the performance advantage of OSLGC confirms that label group correlation and feature interaction can provide additional information for processing streaming features.
OSLGC is able to make use of label group correlation to guide feature selection, and adds online feature interaction analysis to provide hidden information for predictive labels. By combining the potential contributions of the feature space and the label space, OSLGC performs very competitively compared to other mature MFS methods.
Table 4. Results for different algorithms on Average Precision (↑).
Table 4. Results for different algorithms on Average Precision (↑).
Average PrecisionMSDSMLSMFSLSMFSMSFSOMNRSMLFSLCSASGOSLGC
Business0.87480.87050.87070.86670.87460.87500.87550.8782
Computer0.64100.63290.63280.63030.64200.62600.63970.6458
Education0.55380.55150.53190.54750.55470.54780.55990.5636
Entertainment0.56170.54870.56850.56320.57040.56260.56490.5809
Health0.68830.66170.67010.68150.68940.65510.70130.7040
Recreation0.49040.46280.47740.49210.49910.44590.48240.5083
Reference0.62380.62320.62050.62520.63320.61700.62900.6324
Society0.59320.59020.56980.59610.58490.58620.59760.5983
Birds0.48770.46030.45630.51810.48420.46140.52600.5317
Scene0.83720.83310.83310.67560.83750.84280.84300.8451
Average0.63520.62350.62310.61960.63700.62200.64190.6488
Win/Draw/Loss10/0/010/0/010/0/010/0/09/0/110/0/010/0/0-
Table 5. Results for different algorithms on Hamming Loss (↓).
Table 5. Results for different algorithms on Hamming Loss (↓).
Hamming LossMSDSMLSMFSLSMFSMSFSOMNRSMLFSLCSASGOSLGC
Business0.02760.02810.02830.02830.02740.02750.02760.0273
Computer0.03990.04040.04010.04000.03970.04170.03980.0396
Education0.04000.04100.04130.04130.04080.04100.04030.0398
Entertainment0.06160.06240.06150.06210.06170.06150.06060.0600
Health0.04330.04470.04500.04340.04170.04560.04140.0408
Recreation0.06070.06170.06140.06130.05950.06350.06030.0593
Reference0.03110.03110.02910.03040.02940.03110.03060.0301
Society0.05550.05590.05850.05530.05650.05750.05530.0549
Birds0.05130.05360.04970.04840.05180.05130.04890.0463
Scene0.09880.10020.10020.16370.10140.10190.10090.0957
Average0.05100.05190.05150.05740.05100.05230.05060.0494
Win/Draw/Loss10/0/010/0/010/0/010/0/09/0/110/0/010/0/0-
Table 6. Results for different algorithms on One Error (↓).
Table 6. Results for different algorithms on One Error (↓).
One ErrorMSDSMLSMFSLSMFSMSFSOMNRSMLFSLCSASGOSLGC
Business0.12400.13230.13230.13600.12470.12300.12330.1187
Computer0.42730.43870.43870.44570.43300.45830.43130.4197
Education0.56730.59100.60970.58930.58100.59070.57670.5653
Entertainment0.59470.60570.58370.59100.57830.59030.58800.5620
Health0.39670.43830.41400.41570.40430.44970.38600.3747
Recreation0.65570.68970.67300.65170.64130.71570.66670.6277
Reference0.47130.48430.47100.46970.45270.48400.46130.4650
Society0.45000.45870.47800.44730.46130.46770.44530.4427
Birds0.62790.66860.65120.55810.62210.66280.53490.5465
Scene0.26760.27420.27420.50840.26670.25670.25500.2525
Average0.45820.47820.47260.48130.45650.47990.44690.4375
Win/Draw/Loss10/0/010/0/010/0/010/0/09/0/110/0/09/0/1-
Table 7. Results for different algorithms on Macro_F (↑).
Table 7. Results for different algorithms on Macro_F (↑).
Macro_FMSDSMLSMFSLSMFSMSFSOMNRSMLFSLCSASGOSLGC
Business0.16020.13870.13260.12100.08520.14650.15910.1668
Computer0.09960.09020.07030.06120.07140.07270.09100.0919
Education0.13290.13470.10880.13260.07860.13120.12840.1288
Entertainment0.12590.09600.11620.11450.11250.11680.13360.1391
Health0.22700.16820.15370.19970.16230.16320.24350.2261
Recreation0.10330.08540.07900.08340.12430.05560.10190.1307
Reference0.11910.11110.11720.11260.07470.11800.12510.1286
Society0.08550.06140.03460.08550.05810.04420.07710.0791
Birds0.07660.04120.04600.08550.04340.05030.05730.1129
Scene0.71770.69210.69210.35640.68590.70190.68340.7026
Average0.18480.16190.15500.13520.14960.16000.18000.1906
Win/Draw/Loss9/0/19/0/110/0/010/0/010/0/010/0/010/0/0-

5.3.1. Statistical Tests

To assess the statistical significance of the observed differences between the eight algorithms, we used the Friedman test [49]. The Friedman test ranks the prediction performance obtained by each dataset. The best algorithm ranks first, the sub-optimal algorithm ranks second, and so on. For K algorithms and N datasets, r j i represents the rank of the i algorithm on the j dataset, and R i = 1 / N j = 1 N r j i represents the average rank of the i algorithm. Under the null hypothesis (i.e., all algorithms are equivalent), the Friedman statistic F F obeys the Fisher distribution of degrees of freedom ( K 1 ) and ( K 1 ) ( N 1 ) :
F F = ( N 1 ) χ F 2 N ( K 1 ) χ F 2 , where χ F 2 = 12 N K ( K + 1 ) ( i = 1 K R i 2 K ( K + 1 ) 2 4 ) .
Table 8 summarizes the value of F F and the corresponding critical value. Based on the Friedman test, the null hypothesis is rejected at a significance level of 0.10. Consequently, it is necessary to use a post hoc test to further analyze the relative performance between the algorithms. As the experiments focus on the performance difference between OSLGC and other algorithms, we chose the Bonferroni–Dunn test [50] to serve this purpose. In this test, the performance difference between OSLGC and one comparison algorithm is compared using the critical difference (CD), and C D α = q α · K ( K + 1 ) 6 N , where q α = 2.450 at α = 0.10; thus, we can compute C D 0.1 = 2.6838 .
Figure 3 gives the CD diagrams, where the average rank of each algorithm is plotted on the coordinate axis. The best performance rank is on the rightmost side of the coordinate axis, and conversely, the worst rank is on the leftmost side of the coordinate axis. In each subfigure, if the average rank between OSLGC and one comparison algorithm is connected by a CD line, it indicates that the performance of the two algorithms is comparable and statistically indistinguishable. Otherwise, if the average rank of a comparison algorithm is outside a CD, it is considered to have a significantly different performance from OSLGC.
From Figure 3, we can observe that: (1) OSLGC has obvious advantages over LSMFS, MLSMFS, MLFSL, and MSFS with respect to all the evaluation metrics; (2) OSLGC achieves comparable performance with CSASG for each evaluation metric, but, different from the setting of the known static feature space of CSASG, OSLGC selects features by assuming the dynamic arrival of features, which entails a process of selecting the best feature with local feature information; (3) It is noteworthy that, although OSLGC cannot be significantly distinguished from all the algorithms, OSLGC exhibits significant advantages over the other feature selection algorithms. In summary, OSLGC exhibits a stronger statistical performance than LSMFS, MLSMFS, MLFSL, MSFS, MSDS, OMNRS, and CSASG.

5.3.2. Stability Analysis

In this subsection, we employ spiderweb plots to verify the stability of the algorithms. Because the results generated by the algorithm on different evaluation metrics are quite different, to reasonably compare, we standardize the prediction results to a standard range [0.1, 0.5]. The spiderweb diagram has the following characteristics: (1) The larger the area surrounded by the same color line, the better the performance and the stability of the algorithm. (2) The closer the normalized value is to 0.5, the better the performance. (3) The closer the shape of the encircling line corresponding to the algorithm is to a regular polygon, the better the stability of the algorithm. Figure 4 shows spider diagrams for all the evaluation metrics, where each corner denotes a dataset and different colored lines represent different MFS algorithms, respectively.
By analyzing Figure 4, it is found that: (1) Among all the algorithms, the area surrounded by OSLGC is the largest, which indicates that OSLGC has the best performance; (2) The polygon enclosed by OSLGC is approximately a regular polygon with respect to the average precision and macro-F1. This indicates that the performance obtained by OSLGC is relatively stable on different datasets; (3) Furthermore, although the polygon enclosed by OSLGC is not a regular polygon with respect to the Hamming loss and one error metrics, the fluctuation range of OSLGC at each vertex is relatively small. In summary, compared with the other algorithms, the OSLGC algorithm has obvious advantages in terms of performance and stability.

5.3.3. Ablation Experiment

To evaluate the contribution of the label group correlation, we conducted an ablation empirical study by removing the label group correlation in Algorithm 1 and derived a variant of the OSLGC algorithm, called the OSLGC-RLC algorithm. Table 9 displays the results for OSLGC and OSLGC-RLC. Due to space limitations, we select three datasets for experimental verification: Recreation, Entertainment, and Social. Considering the results in Table 9, it is observed that OSLGC significantly outperforms OSLGC-RLC on all the evaluation metrics. In conclusion, the above results suggest that considering the label group correlation is an effective strategy in feature selection.

6. Conclusions

In this paper, we have presented a new online multi-label streaming feature selection method, called OSLGC, to select relevant or interactive features from streaming features. In OSLGC, a set of trees is constructed using graph theory that is able to divide strongly related labels into the same tree, and which applies a streaming feature selection strategy using sliding windows, which identifies the relevant, interactive, and irrelevant features in an online manner. OSLGC can be divided into two parts: online feature relevance analysis and online feature interaction analysis. For online feature relevance analysis, we designed the feature relevance terms to provide a basis for decision-making, such as for selection, delay, and abandonment. For online feature interaction analysis, we defined an enhanced feature relevance item that prefers to select a group of interactive features from the delay decisions corresponding to the online relevance analysis. Based on experiments undertaken, our research showed that OSLGC achieved a high level of competitive performance against other advanced competitors.
In future work, we intend to combine label-specific features and common features to design streaming feature selection strategies. Furthermore, we are committed to building streaming feature selection strategies that are suitable for large-scale data.

Author Contributions

Conceptualization, J.L.; Methodology, J.L.; Software, J.L. and S.Y.; Validation, J.L.; Investigation, Z.S.; Resources, Z.S. and J.D.; Data curation, J.L.; Writing—original draft, J.L.; Writing—review & editing, J.L.; Visualization, S.Y.; Supervision, H.Z. and J.D.; Project administration, H.Z.; Funding acquisition, H.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by grants from the National Natural Science Foundation of China (Nos. 61871196, 61976120, and 62076116), the Guiding Project of Fujian Science and Technology Plan (No. 2021H0019), the Natural Science Foundation of Fujian Province (Nos. 2021J02049, and 2022J01317), and, in part by the Fundamental Research Funds for the Central Universities of Huaqiao University under Grant ZQN-1115. All authors have read and agreed to the published version of the manuscript.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Real-world datasets that we use in our experiments are publicly available. These data can be found here: http://www.lamda.nju.edu.cn/code_MDDM.ashx and https://mulan.sourceforge.net/datasets-mlc.html.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, J.; Lin, Y.; Jiang, M.; Li, S.; Tang, Y.; Tan, K.C. Multi-label Feature Selection via Global Relevance and Redundancy Optimization. In Proceedings of the IJCAI, Yokohama, Japan, 11–17 July 2020; pp. 2512–2518. [Google Scholar]
  2. Li, Y.; Hu, L.; Gao, W. Multi-label feature selection via robust flexible sparse regularization. Pattern Recognit. 2023, 134, 109074. [Google Scholar] [CrossRef]
  3. Huang, J.; Qian, W.; Vong, C.M.; Ding, W.; Shu, W.; Huang, Q. Multi-Label Feature Selection via Label Enhancement and Analytic Hierarchy Process. IEEE Trans. Emerg. Top. Comput. Intell. 2023, 1–17. [Google Scholar] [CrossRef]
  4. Wang, J.; Lin, Y.; Li, L.; Wang, Y.a.; Xu, M.; Chen, J. Multi-label causal feature selection based on neighbourhood mutual information. Int. J. Mach. Learn. Cybern. 2022, 13, 3509–3522. [Google Scholar] [CrossRef]
  5. Wu, X.; Yu, K.; Ding, W.; Wang, H.; Zhu, X. Online feature selection with streaming features. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 1178–1192. [Google Scholar]
  6. Hu, X.; Zhou, P.; Li, P.; Wang, J.; Wu, X. A survey on online feature selection with streaming features. Front. Comput. Sci. 2018, 12, 479–493. [Google Scholar] [CrossRef]
  7. You, D.; Li, R.; Liang, S.; Sun, M.; Ou, X.; Yuan, F.; Shen, L.; Wu, X. Online causal feature selection for streaming features. IEEE Trans. Neural Netw. Learn. Syst. 2021, 34, 1563–1577. [Google Scholar] [CrossRef]
  8. Gomes, H.M.; Read, J.; Bifet, A.; Barddal, J.P.; Gama, J. Machine learning for streaming data: State of the art, challenges, and opportunities. ACM SIGKDD Explor. Newsl. 2019, 21, 6–22. [Google Scholar] [CrossRef]
  9. Liu, J.; Lin, Y.; Li, Y.; Weng, W.; Wu, S. Online multi-label streaming feature selection based on neighborhood rough set. Pattern Recognit. 2018, 84, 273–287. [Google Scholar] [CrossRef]
  10. Lin, Y.; Hu, Q.; Liu, J.; Li, J.; Wu, X. Streaming feature selection for multilabel learning based on fuzzy mutual information. IEEE Trans. Fuzzy Syst. 2017, 25, 1491–1507. [Google Scholar] [CrossRef]
  11. Liu, J.; Lin, Y.; Du, J.; Zhang, H.; Chen, Z.; Zhang, J. ASFS: A novel streaming feature selection for multi-label data based on neighborhood rough set. Appl. Intell. 2023, 53, 1707–1724. [Google Scholar] [CrossRef]
  12. Zhou, P.; Li, P.; Zhao, S.; Wu, X. Feature interaction for streaming feature selection. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4691–4702. [Google Scholar] [CrossRef] [PubMed]
  13. Zhu, Y.; Kwok, J.T.; Zhou, Z.H. Multi-label learning with global and local label correlation. IEEE Trans. Knowl. Data Eng. 2017, 30, 1081–1094. [Google Scholar] [CrossRef] [Green Version]
  14. Che, X.; Chen, D.; Mi, J. Feature distribution-based label correlation in multi-label classification. Int. J. Mach. Learn. Cybern. 2021, 12, 1705–1719. [Google Scholar] [CrossRef]
  15. Zhang, J.; Li, C.; Cao, D.; Lin, Y.; Su, S.; Dai, L.; Li, S. Multi-label learning with label-specific features by resolving label correlations. Knowl.-Based Syst. 2018, 159, 148–157. [Google Scholar] [CrossRef]
  16. Liu, J.; Lin, Y.; Ding, W.; Zhang, H.; Wang, C.; Du, J. Multi-label feature selection based on label distribution and neighborhood rough set. Neurocomputing 2023, 524, 142–157. [Google Scholar] [CrossRef]
  17. Li, J.; Li, P.; Hu, X.; Yu, K. Learning common and label-specific features for multi-label classification with correlation information. Pattern Recognit. 2022, 121, 108259. [Google Scholar] [CrossRef]
  18. You, M.; Liu, J.; Li, G.Z.; Chen, Y. Embedded feature selection for multi-label classification of music emotions. Int. J. Comput. Intell. Syst. 2012, 5, 668–678. [Google Scholar] [CrossRef] [Green Version]
  19. Ma, Y.; Liu, X.; Zhao, L.; Liang, Y.; Zhang, P.; Jin, B. Hybrid embedding-based text representation for hierarchical multi-label text classification. Expert Syst. Appl. 2022, 187, 115905. [Google Scholar] [CrossRef]
  20. Pereira, R.B.; Plastino, A.; Zadrozny, B.; Merschmann, L.H. Categorizing feature selection methods for multi-label classification. Artif. Intell. Rev. 2018, 49, 57–78. [Google Scholar] [CrossRef]
  21. Huang, J.; Qin, F.; Zheng, X.; Cheng, Z.; Yuan, Z.; Zhang, W.; Huang, Q. Improving multi-label classification with missing labels by learning label-specific features. Inf. Sci. 2019, 492, 124–146. [Google Scholar] [CrossRef]
  22. Tan, A.; Ji, X.; Liang, J.; Tao, Y.; Wu, W.Z.; Pedrycz, W. Weak multi-label learning with missing labels via instance granular discrimination. Inf. Sci. 2022, 594, 200–216. [Google Scholar] [CrossRef]
  23. Xiong, C.; Qian, W.; Wang, Y.; Huang, J. Feature selection based on label distribution and fuzzy mutual information. Inf. Sci. 2021, 574, 297–319. [Google Scholar] [CrossRef]
  24. Qian, W.; Xiong, Y.; Yang, J.; Shu, W. Feature selection for label distribution learning via feature similarity and label correlation. Inf. Sci. 2022, 582, 38–59. [Google Scholar] [CrossRef]
  25. Lee, J.; Kim, D.W. Efficient multi-label feature selection using entropy-based label selection. Entropy 2016, 18, 405. [Google Scholar] [CrossRef] [Green Version]
  26. Liu, B.; Blekas, K.; Tsoumakas, G. Multi-label sampling based on local label imbalance. Pattern Recognit. 2022, 122, 108294. [Google Scholar] [CrossRef]
  27. Tarekegn, A.N.; Giacobini, M.; Michalak, K. A review of methods for imbalanced multi-label classification. Pattern Recognit. 2021, 118, 107965. [Google Scholar] [CrossRef]
  28. Lin, Y.; Hu, Q.; Zhang, J.; Wu, X. Multi-label feature selection with streaming labels. Inf. Sci. 2016, 372, 256–275. [Google Scholar] [CrossRef]
  29. Liu, J.; Lin, Y.; Ding, W.; Zhang, H.; Du, J. Fuzzy Mutual Information-Based Multilabel Feature Selection With Label Dependency and Streaming Labels. IEEE Trans. Fuzzy Syst. 2022, 31, 77–91. [Google Scholar] [CrossRef]
  30. Qian, W.; Li, Y.; Ye, Q.; Ding, W.; Shu, W. Disambiguation-based partial label feature selection via feature dependency and label consistency. Inf. Fusion 2023, 94, 152–168. [Google Scholar] [CrossRef]
  31. Wang, J.; Li, P.; Yu, K. Partial Multi-Label Feature Selection. In Proceedings of the 2022 International Joint Conference on Neural Networks (IJCNN), Padua, Italy, 18–23 July 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–9. [Google Scholar]
  32. Zhang, M.L.; Wu, L. Lift: Multi-label learning with label-specific features. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 37, 107–120. [Google Scholar] [CrossRef] [Green Version]
  33. Zhang, J.; Wu, H.; Jiang, M.; Liu, J.; Li, S.; Tang, Y.; Long, J. Group-preserving label-specific feature selection for multi-label learning. Expert Syst. Appl. 2023, 213, 118861. [Google Scholar] [CrossRef]
  34. Lin, Y.; Hu, Q.; Liu, J.; Zhu, X.; Wu, X. MULFE: Multi-label learning via label-specific feature space ensemble. ACM Trans. Knowl. Discov. Data (TKDD) 2021, 16, 1–24. [Google Scholar] [CrossRef]
  35. Zhang, P.; Liu, G.; Song, J. MFSJMI: Multi-label feature selection considering join mutual information and interaction weight. Pattern Recognit. 2023, 138, 109378. [Google Scholar] [CrossRef]
  36. Zhang, P.; Gao, W.; Hu, J.; Li, Y. Multi-label feature selection based on high-order label correlation assumption. Entropy 2020, 22, 797. [Google Scholar] [CrossRef] [PubMed]
  37. Wang, Z.; Wang, T.; Wan, B.; Han, M. Partial classifier chains with feature selection by exploiting label correlation in multi-label classification. Entropy 2020, 22, 1143. [Google Scholar] [CrossRef] [PubMed]
  38. Zhang, P.; Liu, G.; Gao, W.; Song, J. Multi-label feature selection considering label supplementation. Pattern Recognit. 2021, 120, 108137. [Google Scholar] [CrossRef]
  39. Gu, Q.; Li, Z.; Han, J. Correlated multi-label feature selection. In Proceedings of the 20th ACM International Conference on Information and Knowledge Management, Glasgow, UK, 24–28 October 2011; pp. 1087–1096. [Google Scholar]
  40. Li, Y.; Cheng, Y. Streaming feature selection for multi-label data with dynamic sliding windows and feature repulsion loss. Entropy 2019, 21, 1151. [Google Scholar] [CrossRef] [Green Version]
  41. Rafie, A.; Moradi, P.; Ghaderzadeh, A. A multi-objective online streaming multi-label feature selection using mutual information. Expert Syst. Appl. 2023, 216, 119428. [Google Scholar] [CrossRef]
  42. Zou, Y.; Hu, X.; Li, P.; Li, J. Multi-label streaming feature selection via class-imbalance aware rough set. In Proceedings of the 2021 International Joint Conference on Neural Networks (IJCNN), Shenzhen, China, 18–22 July 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–9. [Google Scholar]
  43. Liu, J.; Lin, Y.; Wu, S.; Wang, C. Online multi-label group feature selection. Knowl.-Based Syst. 2018, 143, 42–57. [Google Scholar] [CrossRef]
  44. Paul, D.; Jain, A.; Saha, S.; Mathew, J. Multi-objective PSO based online feature selection for multi-label classification. Knowl.-Based Syst. 2021, 222, 106966. [Google Scholar] [CrossRef]
  45. Wang, H.; Yu, D.; Li, Y.; Li, Z.; Wang, G. Multi-label online streaming feature selection based on spectral granulation and mutual information. In Proceedings of the Rough Sets: International Joint Conference, IJCRS 2018, Quy Nhon, Vietnam, 20–24 August 2018; Proceedings 6. Springer: Berlin/Heidelberg, Germany, 2018; pp. 215–228. [Google Scholar]
  46. Fan, Y.; Liu, J.; Weng, W.; Chen, B.; Chen, Y.; Wu, S. Multi-label feature selection with constraint regression and adaptive spectral graph. Knowl.-Based Syst. 2021, 212, 106621. [Google Scholar] [CrossRef]
  47. Zhang, Y.; Zhou, Z.H. Multilabel dimensionality reduction via dependence maximization. ACM Trans. Knowl. Discov. Data (TKDD) 2010, 4, 1–21. [Google Scholar] [CrossRef]
  48. Zhang, M.L.; Zhou, Z.H. A review on multi-label learning algorithms. IEEE Trans. Knowl. Data Eng. 2013, 26, 1819–1837. [Google Scholar] [CrossRef]
  49. Friedman, M. A comparison of alternative tests of significance for the problem of m rankings. Ann. Math. Stat. 1940, 11, 86–92. [Google Scholar] [CrossRef]
  50. Dunn, O.J. Multiple comparisons among means. J. Am. Stat. Assoc. 1961, 56, 52–64. [Google Scholar] [CrossRef]
Figure 1. The relationship among labels. (a) label correlation matrix, (b) minimum spanning tree, and (c) label groups.
Figure 1. The relationship among labels. (a) label correlation matrix, (b) minimum spanning tree, and (c) label groups.
Entropy 25 01071 g001
Figure 3. The CD diagrams using the Bonferroni–Dunn test. (a) Average precision, (b) Hamming loss, (c) One error, and (d) Macro-F1.
Figure 3. The CD diagrams using the Bonferroni–Dunn test. (a) Average precision, (b) Hamming loss, (c) One error, and (d) Macro-F1.
Entropy 25 01071 g003
Figure 4. Spiderweb diagrams showing the stability of the algorithm. (a) Average precision, (b) Hamming loss, (c) One error, and (d) Macro-F1.
Figure 4. Spiderweb diagrams showing the stability of the algorithm. (a) Average precision, (b) Hamming loss, (c) One error, and (d) Macro-F1.
Entropy 25 01071 g004
Table 1. Example of multi-label data.
Table 1. Example of multi-label data.
Instance f 1 f 2 l 1 l 2 l 3 l 4 l 5 l 6
x 1 10+1−1+1+ 1−1+1
x 2 10+1−1+1+1−1+ 1
x 3 01−1+1−1−1+1−1
x 4 01−1+1+1−1+1−1
x 5 11−1−1+1−1−1+1
x 6 10+1−1−1+1+1−1
x 7 01+1+1−1−1−1+1
x 8 01−1+1+1−1−1−1
x 9 10+1+1−1+1−1+1
x 10 11+1−1+1+1+ 1+1
Table 3. Detailed description of datasets.
Table 3. Detailed description of datasets.
Dataset | S | F ( S ) L ( S ) LCard ( S ) LDen ( S ) Domain
Business5000438301.5990.053Text
Computer5000681331.5070.046Text
Education5000550331.4630.044Text
Entertainment5000640211.4140.067Text
Health5000612321.6620.052Text
Recreation5000606221.4230.065Text
Reference5000793331.1690.035Text
Society5000636271.670.062Text
Birds645260191.0140.053Audio
Scene240729461.0740.179Image
Table 8. Friedman statistics F F and critical value.
Table 8. Friedman statistics F F and critical value.
Evaluation Metric F F Critical Value ( α = 0.10 )
Average Precision15.26971.74
Hamming Loss9.0301
One Error13.5067
Macro-F19.2081
Table 9. Results between OSLGC and OSLGC-RLC.
Table 9. Results between OSLGC and OSLGC-RLC.
Evaluation MetricRecreationEntertainmentSocial
OSLGCOSLGC-RLCOSLGCOSLGC-RLCOSLGCOSLGC-RLC
Average Precision0.50830.49960.58090.57130.71260.7036
Hamming Loss0.05930.06010.06000.06030.02500.0252
One Error0.62770.64400.56200.57830.37670.3957
Macro-F10.13070.11290.13910.13750.15000.1400
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, J.; Yang, S.; Zhang, H.; Sun, Z.; Du, J. Online Multi-Label Streaming Feature Selection Based on Label Group Correlation and Feature Interaction. Entropy 2023, 25, 1071. https://doi.org/10.3390/e25071071

AMA Style

Liu J, Yang S, Zhang H, Sun Z, Du J. Online Multi-Label Streaming Feature Selection Based on Label Group Correlation and Feature Interaction. Entropy. 2023; 25(7):1071. https://doi.org/10.3390/e25071071

Chicago/Turabian Style

Liu, Jinghua, Songwei Yang, Hongbo Zhang, Zhenzhen Sun, and Jixiang Du. 2023. "Online Multi-Label Streaming Feature Selection Based on Label Group Correlation and Feature Interaction" Entropy 25, no. 7: 1071. https://doi.org/10.3390/e25071071

APA Style

Liu, J., Yang, S., Zhang, H., Sun, Z., & Du, J. (2023). Online Multi-Label Streaming Feature Selection Based on Label Group Correlation and Feature Interaction. Entropy, 25(7), 1071. https://doi.org/10.3390/e25071071

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop