1
|
Xu W, Huang M, Jiang Z, Qian Y. Graph-Based Unsupervised Feature Selection for Interval-Valued Information System. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:12576-12589. [PMID: 37067967 DOI: 10.1109/tnnls.2023.3263684] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Feature selection has become one of the hot research topics in the era of big data. At the same time, as an extension of single-valued data, interval-valued data with its inherent uncertainty tend to be more applicable than single-valued data in some fields for characterizing inaccurate and ambiguous information, such as medical test results and qualified product indicators. However, there are relatively few studies on unsupervised attribute reduction for interval-valued information systems (IVISs), and it remains to be studied how to effectively control the dramatic increase of time cost in feature selection of large sample datasets. For these reasons, we propose a feature selection method for IVISs based on graph theory. Then, the model complexity could be greatly reduced after we utilize the properties of the matrix power series to optimize the calculation of the original model. Our approach can be divided into two steps. The first is feature ranking with the principles of relevance and nonredundancy, and the second is selecting top-ranked attributes when the number of features to keep is fixed as a priori. In this article, experiments are performed on 14 public datasets and the corresponding seven comparative algorithms. The results of the experiments verify that our algorithm is effective and efficient for feature selection in IVISs.
Collapse
|
2
|
Yang S, Zhang H, Shi G, Zhang Y. Attribute reductions of quantitative dominance-based neighborhood rough sets with A-stochastic transitivity of fuzzy preference relations. Appl Soft Comput 2023. [DOI: 10.1016/j.asoc.2023.109994] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/09/2023]
|
3
|
Wang P, Qu L, Zhang Q. Information entropy based attribute reduction for incomplete heterogeneous data. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS 2022. [DOI: 10.3233/jifs-212037] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Attribute reduction in an information system (IS) is an important research topic in rough set theory (RST). This paper investigates attribute reduction for incomplete heterogeneous data based on information entropy. Information entropy in an incomplete IS with heterogeneous data (IISH) is first defined. Then, some derived notions of information entropy, such as joint information entropy, conditional information entropy, mutual information entropy, gain and gain ratio in an incomplete decision IS with heterogeneous data (IDISH), are presented. Next, information entropy is applied to perform attribute reduction in an IDISH. Two attribute reduction algorithms, based on gain and gain ratio, respectively, are proposed. Finally, in order to illustrate the feasibility and efficiency of the proposed algorithms, experimental analysis is carried out and comparisons are done. It is worth mentioning that the incomplete rate is used to deal with incomplete heterogeneous data.
Collapse
Affiliation(s)
- Pei Wang
- Key Laboratory of Complex System Optimization and Big Data Processing in Department of Guangxi Education, Yulin Normal University, Yulin, Guangxi, P.R. China
| | - Liangdong Qu
- School of Artificial Intelligence, Guangxi University for Nationalities, Nanning, Guangxi, P.R. China
| | - Qinli Zhang
- School of Big Data and Artificial Intelligence, Chizhou University, Chizhou, Anhui, P.R. China
| |
Collapse
|
4
|
Yang L, Qin K, Sang B, Xu W. Dynamic fuzzy neighborhood rough set approach for interval-valued information systems with fuzzy decision. Appl Soft Comput 2021. [DOI: 10.1016/j.asoc.2021.107679] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
5
|
Liu X, Dai J, Chen J, Zhang C. A fuzzy α-similarity relation-based attribute reduction approach in incomplete interval-valued information systems. Appl Soft Comput 2021. [DOI: 10.1016/j.asoc.2021.107593] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|
6
|
Sang B, Chen H, Yang L, Li T, Xu W, Luo C. Feature selection for dynamic interval-valued ordered data based on fuzzy dominance neighborhood rough set. Knowl Based Syst 2021. [DOI: 10.1016/j.knosys.2021.107223] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
|
7
|
Li L. Feature selection for incomplete set-valued data. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS 2021. [DOI: 10.3233/jifs-210135] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Set-valued data is a significant kind of data, such as data obtained from different search engines, market data, patients’ symptoms and behaviours. An information system (IS) based on incomplete set-valued data is called an incomplete set-valued information system (ISVIS), which generalized model of a single-valued incomplete information system. This paper gives feature selection for an ISVIS by means of uncertainty measurement. Firstly, the similarity degree between two information values on a given feature of an ISVIS is proposed. Then, the tolerance relation on the object set with respect to a given feature subset in an ISVIS is obtained. Next, λ-reduction in an ISVIS is presented. What’s more, connections between the proposed feature selection and uncertainty measurement are exhibited. Lastly, feature selection algorithms based on λ-discernibility matrix, λ-information granulation, λ-information entropy and λ-significance in an ISVIS are provided. In order to better prove the practical significance of the provided algorithms, a numerical experiment is carried out, and experiment results show the number of features and average size of features by each feature selection algorithm.
Collapse
Affiliation(s)
- Lulu Li
- College of Computer Science and Engineering, Yulin Normal University, Yulin, Guangxi, P.R.China
| |
Collapse
|
8
|
Li Z, Liao S, Qu L, Song Y. Attribute selection approaches for incomplete interval-value data. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS 2021. [DOI: 10.3233/jifs-200394] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Attribute selection in an information system (IS) is an important issue when dealing with a large amount of data. An IS with incomplete interval-value data is called an incomplete interval-valued information system (IIVIS). This paper proposes attribute selection approaches for an IIVIS. Firstly, the similarity degree between two information values of a given attribute in an IIVIS is proposed. Then, the tolerance relation on the object set with respect to a given attribute subset is obtained. Next, θ-reduction in an IIVIS is studied. What is more, connections between the proposed reduction and information entropy are revealed. Lastly, three reduction algorithms base on θ-discernibility matrix, θ-information entropy and θ-significance in an IIVIS are given.
Collapse
Affiliation(s)
- Zhaowen Li
- Key Laboratory of Complex System Optimization and Big Data Processing in Department of GuangxiEducation, Yulin Normal University, Yulin, Guangxi, P.R. China
| | - Shimin Liao
- School of Mathematics and Physics, Guangxi University for Nationalities, Nanning, Guangxi, P.R. China
| | - Liangdong Qu
- School of Artificial Intelligence, Guangxi University for Nationalities, Nanning, Guangxi, P.R. China
| | - Yan Song
- School of Mathematics and Statistics, Yulin Normal University, Yulin, Guangxi, P.R. China
| |
Collapse
|
9
|
Zhang C, Dai J, Chen J. Knowledge granularity based incremental attribute reduction for incomplete decision systems. INT J MACH LEARN CYB 2020. [DOI: 10.1007/s13042-020-01089-4] [Citation(s) in RCA: 18] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
10
|
Lin B, Zhang X, Xu W, Wu Y. Dynamically updating approximations based on multi-threshold tolerance relation in incomplete interval-valued decision information systems. Knowl Inf Syst 2019. [DOI: 10.1007/s10115-019-01377-8] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
11
|
Neighborhood Attribute Reduction: A Multicriterion Strategy Based on Sample Selection. INFORMATION 2018. [DOI: 10.3390/info9110282] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
In the rough-set field, the objective of attribute reduction is to regulate the variations of measures by reducing redundant data attributes. However, most of the previous concepts of attribute reductions were designed by one and only one measure, which indicates that the obtained reduct may fail to meet the constraints given by other measures. In addition, the widely used heuristic algorithm for computing a reduct requires to scan all samples in data, and then time consumption may be too high to be accepted if the size of the data is too large. To alleviate these problems, a framework of attribute reduction based on multiple criteria with sample selection is proposed in this paper. Firstly, cluster centroids are derived from data, and then samples that are far away from the cluster centroids can be selected. This step completes the process of sample selection for reducing data size. Secondly, multiple criteria-based attribute reduction was designed, and the heuristic algorithm was used over the selected samples for computing reduct in terms of multiple criteria. Finally, the experimental results over 12 UCI datasets show that the reducts obtained by our framework not only satisfy the constraints given by multiple criteria, but also provide better classification performance and less time consumption.
Collapse
|
12
|
|
13
|
Cavallo B, Brunelli M. A general unified framework for interval pairwise comparison matrices. Int J Approx Reason 2018. [DOI: 10.1016/j.ijar.2017.11.002] [Citation(s) in RCA: 26] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
|
14
|
Dynamic computing rough approximations approach to time-evolving information granule interval-valued ordered information system. Appl Soft Comput 2017. [DOI: 10.1016/j.asoc.2017.06.009] [Citation(s) in RCA: 30] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
|
15
|
Zhang HY, Yang SY. Feature selection and approximate reasoning of large-scale set-valued decision tables based on α -dominance-based quantitative rough sets. Inf Sci (N Y) 2017. [DOI: 10.1016/j.ins.2016.06.028] [Citation(s) in RCA: 37] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
|
16
|
|
17
|
Zhang Y, Li T, Luo C, Zhang J, Chen H. Incremental updating of rough approximations in interval-valued information systems under attribute generalization. Inf Sci (N Y) 2016. [DOI: 10.1016/j.ins.2016.09.018] [Citation(s) in RCA: 43] [Impact Index Per Article: 5.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|
18
|
Yu J, Xu W. Incremental knowledge discovering in interval-valued decision information system with the dynamic data. INT J MACH LEARN CYB 2015. [DOI: 10.1007/s13042-015-0473-z] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
|
19
|
Fast algorithms of attribute reduction for covering decision systems with minimal elements in discernibility matrix. INT J MACH LEARN CYB 2015. [DOI: 10.1007/s13042-015-0438-2] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
20
|
Test-cost-sensitive attribute reduction on heterogeneous data for adaptive neighborhood model. Soft comput 2015. [DOI: 10.1007/s00500-015-1770-x] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|