1
|
Rivolli A, Garcia LP, Soares C, Vanschoren J, de Carvalho AC. Meta-features for meta-learning. Knowl Based Syst 2022. [DOI: 10.1016/j.knosys.2021.108101] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
2
|
Golshanrad P, Rahmani H, Karimian B, Karimkhani F, Weiss G. MEGA: Predicting the best classifier combination using meta-learning and a genetic algorithm. INTELL DATA ANAL 2021. [DOI: 10.3233/ida-205494] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Classifier combination through ensemble systems is one of the most effective approaches to improve the accuracy of classification systems. Ensemble systems are generally used to combine classifiers; However, selecting the best combination of individual classifiers is a challenging task. In this paper, we propose an efficient assembling method that employs both meta-learning and a genetic algorithm for the selection of the best classifiers. Our method is called MEGA, standing for using MEta-learning and a Genetic Algorithm for algorithm recommendation. MEGA has three main components: Training, Model Interpretation and Testing. The Training component extracts meta-features of each training dataset and uses a genetic algorithm to discover the best classifier combination. The Model Interpretation component interprets the relationships between meta-features and classifiers using a priori and multi-label decision tree algorithms. Finally, the Testing component uses a weighted k-nearest-neighbors algorithm to predict the best combination of classifiers for unseen datasets. We present extensive experimental results that demonstrate the performance of MEGA. MEGA achieves superior results in a comparison of three other methods and, most importantly, is able to find novel interpretable rules that can be used to select the best combination of classifiers for an unseen dataset.
Collapse
Affiliation(s)
- Paria Golshanrad
- School of Computer Engineering, Iran University of Science and Technology, Tehran, Iran
| | - Hossein Rahmani
- School of Computer Engineering, Iran University of Science and Technology, Tehran, Iran
- Department of Data Science and Knowledge Engineering, Maastricht University, Maastricht, The Netherlands
| | - Banafsheh Karimian
- School of Computer Engineering, Iran University of Science and Technology, Tehran, Iran
| | - Fatemeh Karimkhani
- School of Computer Engineering, Iran University of Science and Technology, Tehran, Iran
| | - Gerhard Weiss
- Department of Data Science and Knowledge Engineering, Maastricht University, Maastricht, The Netherlands
| |
Collapse
|
3
|
Cost-sensitive meta-learning framework. JOURNAL OF MODELLING IN MANAGEMENT 2021. [DOI: 10.1108/jm2-03-2021-0065] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Purpose
This paper aims to describe the use of a meta-learning framework for recommending cost-sensitive classification methods with the aim of answering an important question that arises in machine learning, namely, “Among all the available classification algorithms, and in considering a specific type of data and cost, which is the best algorithm for my problem?”
Design/methodology/approach
This paper describes the use of a meta-learning framework for recommending cost-sensitive classification methods for the aim of answering an important question that arises in machine learning, namely, “Among all the available classification algorithms, and in considering a specific type of data and cost, which is the best algorithm for my problem?” The framework is based on the idea of applying machine learning techniques to discover knowledge about the performance of different machine learning algorithms. It includes components that repeatedly apply different classification methods on data sets and measures their performance. The characteristics of the data sets, combined with the algorithms and the performance provide the training examples. A decision tree algorithm is applied to the training examples to induce the knowledge, which can then be used to recommend algorithms for new data sets. The paper makes a contribution to both meta-learning and cost-sensitive machine learning approaches. Those both fields are not new, however, building a recommender that recommends the optimal case-sensitive approach for a given data problem is the contribution. The proposed solution is implemented in WEKA and evaluated by applying it on different data sets and comparing the results with existing studies available in the literature. The results show that a developed meta-learning solution produces better results than METAL, a well-known meta-learning system. The developed solution takes the misclassification cost into consideration during the learning process, which is not available in the compared project.
Findings
The proposed solution is implemented in WEKA and evaluated by applying it to different data sets and comparing the results with existing studies available in the literature. The results show that a developed meta-learning solution produces better results than METAL, a well-known meta-learning system.
Originality/value
The paper presents a major piece of new information in writing for the first time. Meta-learning work has been done before but this paper presents a new meta-learning framework that is costs sensitive.
Collapse
|
4
|
A case-based reasoning system for recommendation of data cleaning algorithms in classification and regression tasks. Appl Soft Comput 2020. [DOI: 10.1016/j.asoc.2020.106180] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
|
5
|
Automatically discovering clusters of algorithm and problem instance behaviors as well as their causes from experimental data, algorithm setups, and instance features. Appl Soft Comput 2018. [DOI: 10.1016/j.asoc.2018.08.030] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
|
6
|
Singh D, Singh B. Hybridization of feature selection and feature weighting for high dimensional data. APPL INTELL 2018. [DOI: 10.1007/s10489-018-1348-2] [Citation(s) in RCA: 18] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/01/2022]
|
7
|
Goswami S, Chakrabarti A, Chakraborty B. An efficient feature selection technique for clustering based on a new measure of feature importance. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS 2017. [DOI: 10.3233/ifs-162156] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Affiliation(s)
- Saptarsi Goswami
- Computer Science and Engineering, Institute of Engineering & Management, Salt Lake, Kolkata, India
| | - Amlan Chakrabarti
- A.k.Choudhury School of Information Technology, Calcutta University, Kolkata, India
| | - Basabi Chakraborty
- Faculty of Software and Information Science, Iwate Prefectural University, Iwate, Japan
| |
Collapse
|
8
|
Garcia-Chimeno Y, Garcia-Zapirain B, Gomez-Beldarrain M, Fernandez-Ruanova B, Garcia-Monco JC. Automatic migraine classification via feature selection committee and machine learning techniques over imaging and questionnaire data. BMC Med Inform Decis Mak 2017; 17:38. [PMID: 28407777 PMCID: PMC5390380 DOI: 10.1186/s12911-017-0434-4] [Citation(s) in RCA: 23] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/13/2016] [Accepted: 03/29/2017] [Indexed: 11/10/2022] Open
Abstract
BACKGROUND Feature selection methods are commonly used to identify subsets of relevant features to facilitate the construction of models for classification, yet little is known about how feature selection methods perform in diffusion tensor images (DTIs). In this study, feature selection and machine learning classification methods were tested for the purpose of automating diagnosis of migraines using both DTIs and questionnaire answers related to emotion and cognition - factors that influence of pain perceptions. METHODS We select 52 adult subjects for the study divided into three groups: control group (15), subjects with sporadic migraine (19) and subjects with chronic migraine and medication overuse (18). These subjects underwent magnetic resonance with diffusion tensor to see white matter pathway integrity of the regions of interest involved in pain and emotion. The tests also gather data about pathology. The DTI images and test results were then introduced into feature selection algorithms (Gradient Tree Boosting, L1-based, Random Forest and Univariate) to reduce features of the first dataset and classification algorithms (SVM (Support Vector Machine), Boosting (Adaboost) and Naive Bayes) to perform a classification of migraine group. Moreover we implement a committee method to improve the classification accuracy based on feature selection algorithms. RESULTS When classifying the migraine group, the greatest improvements in accuracy were made using the proposed committee-based feature selection method. Using this approach, the accuracy of classification into three types improved from 67 to 93% when using the Naive Bayes classifier, from 90 to 95% with the support vector machine classifier, 93 to 94% in boosting. The features that were determined to be most useful for classification included are related with the pain, analgesics and left uncinate brain (connected with the pain and emotions). CONCLUSIONS The proposed feature selection committee method improved the performance of migraine diagnosis classifiers compared to individual feature selection methods, producing a robust system that achieved over 90% accuracy in all classifiers. The results suggest that the proposed methods can be used to support specialists in the classification of migraines in patients undergoing magnetic resonance imaging.
Collapse
Affiliation(s)
- Yolanda Garcia-Chimeno
- DeustoTech - Fundacion Deusto, Avda. Universidades, 24, Bilbao, 48007 Spain
- Facultad IngenieriaUniversidad de Deusto, Avda. Universidades, 24, Bilbao, 48007 Spain
| | - Begonya Garcia-Zapirain
- DeustoTech - Fundacion Deusto, Avda. Universidades, 24, Bilbao, 48007 Spain
- Facultad IngenieriaUniversidad de Deusto, Avda. Universidades, 24, Bilbao, 48007 Spain
| | - Marian Gomez-Beldarrain
- Service of Neurology Hospital de Galdakao-Usansolo, Barrio Labeaga, S/N, Galdakao, 48960 Spain
| | | | - Juan Carlos Garcia-Monco
- Research and Innovation Department, Magnetic Resonance Imaging Unit, OSATEK, Alameda Urquijo, 36, Bilbao, 48011 Spain
| |
Collapse
|
9
|
Tiwari S, Singh B, Kaur M. An approach for feature selection using local searching and global optimization techniques. Neural Comput Appl 2017. [DOI: 10.1007/s00521-017-2959-y] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
10
|
Bagherzadeh-Khiabani F, Ramezankhani A, Azizi F, Hadaegh F, Steyerberg EW, Khalili D. A tutorial on variable selection for clinical prediction models: feature selection methods in data mining could improve the results. J Clin Epidemiol 2015; 71:76-85. [PMID: 26475568 DOI: 10.1016/j.jclinepi.2015.10.002] [Citation(s) in RCA: 68] [Impact Index Per Article: 7.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/30/2014] [Revised: 08/30/2015] [Accepted: 10/02/2015] [Indexed: 12/31/2022]
Abstract
OBJECTIVES Identifying an appropriate set of predictors for the outcome of interest is a major challenge in clinical prediction research. The aim of this study was to show the application of some variable selection methods, usually used in data mining, for an epidemiological study. We introduce here a systematic approach. STUDY DESIGN AND SETTING The P-value-based method, usually used in epidemiological studies, and several filter and wrapper methods were implemented to select the predictors of diabetes among 55 variables in 803 prediabetic females, aged ≥ 20 years, followed for 10-12 years. To develop a logistic model, variables were selected from a train data set and evaluated on the test data set. The measures of Akaike information criterion (AIC) and area under the curve (AUC) were used as performance criteria. We also implemented a full model with all 55 variables. RESULTS We found that the worst and the best models were the full model and models based on the wrappers, respectively. Among filter methods, symmetrical uncertainty gave both the best AUC and AIC. CONCLUSION Our experiment showed that the variable selection methods used in data mining could improve the performance of clinical prediction models. An R program was developed to make these methods more feasible and visualize the results.
Collapse
Affiliation(s)
- Farideh Bagherzadeh-Khiabani
- Prevention of Metabolic Disorders Research Center, Research Institute for Endocrine Sciences, Shahid Beheshti University of Medical Sciences, Velenjak, 1985717413 Tehran, Iran
| | - Azra Ramezankhani
- Prevention of Metabolic Disorders Research Center, Research Institute for Endocrine Sciences, Shahid Beheshti University of Medical Sciences, Velenjak, 1985717413 Tehran, Iran
| | - Fereidoun Azizi
- Endocrine Research Center, Research Institute for Endocrine Sciences, Shahid Beheshti University of Medical Sciences, Velenjak, 1985717413 Tehran, Iran
| | - Farzad Hadaegh
- Prevention of Metabolic Disorders Research Center, Research Institute for Endocrine Sciences, Shahid Beheshti University of Medical Sciences, Velenjak, 1985717413 Tehran, Iran
| | | | - Davood Khalili
- Prevention of Metabolic Disorders Research Center, Research Institute for Endocrine Sciences, Shahid Beheshti University of Medical Sciences, Velenjak, 1985717413 Tehran, Iran; Department of Biostatistics and Epidemiology, Research Institute for Endocrine Sciences, Shahid Beheshti University of Medical Sciences, Velenjak, 1985717413 Tehran, Iran.
| |
Collapse
|
11
|
Muñoz MA, Sun Y, Kirley M, Halgamuge SK. Algorithm selection for black-box continuous optimization problems: A survey on methods and challenges. Inf Sci (N Y) 2015. [DOI: 10.1016/j.ins.2015.05.010] [Citation(s) in RCA: 120] [Impact Index Per Article: 13.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
12
|
Khan A, Baig AR. Multi-objective feature subset selection using mRMR based enhanced ant colony optimization algorithm (mRMR-EACO). J EXP THEOR ARTIF IN 2015. [DOI: 10.1080/0952813x.2015.1056240] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
13
|
Nagpal A, Gaur D. A New Proposed Feature Subset Selection Algorithm Based on Maximization of Gain Ratio. BIG DATA ANALYTICS 2015. [DOI: 10.1007/978-3-319-27057-9_13] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022] Open
|