1
|
Pontes ED, Pinto M, Lopes F, Teixeira C. Concept-drifts adaptation for machine learning EEG epilepsy seizure prediction. Sci Rep 2024; 14:8204. [PMID: 38589379 PMCID: PMC11001609 DOI: 10.1038/s41598-024-57744-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2024] [Accepted: 03/21/2024] [Indexed: 04/10/2024] Open
Abstract
Seizure prediction remains a challenge, with approximately 30% of patients unresponsive to conventional treatments. Addressing this issue is crucial for improving patients' quality of life, as timely intervention can mitigate the impact of seizures. In this research field, it is critical to identify the preictal interval, the transition from regular brain activity to a seizure. While previous studies have explored various Electroencephalogram (EEG) based methodologies for prediction, few have been clinically applicable. Recent studies have underlined the dynamic nature of EEG data, characterised by data changes with time, known as concept drifts, highlighting the need for automated methods to detect and adapt to these changes. In this study, we investigate the effectiveness of automatic concept drift adaptation methods in seizure prediction. Three patient-specific seizure prediction approaches with a 10-minute prediction horizon are compared: a seizure prediction algorithm incorporating a window adjustment method by optimising performance with Support Vector Machines (Backwards-Landmark Window), a seizure prediction algorithm incorporating a data-batch (seizures) selection method using a logistic regression (Seizure-batch Regression), and a seizure prediction algorithm with a dynamic integration of classifiers (Dynamic Weighted Ensemble). These methods incorporate a retraining process after each seizure and use a combination of univariate linear features and SVM classifiers. The Firing Power was used as a post-processing technique to generate alarms before seizures. These methodologies were compared with a control approach based on the typical machine learning pipeline, considering a group of 37 patients with Temporal Lobe Epilepsy from the EPILEPSIAE database. The best-performing approach (Backwards-Landmark Window) achieved results of 0.75 ± 0.33 for sensitivity and 1.03 ± 1.00 for false positive rate per hour. This new strategy performed above chance for 89% of patients with the surrogate predictor, whereas the control approach only validated 46%.
Collapse
Affiliation(s)
- Edson David Pontes
- Department of Informatics Engineering, CISUC, University of Coimbra, Coimbra, Portugal.
| | - Mauro Pinto
- Department of Informatics Engineering, CISUC, University of Coimbra, Coimbra, Portugal
| | - Fábio Lopes
- Department of Informatics Engineering, CISUC, University of Coimbra, Coimbra, Portugal
- Epilepsy Center, Department Neurosurgery, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany
| | - César Teixeira
- Department of Informatics Engineering, CISUC, University of Coimbra, Coimbra, Portugal
| |
Collapse
|
2
|
Phan HT, Nguyen NT, Hwang D. Fake news detection: A survey of graph neural network methods. Appl Soft Comput 2023; 139:110235. [PMID: 36999094 PMCID: PMC10036155 DOI: 10.1016/j.asoc.2023.110235] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/05/2022] [Revised: 08/03/2022] [Accepted: 03/19/2023] [Indexed: 04/01/2023]
Abstract
The emergence of various social networks has generated vast volumes of data. Efficient methods for capturing, distinguishing, and filtering real and fake news are becoming increasingly important, especially after the outbreak of the COVID-19 pandemic. This study conducts a multiaspect and systematic review of the current state and challenges of graph neural networks (GNNs) for fake news detection systems and outlines a comprehensive approach to implementing fake news detection systems using GNNs. Furthermore, advanced GNN-based techniques for implementing pragmatic fake news detection systems are discussed from multiple perspectives. First, we introduce the background and overview related to fake news, fake news detection, and GNNs. Second, we provide a GNN taxonomy-based fake news detection taxonomy and review and highlight models in categories. Subsequently, we compare critical ideas, advantages, and disadvantages of the methods in categories. Next, we discuss the possible challenges of fake news detection and GNNs. Finally, we present several open issues in this area and discuss potential directions for future research. We believe that this review can be utilized by systems practitioners and newcomers in surmounting current impediments and navigating future situations by deploying a fake news detection system using GNNs.
Collapse
Affiliation(s)
- Huyen Trang Phan
- Department of Computer Engineering, Yeungnam University, Gyeongsan, South Korea
- Faculty of Information Technology, Nguyen Tat Thanh University, Ho Chi Minh, Vietnam
| | - Ngoc Thanh Nguyen
- Department of Applied Informatics, Wroclaw University of Science and Technology, Wroclaw, Poland
| | - Dosam Hwang
- Department of Computer Engineering, Yeungnam University, Gyeongsan, South Korea
| |
Collapse
|
3
|
Rahmani K, Thapa R, Tsou P, Casie Chetty S, Barnes G, Lam C, Foon Tso C. Assessing the effects of data drift on the performance of machine learning models used in clinical sepsis prediction. Int J Med Inform 2023; 173:104930. [PMID: 36893656 DOI: 10.1016/j.ijmedinf.2022.104930] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/20/2022] [Revised: 10/30/2022] [Accepted: 11/15/2022] [Indexed: 11/21/2022]
Abstract
BACKGROUND Data drift can negatively impact the performance of machine learning algorithms (MLAs) that were trained on historical data. As such, MLAs should be continuously monitored and tuned to overcome the systematic changes that occur in the distribution of data. In this paper, we study the extent of data drift and provide insights about its characteristics for sepsis onset prediction. This study will help elucidate the nature of data drift for prediction of sepsis and similar diseases. This may aid with the development of more effective patient monitoring systems that can stratify risk for dynamic disease states in hospitals. METHODS We devise a series of simulations that measure the effects of data drift in patients with sepsis, using electronic health records (EHR). We simulate multiple scenarios in which data drift may occur, namely the change in the distribution of the predictor variables (covariate shift), the change in the statistical relationship between the predictors and the target (concept shift), and the occurrence of a major healthcare event (major event) such as the COVID-19 pandemic. We measure the impact of data drift on model performances, identify the circumstances that necessitate model retraining, and compare the effects of different retraining methodologies and model architecture on the outcomes. We present the results for two different MLAs, eXtreme Gradient Boosting (XGB) and Recurrent Neural Network (RNN). RESULTS Our results show that the properly retrained XGB models outperform the baseline models in all simulation scenarios, hence signifying the existence of data drift. In the major event scenario, the area under the receiver operating characteristic curve (AUROC) at the end of the simulation period is 0.811 for the baseline XGB model and 0.868 for the retrained XGB model. In the covariate shift scenario, the AUROC at the end of the simulation period for the baseline and retrained XGB models is 0.853 and 0.874 respectively. In the concept shift scenario and under the mixed labeling method, the retrained XGB models perform worse than the baseline model for most simulation steps. However, under the full relabeling method, the AUROC at the end of the simulation period for the baseline and retrained XGB models is 0.852 and 0.877 respectively. The results for the RNN models were mixed, suggesting that retraining based on a fixed network architecture may be inadequate for an RNN. We also present the results in the form of other performance metrics such as the ratio of observed to expected probabilities (calibration) and the normalized rate of positive predictive values (PPV) by prevalence, referred to as lift, at a sensitivity of 0.8. CONCLUSION Our simulations reveal that retraining periods of a couple of months or using several thousand patients are likely to be adequate to monitor machine learning models that predict sepsis. This indicates that a machine learning system for sepsis prediction will probably need less infrastructure for performance monitoring and retraining compared to other applications in which data drift is more frequent and continuous. Our results also show that in the event of a concept shift, a full overhaul of the sepsis prediction model may be necessary because it indicates a discrete change in the definition of sepsis labels, and mixing the labels for the sake of incremental training may not produce the desired results.
Collapse
Affiliation(s)
- Keyvan Rahmani
- Dascena, Inc., 12333 Sowden Rd Ste B PMB 65148, Houston, TX 77080-2059, USA
| | - Rahul Thapa
- Dascena, Inc., 12333 Sowden Rd Ste B PMB 65148, Houston, TX 77080-2059, USA
| | - Peiling Tsou
- Dascena, Inc., 12333 Sowden Rd Ste B PMB 65148, Houston, TX 77080-2059, USA
| | - Satish Casie Chetty
- Dascena, Inc., 12333 Sowden Rd Ste B PMB 65148, Houston, TX 77080-2059, USA.
| | - Gina Barnes
- Dascena, Inc., 12333 Sowden Rd Ste B PMB 65148, Houston, TX 77080-2059, USA
| | - Carson Lam
- Dascena, Inc., 12333 Sowden Rd Ste B PMB 65148, Houston, TX 77080-2059, USA
| | - Chak Foon Tso
- Dascena, Inc., 12333 Sowden Rd Ste B PMB 65148, Houston, TX 77080-2059, USA
| |
Collapse
|
4
|
Peng X, Wang FY, Li L. MixGradient: A gradient-based re-weighting scheme with mixup for imbalanced data streams. Neural Netw 2023; 161:525-534. [PMID: 36805267 DOI: 10.1016/j.neunet.2023.02.017] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2022] [Revised: 11/22/2022] [Accepted: 02/09/2023] [Indexed: 02/15/2023]
Abstract
A challenge for contemporary deep neural networks in real-world problems is learning from an imbalanced data stream, where data tends to be received chunk by chunk over time, and the prior class distribution is severely imbalanced. Although many sophisticated algorithms have been derived, most of them overlook the importance of gradient information. From this perspective, the difficulty of learning from imbalanced data streams lies in the fact that the gradient estimated on an uneven class distribution is not informative enough to reflect the critical pattern of each class. To this end, we propose to assign higher weights on the training samples whose gradients are close to the gradient of corresponding typical samples, thus highlighting the important samples in minority classes and suppressing the noisy samples in majority classes. Such an idea can be combined with Mixup, which exploits the interpolation information of data to further compensate for the information of sample space that the typical samples do not provide and expand the role of the proposed re-weighting scheme. Experiments on artificially induced long-tailed CIFAR data streams and long-tailed MiniPlaces data stream show that the resulting method, termed MixGradient, boosts the generalization performance of DNNs under different imbalance ratios and achieves up to 10% accuracy improvement.
Collapse
Affiliation(s)
- Xinyu Peng
- Department of Automation, Tsinghua University, Beijing, 100084, China.
| | - Fei-Yue Wang
- State Key Laboratory of Management and Control for Complex Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, 100080, China.
| | - Li Li
- Department of Automation, Tsinghua University, Beijing, 100084, China; National Laboratory for Information Science and Technology, Tsinghua University, Beijing, 100084, China.
| |
Collapse
|
5
|
Jiang T, Zeng J. Time-Aware Explainable Recommendation via Updating Enabled Online Prediction. ENTROPY (BASEL, SWITZERLAND) 2022; 24:1639. [PMID: 36421494 PMCID: PMC9689638 DOI: 10.3390/e24111639] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/17/2022] [Revised: 11/06/2022] [Accepted: 11/08/2022] [Indexed: 06/16/2023]
Abstract
There has been growing attention on explainable recommendation that is able to provide high-quality results as well as intuitive explanations. However, most existing studies use offline prediction strategies where recommender systems are trained once while used forever, which ignores the dynamic and evolving nature of user-item interactions. There are two main issues with these methods. First, their random dataset split setting will result in data leakage that knowledge should not be known at the time of training is utilized. Second, the dynamic characteristics of user preferences are overlooked, resulting in a model aging issue where the model's performance degrades along with time. In this paper, we propose an updating enabled online prediction framework for the time-aware explainable recommendation. Specifically, we propose an online prediction scheme to eliminate the data leakage issue and two novel updating strategies to relieve the model aging issue. Moreover, we conduct extensive experiments on four real-world datasets to evaluate the effectiveness of our proposed methods. Compared with the state-of-the-art, our time-aware approach achieves higher accuracy results and more convincing explanations for the entire lifetime of recommendation systems, i.e., both the initial period and the long-term usage.
Collapse
|
6
|
Takada T, Kitajima T. Trend-following with better adaptation to large downside risks. PLoS One 2022; 17:e0276322. [PMID: 36256670 PMCID: PMC9578607 DOI: 10.1371/journal.pone.0276322] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2022] [Accepted: 10/04/2022] [Indexed: 11/18/2022] Open
Abstract
Avoiding losses from long-term trend reversals is challenging, and trend-following is one of the few trading approaches to explore it. While trend-following is popular among investors and has gained increased attention in academia, the recent diminished profitability in equity markets casts doubt on its effectiveness. To clarify its cause and suggest remedies, we thoroughly examine the effect of market conditions and averaging window on recent profitability using four major stock indices in an out-of-sample experiment comparing trend-following rules (moving average and momentum) and a machine-classification-based non-trend-following rule. In addition to the significant advantage of trend-following rules in avoiding downside risks, we find a discrepancy in the optimum averaging window size between trend direction phases, which is exacerbated by a higher positive trend direction ratio. A higher positive trend direction ratio leads to poor performance relative to buy-and-hold returns. This discrepancy creates the dilemma of choosing which trend direction phase to emphasize. Incorporating machine-learning into trend-following is effective for alleviating this dilemma. We find that the profit-maximizing averaging window realizes the level that best balances the dilemma and suggest a simple guideline for selecting the optimum averaging window. We attribute the sluggishness of trend-following in recent equity markets to the insufficient chances of trend reversals rather than their loss of profitability. Our results contribute to improving the performance of trend following by mitigating the dilemma.
Collapse
Affiliation(s)
- Teruko Takada
- Graduate School of Business, Osaka Metropolitan University, Osaka, Japan
- * E-mail:
| | - Takahiro Kitajima
- Graduate School of Business, Osaka Metropolitan University, Osaka, Japan
- Faculty of Commerce, Kumamoto Gakuen University, Kumamoto, Japan
| |
Collapse
|
7
|
Liu H, Zhou Y, Liu B, Zhao J, Yao R, Shao Z. Incremental learning with neural networks for computer vision: a survey. Artif Intell Rev 2022. [DOI: 10.1007/s10462-022-10294-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/01/2022]
|
8
|
Suryawanshi S, Goswami A, Patil P, Mishra V. Adaptive windowing based recurrent neural network for drift adaption in non-stationary environment. JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING 2022; 14:1-15. [PMID: 35789602 PMCID: PMC9243804 DOI: 10.1007/s12652-022-04116-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/24/2021] [Accepted: 06/06/2022] [Indexed: 06/15/2023]
Abstract
In today's digital era, many applications generate massive data streams that must be sequenced and processed immediately. Therefore, storing large amounts of data for analysis is impractical. Now, this infinite amount of evolving data confronts concept drifts in data stream classification. Concept drift is a phenomenon in which the distribution of input data or the relationship between input data and target label changes over time. If the drifts are not addressed, the learning model's performance suffers. Non-stationary data streams must be processed as they arrive, and neural networks' built-in capabilities aid in the processing of huge non-stationary data streams. We proposed an adaptive windowing approach based on a gated recurrent unit, a variant of the recurrent neural network incrementally trained on incoming data (for the real-world airline and synthetic Streaming Ensemble Algorithm (SEA) datasets), and employed elastic weight consolidation with the Fisher information matrix to prevent forgetting. Unlike the traditional fixed window methodology, the proposed model dynamically increases the window size if the prediction is correct and reduces it if drifts occur. As a result, an adaptive recurrent neural network model can adapt to changes in the non-stationary data stream and provide consistent performance. Moreover, the findings revealed that on the airline and the SEA dataset, the proposed model outperforms state-of-the-art methods by achieving 67.74% and 91.70% accuracy, respectively. Further, the results demonstrated that the proposed model has a better accuracy of 3.6% and 1.6% for the SEA and the airline dataset, respectively.
Collapse
Affiliation(s)
- Shubhangi Suryawanshi
- Bennett University, Greater Noida, India
- Dr. D. Y. Patil Institute of Technology, Pimpri, Pune, India
| | | | - Pramod Patil
- Dr. D. Y. Patil Institute of Technology, Pimpri, Pune, India
| | | |
Collapse
|
9
|
Rahmani K, Thapa R, Tsou P, Chetty SC, Barnes G, Lam C, Tso CF. Assessing the effects of data drift on the performance of machine learning models used in clinical sepsis prediction. MEDRXIV : THE PREPRINT SERVER FOR HEALTH SCIENCES 2022:2022.06.06.22276062. [PMID: 35702157 PMCID: PMC9196120 DOI: 10.1101/2022.06.06.22276062] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 12/29/2022]
Abstract
Background Data drift can negatively impact the performance of machine learning algorithms (MLAs) that were trained on historical data. As such, MLAs should be continuously monitored and tuned to overcome the systematic changes that occur in the distribution of data. In this paper, we study the extent of data drift and provide insights about its characteristics for sepsis onset prediction. This study will help elucidate the nature of data drift for prediction of sepsis and similar diseases. This may aid with the development of more effective patient monitoring systems that can stratify risk for dynamic disease states in hospitals. Methods We devise a series of simulations that measure the effects of data drift in patients with sepsis. We simulate multiple scenarios in which data drift may occur, namely the change in the distribution of the predictor variables (covariate shift), the change in the statistical relationship between the predictors and the target (concept shift), and the occurrence of a major healthcare event (major event) such as the COVID-19 pandemic. We measure the impact of data drift on model performances, identify the circumstances that necessitate model retraining, and compare the effects of different retraining methodologies and model architecture on the outcomes. We present the results for two different MLAs, eXtreme Gradient Boosting (XGB) and Recurrent Neural Network (RNN). Results Our results show that the properly retrained XGB models outperform the baseline models in all simulation scenarios, hence signifying the existence of data drift. In the major event scenario, the area under the receiver operating characteristic curve (AUROC) at the end of the simulation period is 0.811 for the baseline XGB model and 0.868 for the retrained XGB model. In the covariate shift scenario, the AUROC at the end of the simulation period for the baseline and retrained XGB models is 0.853 and 0.874 respectively. In the concept shift scenario and under the mixed labeling method, the retrained XGB models perform worse than the baseline model for most simulation steps. However, under the full relabeling method, the AUROC at the end of the simulation period for the baseline and retrained XGB models is 0.852 and 0.877 respectively. The results for the RNN models were mixed, suggesting that retraining based on a fixed network architecture may be inadequate for an RNN. We also present the results in the form of other performance metrics such as the ratio of observed to expected probabilities (calibration) and the normalized rate of positive predictive values (PPV) by prevalence, referred to as lift, at a sensitivity of 0.8. Conclusion Our simulations reveal that retraining periods of a couple of months or using several thousand patients are likely to be adequate to monitor machine learning models that predict sepsis. This indicates that a machine learning system for sepsis prediction will probably need less infrastructure for performance monitoring and retraining compared to other applications in which data drift is more frequent and continuous. Our results also show that in the event of a concept shift, a full overhaul of the sepsis prediction model may be necessary because it indicates a discrete change in the definition of sepsis labels, and mixing the labels for the sake of incremental training may not produce the desired results.
Collapse
Affiliation(s)
- Keyvan Rahmani
- Dascena, Inc., 12333 Sowden Rd Ste B PMB 65148, Houston, Texas 77080-2059
| | - Rahul Thapa
- Dascena, Inc., 12333 Sowden Rd Ste B PMB 65148, Houston, Texas 77080-2059
| | - Peiling Tsou
- Dascena, Inc., 12333 Sowden Rd Ste B PMB 65148, Houston, Texas 77080-2059
| | | | - Gina Barnes
- Dascena, Inc., 12333 Sowden Rd Ste B PMB 65148, Houston, Texas 77080-2059
| | - Carson Lam
- Dascena, Inc., 12333 Sowden Rd Ste B PMB 65148, Houston, Texas 77080-2059
| | - Chak Foon Tso
- Dascena, Inc., 12333 Sowden Rd Ste B PMB 65148, Houston, Texas 77080-2059
| |
Collapse
|
10
|
Bayram F, Ahmed BS, Kassler A. From concept drift to model degradation: An overview on performance-aware drift detectors. Knowl Based Syst 2022. [DOI: 10.1016/j.knosys.2022.108632] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
11
|
Forecasting Charging Point Occupancy Using Supervised Learning Algorithms. ENERGIES 2022. [DOI: 10.3390/en15093409] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/04/2023]
Abstract
The prediction of charging point occupancy enables electric vehicle users to better plan their charging processes and thus promotes the acceptance of electromobility. The study uses Adaptive Charging Network data to investigate a public and a workplace site for predicting individual charging station occupancy as well as overall site occupancy. Predicting individual charging point occupancy is formulated as a classification problem, while predicting total occupancy is formulated as a regression problem. The effects of different feature sets on the predictions are investigated, as well as whether a model trained on data of all charging points per site performs better than one trained on the data of a specific charging point. Reviewed studies so far, however, have failed to compare these two approaches to benchmarks, to use more than one algorithm, or to consider more than one site. Therefore, the following supervised machine-learning algorithms were applied for both tasks: linear and logistic regression, k-nearest neighbor, random forest, and XGBoost. Further, the model results are compared to three different naïve approaches which provide a robust benchmark, and the two training approaches were applied to two different sites. By adding features, the prediction quality can be increased considerably, which resulted in some models performing better than the naïve approaches. In general, models trained on data of all charging points of a site perform slightly better on median than models trained on individual charging points. In certain cases, however, individually trained models achieve the best results, while charging points with very low relative charging point occupancy can benefit from a model that has been trained on all data.
Collapse
|
12
|
Fake news detection based on news content and social contexts: a transformer-based approach. INTERNATIONAL JOURNAL OF DATA SCIENCE AND ANALYTICS 2022; 13:335-362. [PMID: 35128038 PMCID: PMC8800852 DOI: 10.1007/s41060-021-00302-z] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/24/2021] [Accepted: 12/13/2021] [Indexed: 11/30/2022]
Abstract
Fake news is a real problem in today’s world, and it has become more extensive and harder to identify. A major challenge in fake news detection is to detect it in the early phase. Another challenge in fake news detection is the unavailability or the shortage of labelled data for training the detection models. We propose a novel fake news detection framework that can address these challenges. Our proposed framework exploits the information from the news articles and the social contexts to detect fake news. The proposed model is based on a Transformer architecture, which has two parts: the encoder part to learn useful representations from the fake news data and the decoder part that predicts the future behaviour based on past observations. We also incorporate many features from the news content and social contexts into our model to help us classify the news better. In addition, we propose an effective labelling technique to address the label shortage problem. Experimental results on real-world data show that our model can detect fake news with higher accuracy within a few minutes after it propagates (early detection) than the baselines.
Collapse
|
13
|
|
14
|
Goretzko D, Israel LSF. Pitfalls of Machine Learning-Based Personnel Selection. JOURNAL OF PERSONNEL PSYCHOLOGY 2022. [DOI: 10.1027/1866-5888/a000287] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
Abstract
Abstract. In recent years, machine learning (ML) modeling (often referred to as artificial intelligence) has become increasingly popular for personnel selection purposes. Numerous organizations use ML-based procedures for screening large candidate pools, while some companies try to automate the hiring process as far as possible. Since ML models can handle large sets of predictor variables and are therefore able to incorporate many different data sources (often more than common procedures can consider), they promise a higher predictive accuracy and objectivity in selecting the best candidate than traditional personal selection processes. However, there are some pitfalls and challenges that have to be taken into account when using ML for a sensitive issue as personnel selection. In this paper, we address these major challenges – namely the definition of a valid criterion, transparency regarding collected data and decision mechanisms, algorithmic fairness, changing data conditions, and adequate performance evaluation – and discuss some recommendations for implementing fair, transparent, and accurate ML-based selection algorithms.
Collapse
Affiliation(s)
- David Goretzko
- Department of Psychology, Ludwig-Maximilians-Universität München, Munich, Germany
| | | |
Collapse
|
15
|
Nurcahyani I, Lee JW. Role of Machine Learning in Resource Allocation Strategy over Vehicular Networks: A Survey. SENSORS 2021; 21:s21196542. [PMID: 34640858 PMCID: PMC8512744 DOI: 10.3390/s21196542] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 08/20/2021] [Revised: 09/26/2021] [Accepted: 09/28/2021] [Indexed: 12/20/2022]
Abstract
The increasing demand for smart vehicles with many sensing capabilities will escalate data traffic in vehicular networks. Meanwhile, available network resources are limited. The emergence of AI implementation in vehicular network resource allocation opens the opportunity to improve resource utilization to provide more reliable services. Accordingly, many resource allocation schemes with various machine learning algorithms have been proposed to dynamically manage and allocate network resources. This survey paper presents how machine learning is leveraged in the vehicular network resource allocation strategy. We focus our study on determining its role in the mechanism. First, we provide an analysis of how authors designed their scenarios to orchestrate the resource allocation strategy. Secondly, we classify the mechanisms based on the parameters they chose when designing the algorithms. Finally, we analyze the challenges in designing a resource allocation strategy in vehicular networks using machine learning. Therefore, a thorough understanding of how machine learning algorithms are utilized to offer a dynamic resource allocation in vehicular networks is provided in this study.
Collapse
Affiliation(s)
- Ida Nurcahyani
- School of Electrical and Electronics Engineering, Chung-Ang University, Seoul 06974, Korea;
- Department of Electrical Engineering, Universitas Islam Indonesia, Yogyakarta 55584, Indonesia
| | - Jeong Woo Lee
- School of Electrical and Electronics Engineering, Chung-Ang University, Seoul 06974, Korea;
- Correspondence: ; Tel.: +82-2-820-5734
| |
Collapse
|
16
|
Singh MN, Khaiyum S. Enhanced Data Stream Classification by Optimized Weight Updated Meta-learning: Continuous learning-based on Concept-Drift. INTERNATIONAL JOURNAL OF WEB INFORMATION SYSTEMS 2021. [DOI: 10.1108/ijwis-01-2021-0007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Purpose
The aim of continuous learning is to obtain and fine-tune information gradually without removing the already existing information. Many conventional approaches in streaming data classification assume that all arrived new data is completely labeled. To regularize Neural Networks (NNs) by merging side information like user-provided labels or pair-wise constraints, incremental semi-supervised learning models need to be introduced. However, they are hard to implement, specifically in non-stationary environments because of the efficiency and sensitivity of such algorithms to parameters. The periodic update and maintenance of the decision method is the significant challenge in incremental algorithms whenever the new data arrives.
Design/methodology/approach
Hence, this paper plans to develop the meta-learning model for handling continuous or streaming data. Initially, the data pertain to continuous behavior is gathered from diverse benchmark source. Further, the classification of the data is performed by the Recurrent Neural Network (RNN), in which testing weight is adjusted or optimized by the new meta-heuristic algorithm. Here, the weight is updated for reducing the error difference between the target and the measured data when new data is given for testing. The optimized weight updated testing is performed by evaluating the concept-drift and classification accuracy. The new continuous learning by RNN is accomplished by the improved Opposition-based Novel Updating Spotted Hyena Optimization (ONU-SHO). Finally, the experiments with different datasets show that the proposed learning is improved over the conventional models.
Findings
From the analysis, the accuracy of the ONU-SHO based RNN (ONU-SHO-RNN) was 10.1% advanced than Decision Tree (DT), 7.6% advanced than Naive Bayes (NB), 7.4% advanced than k-nearest neighbors (KNN), 2.5% advanced than Support Vector Machine (SVM) 9.3% advanced than NN, and 10.6% advanced than RNN. Hence, it is confirmed that the ONU-SHO algorithm is performing well for acquiring the best data stream classification.
Originality/value
This paper introduces a novel meta-learning model using Opposition-based Novel Updating Spotted Hyena Optimization (ONU-SHO)-based Recurrent Neural Network (RNN) for handling continuous or streaming data. This is the first work utilizes a novel meta-learning model using Opposition-based Novel Updating Spotted Hyena Optimization (ONU-SHO)-based Recurrent Neural Network (RNN) for handling continuous or streaming data.
Collapse
|
17
|
A Regularization-Based Big Data Framework for Winter Precipitation Forecasting on Streaming Data. ELECTRONICS 2021. [DOI: 10.3390/electronics10161872] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
In the current paper, we propose a machine learning forecasting model for the accurate prediction of qualitative weather information on winter precipitation types, utilized in Apache Spark Streaming distributed framework. The proposed model receives storage and processes data in real-time, in order to extract useful knowledge from different sensors related to weather data. In following, the numerical weather prediction model aims at forecasting the weather type given three precipitation classes namely rain, freezing rain, and snow as recorded in the Automated Surface Observing System (ASOS) network. For depicting the effectiveness of our proposed schema, a regularization technique for feature selection so as to avoid overfitting is implemented. Several classification models covering three different categorization methods namely the Bayesian, decision trees, and meta/ensemble methods, have been investigated in a real dataset. The experimental analysis illustrates that the utilization of the regularization technique could offer a significant boost in forecasting performance.
Collapse
|
18
|
Credit Card Fraud Detection in Card-Not-Present Transactions: Where to Invest? APPLIED SCIENCES-BASEL 2021. [DOI: 10.3390/app11156766] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Online shopping, already on a steady rise, was propelled even further with the advent of the COVID-19 pandemic. Of course, credit cards are a dominant way of doing business online. The credit card fraud detection problem has become relevant more than ever as the losses due to fraud accumulate. Most research on this topic takes an isolated, focused view of the problem, typically concentrating on tuning the data mining models. We noticed a significant gap between the academic research findings and the rightfully conservative businesses, which are careful when adopting new, especially black-box, models. In this paper, we took a broader perspective and considered this problem from both the academic and the business angle: we detected challenges in the fraud detection problem such as feature engineering and unbalanced datasets and distinguished between more and less lucrative areas to invest in when upgrading fraud detection systems. Our findings are based on the real-world data of CNP (card not present) fraud transactions, which are a dominant type of fraud transactions. Data were provided by our industrial partner, an international card-processing company. We tested different data mining models and approaches to the outlined challenges and compared them to their existing production systems to trace a cost-effective fraud detection system upgrade path.
Collapse
|
19
|
Wassermann S, Cuvelier T, Mulinka P, Casas P. Adaptive and Reinforcement Learning Approaches for Online Network Monitoring and Analysis. IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT 2021. [DOI: 10.1109/tnsm.2020.3037486] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
|
20
|
Data-driven anomaly detection and diagnostics for changeover processes in biopharmaceutical drug product manufacturing. Chem Eng Res Des 2021. [DOI: 10.1016/j.cherd.2020.12.018] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022]
|
21
|
Anowar F, Sadaoui S. Incremental learning framework for real‐world fraud detection environment. Comput Intell 2021. [DOI: 10.1111/coin.12434] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
Affiliation(s)
- Farzana Anowar
- Department of Computer Science University of Regina Regina Saskatchewan Canada
| | - Samira Sadaoui
- Department of Computer Science University of Regina Regina Saskatchewan Canada
| |
Collapse
|
22
|
A divided and prioritized experience replay approach for streaming regression. MethodsX 2021; 8:101571. [PMID: 35004205 PMCID: PMC8720895 DOI: 10.1016/j.mex.2021.101571] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/02/2021] [Accepted: 10/31/2021] [Indexed: 11/21/2022] Open
Abstract
In the streaming learning setting, an agent is presented with a data stream on which to learn from in an online fashion. A common problem is catastrophic forgetting of old knowledge due to updates to the model. Mitigating catastrophic forgetting has received a lot of attention, and a variety of methods exist to solve this problem. In this paper, we present a divided and prioritized experience replay approach for streaming regression, in which relevant observations are retained in the replay, and extra focus is added to poorly estimated observations through prioritization. Using a real-world dataset, the method is compared to the standard sliding window approach. A statistical power analysis is performed, showing how our approach improves performance on rare, important events at a trade-off in performance for more common observations. Close inspections of the dataset are provided, with emphasis on areas where the standard approach fails. A rephrasing of the problem to a binary classification problem is performed to separate common and rare, important events. These results provide an added perspective regarding the improvement made on rare events.We divide the prediction space in a streaming regression setting Observations in the experience replay are prioritized for further training by the model’s current error
Collapse
|
23
|
|
24
|
Abstract
Spam emails classification using data mining and machine learning approaches has enticed the researchers' attention duo to its obvious positive impact in protecting internet users. Several features can be used for creating data mining and machine learning based spam classification models. Yet, spammers know that the longer they will use the same set of features for tricking email users the more probably the anti-spam parties might develop tools for combating this kind of annoying email messages. Spammers, so, adapt by continuously reforming the group of features utilized for composing spam emails. For that reason, even though traditional classification methods possess sound classification results, they were ineffective for lifelong classification of spam emails duo to the fact that they might be prone to the so-called “Concept Drift”. In the current study, an enhanced model is proposed for ensuring lifelong spam classification model. For the evaluation purposes, the overall performance of the suggested model is contrasted against various other stream mining classification techniques. The results proved the success of the suggested model as a lifelong spam emails classification method.
Collapse
|
25
|
Souza VMA, dos Reis DM, Maletzke AG, Batista GEAPA. Challenges in benchmarking stream learning algorithms with real-world data. Data Min Knowl Discov 2020. [DOI: 10.1007/s10618-020-00698-5] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
26
|
Xu D, Shi Y, Tsang IW, Ong YS, Gong C, Shen X. Survey on Multi-Output Learning. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2020; 31:2409-2429. [PMID: 31714241 DOI: 10.1109/tnnls.2019.2945133] [Citation(s) in RCA: 22] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
The aim of multi-output learning is to simultaneously predict multiple outputs given an input. It is an important learning problem for decision-making since making decisions in the real world often involves multiple complex factors and criteria. In recent times, an increasing number of research studies have focused on ways to predict multiple outputs at once. Such efforts have transpired in different forms according to the particular multi-output learning problem under study. Classic cases of multi-output learning include multi-label learning, multi-dimensional learning, multi-target regression, and others. From our survey of the topic, we were struck by a lack in studies that generalize the different forms of multi-output learning into a common framework. This article fills that gap with a comprehensive review and analysis of the multi-output learning paradigm. In particular, we characterize the four Vs of multi-output learning, i.e., volume, velocity, variety, and veracity, and the ways in which the four Vs both benefit and bring challenges to multi-output learning by taking inspiration from big data. We analyze the life cycle of output labeling, present the main mathematical definitions of multi-output learning, and examine the field's key challenges and corresponding solutions as found in the literature. Several model evaluation metrics and popular data repositories are also discussed. Last but not least, we highlight some emerging challenges with multi-output learning from the perspective of the four Vs as potential research directions worthy of further studies.
Collapse
|
27
|
Abstract
Abstract
Trust between agents in multi-agent systems (MASs) is critical to encourage high levels of cooperation. Existing methods to assess trust and reputation use direct and indirect past experiences about an agent to estimate their future performance; however, these will not always be representative if agents change their behaviour over time.
Real-world distributed networks such as online market places, P2P networks, pervasive computing and the Smart Grid can be viewed as MAS. Dynamic agent behaviour in such MAS can arise from seasonal changes, cheaters, supply chain faults, network traffic and many other reasons. However, existing trust and reputation models use limited techniques, such as forgetting factors and sliding windows, to account for dynamic behaviour.
In this paper, we propose Reacting and Predicting in Trust and Reputation (RaPTaR), a method to extend existing trust and reputation models to give agents the ability to monitor the output of interactions with a group of agents over time to identify any likely changes in behaviour and adapt accordingly. Additionally, RaPTaR can provide an a priori estimate of trust when there is little or no interaction data (either because an agent is new or because a detected behaviour change suggests recent past experiences are no longer representative). Our results show that RaPTaR has improved performance compared to existing trust and reputation methods when dynamic behaviour causes the ranking of the best agents to interact with to change.
Collapse
|
28
|
Ensemble-Based Online Machine Learning Algorithms for Network Intrusion Detection Systems Using Streaming Data. INFORMATION 2020. [DOI: 10.3390/info11060315] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
As new cyberattacks are launched against systems and networks on a daily basis, the ability for network intrusion detection systems to operate efficiently in the big data era has become critically important, particularly as more low-power Internet-of-Things (IoT) devices enter the market. This has motivated research in applying machine learning algorithms that can operate on streams of data, trained online or “live” on only a small amount of data kept in memory at a time, as opposed to the more classical approaches that are trained solely offline on all of the data at once. In this context, one important concept from machine learning for improving detection performance is the idea of “ensembles”, where a collection of machine learning algorithms are combined to compensate for their individual limitations and produce an overall superior algorithm. Unfortunately, existing research lacks proper performance comparison between homogeneous and heterogeneous online ensembles. Hence, this paper investigates several homogeneous and heterogeneous ensembles, proposes three novel online heterogeneous ensembles for intrusion detection, and compares their performance accuracy, run-time complexity, and response to concept drifts. Out of the proposed novel online ensembles, the heterogeneous ensemble consisting of an adaptive random forest of Hoeffding Trees combined with a Hoeffding Adaptive Tree performed the best, by dealing with concept drift in the most effective way. While this scheme is less accurate than a larger size adaptive random forest, it offered a marginally better run-time, which is beneficial for online training.
Collapse
|
29
|
Fernandes Lopes J, Santana EJ, Turrisi da Costa VG, Bogaz Zarpelao B, Barbon S. Evaluating the Four-Way Performance Trade-Off for Data Stream Classification in Edge Computing. IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT 2020. [DOI: 10.1109/tnsm.2020.2983921] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
|
30
|
Wu Y, Zhang W, Zhang L, Qiao Y, Yang J, Cheng C. A Multi-Clustering Algorithm to Solve Driving Cycle Prediction Problems Based on Unbalanced Data Sets: A Chinese Case Study. SENSORS 2020; 20:s20092448. [PMID: 32344855 PMCID: PMC7248886 DOI: 10.3390/s20092448] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/07/2020] [Revised: 04/18/2020] [Accepted: 04/21/2020] [Indexed: 11/16/2022]
Abstract
Vehicle evaluation parameters, which are increasingly of concern for governments and consumers, quantify performance indicators, such as vehicle performance, emissions, and driving experience to help guide consumers in purchasing cars. While past approaches for driving cycle prediction have been proven effective and used in many countries, these algorithms are difficult to use in China with its complex traffic environment and increasingly high frequency of traffic jams. Meanwhile, we found that the vehicle dataset used by the driving cycle prediction problem is usually unbalanced in real cases, which means that there are more medium and high speed samples and very few samples at low and ultra-high speeds. If the ordinary clustering algorithm is directly applied to the unbalanced data, it will have a huge impact on the performance to build driving cycle maps, and the parameters of the map will deviate considerable from actual ones. In order to address these issues, this paper propose a novel driving cycle map algorithm framework based on an ensemble learning method named multi-clustering algorithm, to improve the performance of traditional clustering algorithms on unbalanced data sets. It is noteworthy that our model framework can be easily extended to other complicated structure areas due to its flexible modular design and parameter configuration. Finally, we tested our method based on actual traffic data generated in Fujian Province in China. The results prove the multi-clustering algorithm has excellent performance on our dataset.
Collapse
Affiliation(s)
- Yuewei Wu
- Correspondence: ; Tel.: +86-135-2020-2168
| | | | | | | | | | | |
Collapse
|
31
|
Mahdi OA, Pardede E, Ali N, Cao J. Diversity measure as a new drift detection method in data streaming. Knowl Based Syst 2020. [DOI: 10.1016/j.knosys.2019.105227] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
32
|
Goldenberg I, Webb GI. PCA-based drift and shift quantification framework for multidimensional data. Knowl Inf Syst 2020. [DOI: 10.1007/s10115-020-01438-3] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
33
|
|
34
|
Incremental learning of concept drift in Multiple Instance Learning for industrial visual inspection. COMPUT IND 2019. [DOI: 10.1016/j.compind.2019.04.006] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
|
35
|
Sweetlin Hemalatha C, Pathak R, Vaidehi V. Hybrid decision trees for data streams based on Incremental Flexible Naive Bayes prediction at leaf nodes. EVOLUTIONARY INTELLIGENCE 2019. [DOI: 10.1007/s12065-019-00252-3] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
36
|
Folino G, Guarascio M, Papuzzo G. Exploiting fractal dimension and a distributed evolutionary approach to classify data streams with concept drifts. Appl Soft Comput 2019. [DOI: 10.1016/j.asoc.2018.11.009] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|
37
|
Lopez-Lopez E, V. Regueiro C, M. Pardo X, Franco A, Lumini A. Incremental Learning Techniques Within a Self-updating Approach for Face Verification in Video-Surveillance. PATTERN RECOGNITION AND IMAGE ANALYSIS 2019. [DOI: 10.1007/978-3-030-31321-0_3] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
38
|
Vong CM, Du J, Wong CM, Cao JW. Postboosting Using Extended G-Mean for Online Sequential Multiclass Imbalance Learning. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2018; 29:6163-6177. [PMID: 29993897 DOI: 10.1109/tnnls.2018.2826553] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
In this paper, a novel learning method called postboosting using extended G-mean (PBG) is proposed for online sequential multiclass imbalance learning (OS-MIL) in neural networks. PBG is effective due to three reasons. 1) Through postadjusting a classification boundary under extended G-mean, the challenging issue of imbalanced class distribution for sequentially arriving multiclass data can be effectively resolved. 2) A newly derived update rule for online sequential learning is proposed, which produces a high G-mean for current model and simultaneously possesses almost the same information of its previous models. 3) A dynamic adjustment mechanism provided by extended G-mean is valid to deal with the unresolved challenging dense-majority problem and two dynamic changing issues, namely, dynamic changing data scarcity (DCDS) and dynamic changing data diversity (DCDD). Compared to other OS-MIL methods, PBG is highly effective on resolving DCDS, while PBG is the only method to resolve dense-majority and DCDD. Furthermore, PBG can directly and effectively handle unscaled data stream. Experiments have been conducted for PBG and two popular OS-MIL methods for neural networks under massive binary and multiclass data sets. Through the analyses of experimental results, PBG is shown to outperform the other compared methods on all data sets in various aspects including the issues of data scarcity, dense-majority, DCDS, DCDD, and unscaled data.
Collapse
|
39
|
Wang S, Minku LL, Yao X. A Systematic Study of Online Class Imbalance Learning With Concept Drift. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2018; 29:4802-4821. [PMID: 29993955 DOI: 10.1109/tnnls.2017.2771290] [Citation(s) in RCA: 49] [Impact Index Per Article: 8.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
As an emerging research topic, online class imbalance learning often combines the challenges of both class imbalance and concept drift. It deals with data streams having very skewed class distributions, where concept drift may occur. It has recently received increased research attention; however, very little work addresses the combined problem where both class imbalance and concept drift coexist. As the first systematic study of handling concept drift in class-imbalanced data streams, this paper first provides a comprehensive review of current research progress in this field, including current research focuses and open challenges. Then, an in-depth experimental study is performed, with the goal of understanding how to best overcome concept drift in online learning with class imbalance.
Collapse
|
40
|
Goldenberg I, Webb GI. Survey of distance measures for quantifying concept drift and shift in numeric data. Knowl Inf Syst 2018. [DOI: 10.1007/s10115-018-1257-z] [Citation(s) in RCA: 18] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
41
|
Karlsen MR, Moschoyiannis S. Evolution of control with learning classifier systems. APPLIED NETWORK SCIENCE 2018; 3:30. [PMID: 30839802 PMCID: PMC6214302 DOI: 10.1007/s41109-018-0088-x] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 03/06/2018] [Accepted: 07/20/2018] [Indexed: 06/09/2023]
Abstract
In this paper we describe the application of a learning classifier system (LCS) variant known as the eXtended classifier system (XCS) to evolve a set of 'control rules' for a number of Boolean network instances. We show that (1) it is possible to take the system to an attractor, from any given state, by applying a set of 'control rules' consisting of ternary conditions strings (i.e. each condition component in the rule has three possible states; 0, 1 or #) with associated bit-flip actions, and (2) that it is possible to discover such rules using an evolutionary approach via the application of a learning classifier system. The proposed approach builds on learning (reinforcement learning) and discovery (a genetic algorithm) and therefore the series of interventions for controlling the network are determined but are not fixed. System control rules evolve in such a way that they mirror both the structure and dynamics of the system, without having 'direct' access to either.
Collapse
Affiliation(s)
- Matthew R. Karlsen
- Department of Computer Science, Faculty of Engineering and Physical Sciences, University of Surrey, Guildford, GU2 7XH Surrey UK
| | - Sotiris Moschoyiannis
- Department of Computer Science, Faculty of Engineering and Physical Sciences, University of Surrey, Guildford, GU2 7XH Surrey UK
| |
Collapse
|
42
|
|
43
|
Parallel and incremental credit card fraud detection model to handle concept drift and data imbalance. Neural Comput Appl 2018. [DOI: 10.1007/s00521-018-3633-8] [Citation(s) in RCA: 15] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/28/2022]
|
44
|
|
45
|
Ding S, Mirza B, Lin Z, Cao J, Lai X, Nguyen TV, Sepulveda J. Kernel based online learning for imbalance multiclass classification. Neurocomputing 2018. [DOI: 10.1016/j.neucom.2017.02.102] [Citation(s) in RCA: 23] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
|
46
|
|
47
|
Krell MM, Wilshusen N, Seeland A, Kim SK. Classifier transfer with data selection strategies for online support vector machine classification with class imbalance. J Neural Eng 2017; 14:025003. [DOI: 10.1088/1741-2552/aa5166] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022]
|
48
|
Gligorijević V, Malod-Dognin N, Pržulj N. Integrative methods for analyzing big data in precision medicine. Proteomics 2016; 16:741-58. [PMID: 26677817 DOI: 10.1002/pmic.201500396] [Citation(s) in RCA: 98] [Impact Index Per Article: 12.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/08/2015] [Revised: 11/16/2015] [Accepted: 12/09/2015] [Indexed: 12/19/2022]
Abstract
We provide an overview of recent developments in big data analyses in the context of precision medicine and health informatics. With the advance in technologies capturing molecular and medical data, we entered the area of "Big Data" in biology and medicine. These data offer many opportunities to advance precision medicine. We outline key challenges in precision medicine and present recent advances in data integration-based methods to uncover personalized information from big data produced by various omics studies. We survey recent integrative methods for disease subtyping, biomarkers discovery, and drug repurposing, and list the tools that are available to domain scientists. Given the ever-growing nature of these big data, we highlight key issues that big data integration methods will face.
Collapse
Affiliation(s)
| | | | - Nataša Pržulj
- Department of Computing, Imperial College London, London, UK
| |
Collapse
|
49
|
|
50
|
Mirza B, Lin Z. Meta-cognitive online sequential extreme learning machine for imbalanced and concept-drifting data classification. Neural Netw 2016; 80:79-94. [DOI: 10.1016/j.neunet.2016.04.008] [Citation(s) in RCA: 60] [Impact Index Per Article: 7.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/03/2015] [Revised: 04/08/2016] [Accepted: 04/21/2016] [Indexed: 11/25/2022]
|