1
|
Yamauchi K. Quick continual kernel learning on bounded memory space based on balancing between adaptation and forgetting. EVOLVING SYSTEMS 2022. [DOI: 10.1007/s12530-022-09476-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/31/2022]
|
2
|
Blondé L, Strasser P, Kalousis A. Lipschitzness is all you need to tame off-policy generative adversarial imitation learning. Mach Learn 2022; 111:1431-1521. [PMID: 35602587 PMCID: PMC9114147 DOI: 10.1007/s10994-022-06144-5] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2020] [Revised: 01/18/2022] [Accepted: 01/27/2022] [Indexed: 11/21/2022]
Abstract
Despite the recent success of reinforcement learning in various domains, these approaches remain, for the most part, deterringly sensitive to hyper-parameters and are often riddled with essential engineering feats allowing their success. We consider the case of off-policy generative adversarial imitation learning, and perform an in-depth review, qualitative and quantitative, of the method. We show that forcing the learned reward function to be local Lipschitz-continuous is a sine qua non condition for the method to perform well. We then study the effects of this necessary condition and provide several theoretical results involving the local Lipschitzness of the state-value function. We complement these guarantees with empirical evidence attesting to the strong positive effect that the consistent satisfaction of the Lipschitzness constraint on the reward has on imitation performance. Finally, we tackle a generic pessimistic reward preconditioning add-on spawning a large class of reward shaping methods, which makes the base method it is plugged into provably more robust, as shown in several additional theoretical guarantees. We then discuss these through a fine-grained lens and share our insights. Crucially, the guarantees derived and reported in this work are valid for any reward satisfying the Lipschitzness condition, nothing is specific to imitation. As such, these may be of independent interest.
Collapse
|
3
|
Hastie T, Montanari A, Rosset S, Tibshirani RJ. SURPRISES IN HIGH-DIMENSIONAL RIDGELESS LEAST SQUARES INTERPOLATION. Ann Stat 2022; 50:949-986. [PMID: 36120512 PMCID: PMC9481183 DOI: 10.1214/21-aos2133] [Citation(s) in RCA: 11] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/25/2023]
Abstract
Interpolators-estimators that achieve zero training error-have attracted growing attention in machine learning, mainly because state-of-the art neural networks appear to be models of this type. In this paper, we study minimum ℓ 2 norm ("ridgeless") interpolation least squares regression, focusing on the high-dimensional regime in which the number of unknown parameters p is of the same order as the number of samples n. We consider two different models for the feature distribution: a linear model, where the feature vectors x i ∈ ℝ p are obtained by applying a linear transform to a vector of i.i.d. entries, x i = Σ1/2 z i (with z i ∈ ℝ p ); and a nonlinear model, where the feature vectors are obtained by passing the input through a random one-layer neural network, xi = φ(Wz i ) (with z i ∈ ℝ d , W ∈ ℝ p × d a matrix of i.i.d. entries, and φ an activation function acting componentwise on Wz i ). We recover-in a precise quantitative way-several phenomena that have been observed in large-scale neural networks and kernel machines, including the "double descent" behavior of the prediction risk, and the potential benefits of overparametrization.
Collapse
Affiliation(s)
- Trevor Hastie
- Department of Statistics and Department of Biomedical Data Science, Stanford University
| | - Andrea Montanari
- Department of Statistics and Department of Electrical Engineering, Stanford University
| | | | - Ryan J. Tibshirani
- Department of Statistics and Department of Machine Learning, Carnegie Mellon University
| |
Collapse
|
4
|
Lin P, Lyu DC, Chen F, Wang SS, Tsao Y. Multi-style learning with denoising autoencoders for acoustic modeling in the internet of things (IoT). COMPUT SPEECH LANG 2017. [DOI: 10.1016/j.csl.2017.02.001] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
|
5
|
|
6
|
Yamauchi K. Incremental Learning on a Budget and its Application to Quick Maximum Power Point Tracking of Photovoltaic Systems. JOURNAL OF ADVANCED COMPUTATIONAL INTELLIGENCE AND INTELLIGENT INFORMATICS 2014. [DOI: 10.20965/jaciii.2014.p0682] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Recent improvements in embedded systems has enabled learning algorithms to provide realistic solutions for system identification problems. Existing learning algorithms, however, continue to have limitations in learning on embedded systems, where physical memory space is constrained. To overcome this problem, we propose a Limited General Regression Neural Network (LGRNN), which is a variation of general regression neural network proposed by Specht or of simplified fuzzy inference systems. The LGRNN continues incremental learning even if the number of instances exceeds the maximum number of kernels in the LGRNN. We demonstrate LGRNN advantages by comparing it to other kernel-based perceptron learning methods. We also propose a light-weighted LGRNN algorithm, -LGRNNLight- for reducing computational complexity. As an example of its application, we present a Maximum Power Point Tracking (MPPT) microconverter for photovoltaic power generation systems. MPPT is essential for improving the efficiency of renewable energy systems. Although various techniques exist that can realize MPPT, few techniques are able to realize quick control using conventional circuit design. The LGRNN enables the MPPT converter to be constructed at low cost using the conventional combination of a chopper circuit and microcomputer control. The LGRNN learns the Maximum Power Point (MPP) found by Perturb and Observe (P&O), and immediately sets the converter reference voltage after a sudden irradiation change. By using this strategy, the MPPT quickly responds without a predetermination of parameters. The experimental results suggest that, after learning, the proposed converter controls a chopper circuit within 14 ms after a sudden irradiation change. This rapid response property is suitable for efficient power generation, even under shadow flicker conditions that often occur in solar panels located near large wind turbines.
Collapse
|
7
|
Wu Y, Wang H, Zhang B, Du KL. Using Radial Basis Function Networks for Function Approximation and Classification. ACTA ACUST UNITED AC 2012. [DOI: 10.5402/2012/324194] [Citation(s) in RCA: 94] [Impact Index Per Article: 7.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
Abstract
The radial basis function (RBF) network has its foundation in the conventional approximation theory. It has the capability of universal approximation. The RBF network is a popular alternative to the well-known multilayer perceptron (MLP), since it has a simpler structure and a much faster training process. In this paper, we give a comprehensive survey on the RBF network and its learning. Many aspects associated with the RBF network, such as network structure, universal approimation capability, radial basis functions, RBF network learning, structure optimization, normalized RBF networks, application to dynamic system modeling, and nonlinear complex-valued signal processing, are described. We also compare the features and capability of the two models.
Collapse
Affiliation(s)
- Yue Wu
- Enjoyor Laboratories, Enjoyor Inc., Hangzhou 310030, China
| | - Hui Wang
- Enjoyor Laboratories, Enjoyor Inc., Hangzhou 310030, China
| | - Biaobiao Zhang
- Enjoyor Laboratories, Enjoyor Inc., Hangzhou 310030, China
| | - K.-L. Du
- Enjoyor Laboratories, Enjoyor Inc., Hangzhou 310030, China
- Department of Electrical and Computer Engineering, Concordia University, Montreal, QC, Canada H3G 1M8
| |
Collapse
|
8
|
|
9
|
Grandvalet Y. Anisotropic noise injection for input variables relevance determination. ACTA ACUST UNITED AC 2010; 11:1201-12. [PMID: 18249847 DOI: 10.1109/72.883393] [Citation(s) in RCA: 12] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
There are two archetypal ways to control the complexity of a flexible regressor: subset selection and ridge regression. In neural-networks jargon, they are, respectively, known as pruning and weight decay. These techniques may also be adapted to estimate which features of the input space are relevant for predicting the output variables. Relevance is given by a binary indicator for subset selection, and by a continuous rating for ridge regression. This paper shows how to achieve such a rating for a multilayer perceptron trained with noise (or jitter). Noise injection (NI) is modified in order to penalize heavily irrelevant features. The proposed algorithm is attractive as it requires the tuning of a single parameter. This parameter controls the complexity of the model (effective number of parameters) together with the rating of feature relevances (effective input space dimension). Bounds on the effective number of parameters support that the stability of this adaptive scheme is enforced by the constraints applied to the admissible set of relevance indices. The good properties of the algorithm are confirmed by satisfactory experimental results on simulated data sets.
Collapse
Affiliation(s)
- Y Grandvalet
- Heudiasyc, UMR CNRS 6599, Université de Technologie de Compiègne, 60205 Compiègne Cedex, France.
| |
Collapse
|
10
|
Tung WL, Quek C. A brain-inspired fuzzy semantic memory model for learning and reasoning with uncertainty. Neural Comput Appl 2007. [DOI: 10.1007/s00521-007-0101-2] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
11
|
da Silva FLB, Olivares-Rivas W, Colmenares PJ. Basic statistics and variational concepts behind the reverse Monte Carlo technique. MOLECULAR SIMULATION 2007. [DOI: 10.1080/08927020701361884] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
|
12
|
Adaptive noise injection for input variables relevance determination. ACTA ACUST UNITED AC 2005. [DOI: 10.1007/bfb0020198] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register]
|
13
|
Gas B, Zarader J, Chavy C, Chetouani M. Discriminant neural predictive coding applied to phoneme recognition. Neurocomputing 2004. [DOI: 10.1016/j.neucom.2002.08.001] [Citation(s) in RCA: 16] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
14
|
Dominguez M, Jacobs RA. Developmental constraints aid the acquisition of binocular disparity sensitivities. Neural Comput 2003; 15:161-82. [PMID: 12590824 DOI: 10.1162/089976603321043748] [Citation(s) in RCA: 17] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
This article considers the hypothesis that systems learning aspects of visual perception may benefit from the use of suitably designed developmental progressions during training. We report the results of simulations in which four models were trained to detect binocular disparities in pairs of visual images. Three of the models were developmental models in the sense that the nature of their visual input changed during the course of training. These models received a relatively impoverished visual input early in training, and the quality of this input improved as training progressed. One model used a coarse-scale-to-multiscale developmental progression, another used a fine-scale-to-multiscale progression, and the third used a random progression. The final model was nondevelopmental in the sense that the nature of its input remained the same throughout the training period. The simulation results show that the two developmental models whose progressions were organized by spatial frequency content consistently outperformed the nondevelopmental and random developmental models. We speculate that the superior performance of these two models is due to two important features of their developmental progressions: (1) these models were exposed to visual inputs at a single scale early in training, and (2) the spatial scale of their inputs progressed in an orderly fashion from one scale to a neighboring scale during training. Simulation results consistent with these speculations are presented. We conclude that suitably designed developmental sequences can be useful to systems learning to detect binocular disparities. The idea that visual development can aid visual learning is a viable hypothesis in need of study.
Collapse
Affiliation(s)
- Melissa Dominguez
- Department of Computer Science, University of Rochester, Rochester, NY 14627, USA.
| | | |
Collapse
|
15
|
Man-Wai Mak, Sun-Yuan Kung. Estimation of elliptical basis function parameters by the EM algorithm with application to speaker verification. ACTA ACUST UNITED AC 2000; 11:961-9. [DOI: 10.1109/72.857775] [Citation(s) in RCA: 55] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
|
16
|
Chuan Wang, Principe J. Training neural networks with additive noise in the desired signal. ACTA ACUST UNITED AC 1999; 10:1511-7. [DOI: 10.1109/72.809097] [Citation(s) in RCA: 56] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
17
|
|
18
|
Abstract
Noise injection consists of adding noise to the inputs during neural network training. Experimental results suggest that it might improve the generalization ability of the resulting neural network. A justification of this improvement remains elusive: describing analytically the average perturbed cost function is difficult, and controlling the fluctuations of the random perturbed cost function is hard. Hence, recent papers suggest replacing the random perturbed cost by a (deterministic) Taylor approximation of the average perturbed cost function. This article takes a different stance: when the injected noise is gaussian, noise injection is naturally connected to the action of the heat kernel. This provides indications on the relevance domain of traditional Taylor expansions and shows the dependence of the quality of Taylor approximations on global smoothness properties of neural networks under consideration. The connection between noise injection and heat kernel also enables controlling the fluctuations of the random perturbed cost function. Under the global smoothness assumption, tools from gaussian analysis provide bounds on the tail behavior of the perturbed cost. This finally suggests that mixing input perturbation with smoothness-based penalization might be profitable.
Collapse
Affiliation(s)
- Yves Grandvalet
- CNRS UMR 6599 Heudiasyc, Université de Technologie de Compiègne, Compiègne, France
| | - Stéphane Canu
- CNRS UMR 6599 Heudiasyc, Université de Technologie de Compiègne, Compiègne, France
| | | |
Collapse
|
19
|
Use of some sensitivity criteria for choosing networks with good generalization ability. Neural Process Lett 1995. [DOI: 10.1007/bf02309007] [Citation(s) in RCA: 110] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
|