Improving model selection by nonconvergent methods

التفاصيل البيبلوغرافية
العنوان: Improving model selection by nonconvergent methods
المؤلفون: F. Hergert, Hans-Georg Zimmermann, W. Finnoff
المصدر: Neural Networks. 6:771-783
بيانات النشر: Elsevier BV, 1993.
سنة النشر: 1993
مصطلحات موضوعية: Mathematical optimization, Artificial neural network, business.industry, Cognitive Neuroscience, Model selection, Overfitting, Machine learning, computer.software_genre, Regularization (mathematics), Cross-validation, Regression, Weighting, Artificial Intelligence, Artificial intelligence, business, computer, Statistical hypothesis testing, Mathematics
الوصف: Many techniques for model selection in the field of neural networks correspond to well established statistical methods. For example, architecture modifications based on test variables calculated after convergence of the training process can be viewed as part of a hypothesis testing search, and the use of complexity penalty terms is essentially a type of regularization or biased regression. The method of ''stopped'' or ''cross-validation'' training, on the other hand, in which an oversized network is trained until the error on a further validation set of examples deteriorates, then training is stopped, is a true innovation since model selection doesn't require convergence of the training process. Here, the training process is used to perform a directed search of the parameter space for a model which doesn't overfit the data and thus demonstrates superior generalization performance. In this paper we show that this performance can be significantly enhanced by expanding the ''nonconvergent method'' of stopped training to include dynamic topology modifications (dynamic weight pruning) and modified complexity penalty term methods in which the weighting of the penalty term is adjusted during the training process. On an extensive sequence of simulation examples we demonstrate the general superiority of the ''extended'' nonconvergent methods compared to classical penalty term methods, simple stopped training, and methods which only vary the number of hidden units.
تدمد: 0893-6080
الوصول الحر: https://explore.openaire.eu/search/publication?articleId=doi_________::77423d7a15e4cbe27360b0512e01ca0fTest
https://doi.org/10.1016/s0893-6080Test(05)80122-4
حقوق: CLOSED
رقم الانضمام: edsair.doi...........77423d7a15e4cbe27360b0512e01ca0f
قاعدة البيانات: OpenAIRE