scholarly journals Triple descent and the two kinds of overfitting: where and why do they appear?*

2021 ◽  
Vol 2021 (12) ◽  
pp. 124002
Author(s):  
Stéphane d’Ascoli ◽  
Levent Sagun ◽  
Giulio Biroli

Abstract A recent line of research has highlighted the existence of a ‘double descent’ phenomenon in deep learning, whereby increasing the number of training examples N causes the generalization error of neural networks (NNs) to peak when N is of the same order as the number of parameters P. In earlier works, a similar phenomenon was shown to exist in simpler models such as linear regression, where the peak instead occurs when N is equal to the input dimension D. Since both peaks coincide with the interpolation threshold, they are often conflated in the literature. In this paper, we show that despite their apparent similarity, these two scenarios are inherently different. In fact, both peaks can co-exist when NNs are applied to noisy regression tasks. The relative size of the peaks is then governed by the degree of nonlinearity of the activation function. Building on recent developments in the analysis of random feature models, we provide a theoretical ground for this sample-wise triple descent. As shown previously, the nonlinear peak at N = P is a true divergence caused by the extreme sensitivity of the output function to both the noise corrupting the labels and the initialization of the random features (or the weights in NNs). This peak survives in the absence of noise, but can be suppressed by regularization. In contrast, the linear peak at N = D is solely due to overfitting the noise in the labels, and forms earlier during training. We show that this peak is implicitly regularized by the nonlinearity, which is why it only becomes salient at high noise and is weakly affected by explicit regularization. Throughout the paper, we compare analytical results obtained in the random feature model with the outcomes of numerical experiments involving deep NNs.

2015 ◽  
Vol 12 (3) ◽  
pp. 961-977 ◽  
Author(s):  
Sinisa Neskovic ◽  
Rade Matic

This paper presents an approach for context modeling in complex self adapted systems consisting of many independent context-aware applications. The contextual information used for adaptation of all system applications is described by an ontology treated as a global context model. A local context model tailored to the specific needs of a particular application is defined as a view over the global context in the form of a feature model. Feature models and their configurations derived from the global context state are then used by a specific dynamic software product line in order to adapt applications at runtime. The main focus of the paper is on the realization of mappings between global and local contexts. The paper describes an overall model architecture and provides corresponding metamodels as well as rules for a mapping between feature models and ontologies.


Author(s):  
Jan Wouter Versluis ◽  
Willem F. Bronsvoort ◽  
Klaas Jan de Kraker ◽  
Kees Seebregts

Abstract New techniques for visualizing feature models are presented. These do not only provide better geometric and spatial model insight than the standard display techniques, but also functional insight into a feature model by visualizing engineering information. Geometry is visualized by combining shaded and line visualization techniques, resulting in clearer images. Among the engineering information that is visualized are feature intersections, closure faces and feature parameters. This is done by explicitly displaying feature geometry instead of the geometry of the final shape only, and by displaying additional information in a model image. Combined, these techniques provide powerful possibilities to visualize feature models. The implementation uses specified feature and camera properties and a cellular geometric datastructure for generating images. The cellular data structure contains the feature geometry information required for feature visualization.


Author(s):  
Rafael Bidarra ◽  
Andre´ van Bunnik ◽  
Willem F. Bronsvoort

Providing advanced 3D interactive facilities to users of a client-server collaborative modeling system presents a great challenge when thin clients are involved, mainly due to their lack of both a full-fledged CAD model and adequate modeling and solving functionalities. This paper presents a new approach that provides a convenient representation of feature model data suitable for direct manipulation of feature models at such clients. In particular, feature handles are proposed to support interactive feature editing. This approach combines all advantages of a thin client approach with the sort of 3D direct manipulation facilities usually only found in powerful standalone CAD systems.


2001 ◽  
Vol 13 (8) ◽  
pp. 1863-1889 ◽  
Author(s):  
Masashi Sugiyama ◽  
Hidemitsu Ogawa

The problem of model selection is considerably important for acquiring higher levels of generalization capability in supervised learning. In this article, we propose a new criterion for model selection, the subspace information criterion (SIC), which is a generalization of Mallows's CL. It is assumed that the learning target function belongs to a specified functional Hilbert space and the generalization error is defined as the Hilbert space squared norm of the difference between the learning result function and target function. SIC gives an unbiased estimate of the generalization error so defined. SIC assumes the availability of an unbiased estimate of the target function and the noise covariance matrix, which are generally unknown. A practical calculation method of SIC for least-mean-squares learning is provided under the assumption that the dimension of the Hilbert space is less than the number of training examples. Finally, computer simulations in two examples show that SIC works well even when the number of training examples is small.


1997 ◽  
Vol 15 (1) ◽  
pp. 101-112 ◽  
Author(s):  
J. Small ◽  
L. Shackleford ◽  
G. Pavey

Abstract. The aim of this paper is to test the effectiveness of feature models in ocean acoustic forecasting. Feature models are simple mathematical representations of the horizontal and vertical structures of ocean features (such as fronts and eddies), and have been used primarily for assimilating new observations into forecasts and for compressing data. In this paper we describe the results of experiments in which the models have been tested in acoustic terms in eddy and frontal environments in the Iceland Faeroes region. Propagation-loss values were obtained with a 2D parabolic-equation (PE) model, for the observed fields, and compared to PE results from the corresponding feature models and horizontally uniform (range-independent) fields. The feature models were found to represent the smoothed observed propagation-loss field to within an rms error of 5 dB for the eddy and 7 dB for the front, compared to 10–15-dB rms errors obtained with the range-independent field. Some of the errors in the feature-model propagation loss were found to be due to high-amplitude 'oceanographic noise' in the field. The main conclusion is that the feature models represent the main acoustic properties of the ocean but do not show the significant effects of small-scale internal waves and fine-structure. It is recommended that feature models be used in conjunction with stochastic models of the internal waves, to represent the complete environmental variability.


Author(s):  
Paulos J. Nyirenda ◽  
Willem F. Bronsvoort

Feature modelling is a valuable technology to define and manipulate product models through advanced operations. Parameters and constraints can be included in feature models to represent particular design intent. In this way, entire families of shapes, not just specific instances, can be modelled. Freeform surface feature modelling, which extends the concept of feature modelling to freeform shapes, still suffers from a number of shortcomings as regards parametric modelling. In particular, it lacks good facilities to add new parameters to a model. We present a multi-level freeform surface feature model that can be used to implement such facilities. The model contains an unevaluated, a partially evaluated and an evaluated level. The paper describes the three levels and how these are generated.


Author(s):  
Mehdi Noorian ◽  
Mohsen Asadi ◽  
Ebrahim Bagheri ◽  
Weichang Du

Software Product Line (SPL) engineering is a systematic reuse-based software development approach which is founded on the idea of building software products using a set of core assets rather than developing individual software systems from scratch. Feature models are among the widely used artefacts for SPL development that mostly capture functional and operational variability of a system. Researchers have argued that connecting intentional variability models such as goal models with feature variability models in a target domain can enrich feature models with valuable quality and non-functional information. Interrelating goal models and feature models has already been proposed in the literature for capturing non-functional properties in software product lines; however, this manual integration process is cumbersome and tedious. In this paper, we propose a (semi) automated approach that systematically integrates feature models and goal models through standard ontologies. Our proposed approach connects feature model and goal model elements through measuring the semantic similarity of their annotated ontological concepts. Our work not only provides the means to systematically interrelate feature models and goal models but also allows domain engineers to identify and model the role and significance of non-functional properties in the domain represented by the feature model.


2011 ◽  
Vol 10 (01) ◽  
pp. 151-158
Author(s):  
LEI WU ◽  
ZHEN WEI

Aiming at the shortcomings of the modeling analysis of traditional Feature-Oriented Analysis Approach under service oriented architecture SOA, and providing more reusability and flexibility to the development of SOA system, this paper makes an improvement on Feature-Oriented Analysis Approach. It introduces the concept of service feature and improves the refinement and interaction description of feature models. On the basis of this, it proposes a method of domain analysis in SOA. In addition, in view of the fact that web services act as a technology available to implement SOA, it presents a method to transform feature model into interface model and composite model of web services. Finally, this method's application in ERP system project in publishing is verified as an example to show that it is feasible to improve software development efficiency.


2021 ◽  
Author(s):  
Shuo Yang ◽  
Songhua Wu ◽  
Tongliang Liu ◽  
Min Xu

A major gap between few-shot and many-shot learning is the data distribution empirically observed by the model during training. In few-shot learning, the learned model can easily become over-fitted based on the biased distribution formed by only a few training examples, while the ground-truth data distribution is more accurately uncovered in many-shot learning to learn a well-generalized model. In this paper, we propose to calibrate the distribution of these few-sample classes to be more unbiased to alleviate such an over-fitting problem. The distribution calibration is achieved by transferring statistics from the classes with sufficient examples to those few-sample classes. After calibration, an adequate number of examples can be sampled from the calibrated distribution to expand the inputs to the classifier. Extensive experiments on three datasets, miniImageNet, tieredImageNet, and CUB, show that a simple linear classifier trained using the features sampled from our calibrated distribution can outperform the state-of-the-art accuracy by a large margin. We also establish a generalization error bound for the proposed distribution-calibration-based few-shot learning, which consists of the distribution assumption error, the distribution approximation error, and the estimation error. This generalization error bound theoretically justifies the effectiveness of the proposed method.


Author(s):  
Maurice Dohmen ◽  
Klaas Jan de Kraker ◽  
Willem F. Bronsvoort

Abstract A new approach to specification and maintenance of feature validity conditions in a multiple-view feature modeling system is presented. Each view of a product contains a feature model. Features are specified declaratively in an object-oriented language, using constraints to specify feature validity conditions. Constraints are also used to specify relations between features. Validation of the feature models is done by a constraint manager that integrates different solving techniques. The constraint graph is mapped by the constraint manager onto constraints that are handled by dedicated solvers. If views are consistent, i.e. their feature models represent the same geometry, feature parameters can be changed. Changes are propagated through link constraints between different views.


Sign in / Sign up

Export Citation Format

Share Document