scholarly journals Principal manifold estimation via model complexity selection

Author(s):  
Kun Meng ◽  
Ani Eloyan
Author(s):  
Thorsten Meiser

Stochastic dependence among cognitive processes can be modeled in different ways, and the family of multinomial processing tree models provides a flexible framework for analyzing stochastic dependence among discrete cognitive states. This article presents a multinomial model of multidimensional source recognition that specifies stochastic dependence by a parameter for the joint retrieval of multiple source attributes together with parameters for stochastically independent retrieval. The new model is equivalent to a previous multinomial model of multidimensional source memory for a subset of the parameter space. An empirical application illustrates the advantages of the new multinomial model of joint source recognition. The new model allows for a direct comparison of joint source retrieval across conditions, it avoids statistical problems due to inflated confidence intervals and does not imply a conceptual imbalance between source dimensions. Model selection criteria that take model complexity into account corroborate the new model of joint source recognition.


2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Nicholas Garside ◽  
Hamed Zaribafzadeh ◽  
Ricardo Henao ◽  
Royce Chung ◽  
Daniel Buckland

AbstractMethods used to predict surgical case time often rely upon the current procedural terminology (CPT) code as a nominal variable to train machine-learned models, however this limits the ability of the model to incorporate new procedures and adds complexity as the number of unique procedures increases. The relative value unit (RVU, a consensus-derived billing indicator) can serve as a proxy for procedure workload and could replace the CPT code as a primary feature for models that predict surgical case length. Using 11,696 surgical cases from Duke University Health System electronic health records data, we compared boosted decision tree models that predict individual case length, changing the method by which the model coded procedure type; CPT, RVU, and CPT–RVU combined. Performance of each model was assessed by inference time, MAE, and RMSE compared to the actual case length on a test set. Models were compared to each other and to the manual scheduler method that currently exists. RMSE for the RVU model (60.8 min) was similar to the CPT model (61.9 min), both of which were lower than scheduler (90.2 min). 65.2% of our RVU model’s predictions (compared to 43.2% from the current human scheduler method) fell within 20% of actual case time. Using RVUs reduced model prediction time by ninefold and reduced the number of training features from 485 to 44. Replacing pre-operative CPT codes with RVUs maintains model performance while decreasing overall model complexity in the prediction of surgical case length.


Sensors ◽  
2021 ◽  
Vol 21 (16) ◽  
pp. 5312
Author(s):  
Yanni Zhang ◽  
Yiming Liu ◽  
Qiang Li ◽  
Jianzhong Wang ◽  
Miao Qi ◽  
...  

Recently, deep learning-based image deblurring and deraining have been well developed. However, most of these methods fail to distill the useful features. What is more, exploiting the detailed image features in a deep learning framework always requires a mass of parameters, which inevitably makes the network suffer from a high computational burden. We propose a lightweight fusion distillation network (LFDN) for image deblurring and deraining to solve the above problems. The proposed LFDN is designed as an encoder–decoder architecture. In the encoding stage, the image feature is reduced to various small-scale spaces for multi-scale information extraction and fusion without much information loss. Then, a feature distillation normalization block is designed at the beginning of the decoding stage, which enables the network to distill and screen valuable channel information of feature maps continuously. Besides, an information fusion strategy between distillation modules and feature channels is also carried out by the attention mechanism. By fusing different information in the proposed approach, our network can achieve state-of-the-art image deblurring and deraining results with a smaller number of parameters and outperform the existing methods in model complexity.


Author(s):  
Qiang Yu ◽  
Feiqiang Liu ◽  
Long Xiao ◽  
Zitao Liu ◽  
Xiaomin Yang

Deep-learning (DL)-based methods are of growing importance in the field of single image super-resolution (SISR). The practical application of these DL-based models is a remaining problem due to the requirement of heavy computation and huge storage resources. The powerful feature maps of hidden layers in convolutional neural networks (CNN) help the model learn useful information. However, there exists redundancy among feature maps, which can be further exploited. To address these issues, this paper proposes a lightweight efficient feature generating network (EFGN) for SISR by constructing the efficient feature generating block (EFGB). Specifically, the EFGB can conduct plain operations on the original features to produce more feature maps with parameters slightly increasing. With the help of these extra feature maps, the network can extract more useful information from low resolution (LR) images to reconstruct the desired high resolution (HR) images. Experiments conducted on the benchmark datasets demonstrate that the proposed EFGN can outperform other deep-learning based methods in most cases and possess relatively lower model complexity. Additionally, the running time measurement indicates the feasibility of real-time monitoring.


2021 ◽  
Vol 8 (1) ◽  
Author(s):  
Federico Calesella ◽  
Alberto Testolin ◽  
Michele De Filippo De Grazia ◽  
Marco Zorzi

AbstractMultivariate prediction of human behavior from resting state data is gaining increasing popularity in the neuroimaging community, with far-reaching translational implications in neurology and psychiatry. However, the high dimensionality of neuroimaging data increases the risk of overfitting, calling for the use of dimensionality reduction methods to build robust predictive models. In this work, we assess the ability of four well-known dimensionality reduction techniques to extract relevant features from resting state functional connectivity matrices of stroke patients, which are then used to build a predictive model of the associated deficits based on cross-validated regularized regression. In particular, we investigated the prediction ability over different neuropsychological scores referring to language, verbal memory, and spatial memory domains. Principal Component Analysis (PCA) and Independent Component Analysis (ICA) were the two best methods at extracting representative features, followed by Dictionary Learning (DL) and Non-Negative Matrix Factorization (NNMF). Consistent with these results, features extracted by PCA and ICA were found to be the best predictors of the neuropsychological scores across all the considered cognitive domains. For each feature extraction method, we also examined the impact of the regularization method, model complexity (in terms of number of features that entered in the model) and quality of the maps that display predictive edges in the resting state networks. We conclude that PCA-based models, especially when combined with L1 (LASSO) regularization, provide optimal balance between prediction accuracy, model complexity, and interpretability.


2021 ◽  
pp. 193229682110156
Author(s):  
Thibault Gautier ◽  
Rupesh Silwal ◽  
Aramesh Saremi ◽  
Anders Boss ◽  
Marc D. Breton

Background: As type 2 diabetes (T2D) progresses, intensification to combination therapies, such as iGlarLixi (a fixed-ratio GLP-1 RA and basal insulin combination), may be required. Here a simulation study was used to assess the effect of iGlarLixi administration timing (am vs pm) on blood sugar profiles. Methods: Models of lixisenatide were built with a selection procedure, optimizing measurement fits and model complexity, and were included in a pre-existing T2D simulation platform containing glargine models. With the resulting tool, a simulated trial was conducted with 100 in-silico participants with T2D. Individuals were given iGLarLixi either before breakfast or before an evening meal for 2 weeks and daily glycemic profiles were analyzed. In the model, breakfast was considered the largest meal of the day. Results: A similar percentage of time within 24 hours was spent with blood sugar levels between 70 to 180 mg/dL when iGlarLixi was administered pre-breakfast or pre-evening meal (73% vs 71%, respectively). Overall percent of time with blood glucose levels above 180 mg/dL within a 24-hour period was similar when iGlarLixi was administered pre-breakfast or pre-evening meal (26% vs 28%, respectively). Rates of hypoglycemia were low in both regimens, with a blood glucose concentration of below 70 mg/dL only observed for 1% of the 24-hour time period for either timing of administration. Conclusions: Good efficacy was observed when iGlarlixi was administered pre-breakfast; however, administration of iGlarlixi pre-evening meal was also deemed to be effective, even though in the model the size of the evening meal was smaller than that of the breakfast.


Author(s):  
Susanne Charlotta Åberg ◽  
Annika Katarina Åberg ◽  
Kirsti Korkka-Niemi

AbstractGreater complexity in three-dimensional (3D) model structures yields more plausible groundwater recharge/discharge patterns, especially in groundwater/surface-water interactions. The construction of a 3D hydrostratigraphic model prior to flow modelling is beneficial when the hydraulic conductivity of geological units varies considerably. A workflow for 3D hydrostratigraphic modelling with Leapfrog Geo and flow modelling with MODFLOW-NWT was developed. It was used to evaluate how the modelling results for groundwater flow and recharge/discharge patterns differ when using simple or more complex hydrostratigraphic models. The workflow was applied to a study site consisting of complex Quaternary sediments underlain by fractured and weathered crystalline bedrock. Increasing the hydrostratigraphic detail appeared to improve the fit between the observed and simulated water table, and created more plausible groundwater flow patterns. Interlayered zones of low and high conductivity disperse the recharge/discharge patterns, increasing the vertical flow component. Groundwater flow was predominantly horizontal in models in which Quaternary sediments and bedrock were simplified as one layer per unit. It appears to be important to define the interlayered low-conductivity units, which can limit groundwater infiltration and also affect groundwater discharge patterns. Explicit modelling with Leapfrog Geo was found to be effective but time-consuming in the generation of scattered and thin-layered strata.


2021 ◽  
Vol 2021 (3) ◽  
Author(s):  
Thomas G. Rizzo ◽  
George N. Wojcik

Abstract Extra dimensions have proven to be a very useful tool in constructing new physics models. In earlier work, we began investigating toy models for the 5-D analog of the kinetic mixing/vector portal scenario where the interactions of dark matter, taken to be, e.g., a complex scalar, with the brane-localized fields of the Standard Model (SM) are mediated by a massive U(1)D dark photon living in the bulk. These models were shown to have many novel features differentiating them from their 4-D analogs and which, in several cases, avoided some well-known 4-D model building constraints. However, these gains were obtained at the cost of the introduction of a fair amount of model complexity, e.g., dark matter Kaluza-Klein excitations. In the present paper, we consider an alternative setup wherein the dark matter and the dark Higgs, responsible for U(1)D breaking, are both localized to the ‘dark’ brane at the opposite end of the 5-D interval from where the SM fields are located with only the dark photon now being a 5-D field. The phenomenology of such a setup is explored for both flat and warped extra dimensions and compared to the previous more complex models.


2020 ◽  
pp. 1-11
Author(s):  
Dawei Yu ◽  
Jie Yang ◽  
Yun Zhang ◽  
Shujuan Yu

The Densely Connected Network (DenseNet) has been widely recognized as a highly competitive architecture in Deep Neural Networks. And its most outstanding property is called Dense Connections, which represent each layer’s input by concatenating all the preceding layers’ outputs and thus improve the performance by encouraging feature reuse to the extreme. However, it is Dense Connections that cause the challenge of dimension-enlarging, making DenseNet very resource-intensive and low efficiency. In the light of this, inspired by the Residual Network (ResNet), we propose an improved DenseNet named Additive DenseNet, which features replacing concatenation operations (used in Dense Connections) with addition operations (used in ResNet), and in terms of feature reuse, it upgrades addition operations to accumulating operations (namely ∑ (·)), thus enables each layer’s input to be the summation of all the preceding layers’ outputs. Consequently, Additive DenseNet can not only preserve the dimension of input from enlarging, but also retain the effect of Dense Connections. In this paper, Additive DenseNet is applied to text classification task. The experimental results reveal that compared to DenseNet, our Additive DenseNet can reduce the model complexity by a large margin, such as GPU memory usage and quantity of parameters. And despite its high resource economy, Additive DenseNet can still outperform DenseNet on 6 text classification datasets in terms of accuracy and show competitive performance for model training.


2020 ◽  
Vol 12 (8) ◽  
pp. 1319
Author(s):  
Xiaofan Sun ◽  
Bingnan Wang ◽  
Maosheng Xiang ◽  
Liangjiang Zhou ◽  
Shuai Jiang

The Gaussian vertical backscatter (GVB) model has a pivotal role in describing the forest vertical structure more accurately, which is reflected by P-band polarimetric interferometric synthetic aperture radar (Pol-InSAR) with strong penetrability. The model uses a three-dimensional parameter space (forest height, Gaussian mean representing the strongest backscattered power elevation, and the corresponding standard deviation) to interpret the forest vertical structure. This paper establishes a two-dimensional GVB model by simplifying the three-dimensional one. Specifically, the two-dimensional GVB model includes the following three cases: the Gaussian mean is located at the bottom of the canopy, the Gaussian mean is located at the top of the canopy, as well as a constant volume profile. In the first two cases, only the forest height and the Gaussian standard deviation are variable. The above approximation operation generates a two-dimensional volume only coherence solution space on the complex plane. Based on the established two-dimensional GVB model, the three-baseline inversion is achieved without the null ground-to-volume ratio assumption. The proposed method improves the performance by 18.62% compared to the three-baseline Random Volume over Ground (RVoG) model inversion. In particular, in the area where the radar incidence angle is less than 0.6 rad, the proposed method improves the inversion accuracy by 34.71%. It suggests that the two-dimensional GVB model reduces the GVB model complexity while maintaining a strong description ability.


Sign in / Sign up

Export Citation Format

Share Document