statistical framework
Recently Published Documents


TOTAL DOCUMENTS

913
(FIVE YEARS 284)

H-INDEX

56
(FIVE YEARS 9)

2022 ◽  
Vol 14 (2) ◽  
pp. 904
Author(s):  
William O. Taylor ◽  
Peter L. Watson ◽  
Diego Cerrai ◽  
Emmanouil Anagnostou

This paper develops a statistical framework to analyze the effectiveness of vegetation management at reducing power outages during storms of varying severity levels. The framework was applied on the Eversource Energy distribution grid in Connecticut, USA based on 173 rain and wind events from 2005–2020, including Hurricane Irene, Hurricane Sandy, and Tropical Storm Isaias. The data were binned by storm severity (high/low) and vegetation management levels, where a maximum applicable length of vegetation management for each circuit was determined, and the data were divided into four bins based on the actual length of vegetation management performed divided by the maximum applicable value (0–25%, 25–50%, 50–75%, and 75–100%). Then, weather and overhead line length normalized outage statistics were taken for each group. The statistics were used to determine the effectiveness of vegetation management and its dependence on storm severity. The results demonstrate a higher reduction in damages for lower-severity storms, with a reduction in normalized outages between 45.8% and 63.8%. For high-severity events, there is a large increase in effectiveness between the highest level of vegetation management and the two lower levels, with 75–100% vegetation management leading to a 37.3% reduction in trouble spots. Yet, when evaluating system reliability, it is important to look at all storms combined, and the results of this study provide useful information on total annual trouble spots and allow for analysis of how various vegetation management scenarios would impact trouble spots in the electric grid. This framework can also be used to better understand how more rigorous vegetation management standards (applying ETT) help reduce outages at an individual event level. In future work, a similar framework may be used to evaluate other resilience improvements.


2022 ◽  

Species delimitation is the process of determining whether a group of sampled individuals belong to the same species or to different species. The criteria used to delimit species differ across taxonomic groups, and the methods for delimiting species have changed over time, with a dramatic rise in the popularity of genomic approaches recently. Because inferred species boundaries have ramifications that extend beyond systematics, affecting all fields that rely upon species as a foundational unit, controversy has unsurprisingly surrounded not only the practices used to delimit species boundaries, but also the idea of what species are, which varies across taxa (e.g., the use of subspecies varies across the tree of life). This lack of consensus has no doubt contributed to the appeal of genetic-based delimitation. Specifically, genomic data can be collected from any taxon. Moreover, it can be analyzed in a common statistical framework (as popularized by the multispecies coalescent as a model for species delimitation). With the ease of collecting genetic data, the power of genomics, and the purported standardization for diagnosing species limits, genetic-based species delimitation is displacing traditional time-honored (albeit time-consuming) taxonomic practices of species diagnosis. It has also become an invaluable tool for discovering species in understudied groups, and genetic-based approaches are the foundation of international endeavors to generate a catalogue of DNA barcodes to illuminate biodiversity for all of life on the planet. Yet, genomic applications, and especially the sole reliance upon genetic data for inferring species boundaries, are not without their own set of challenges.


2022 ◽  
Vol 15 (1) ◽  
pp. 45-73
Author(s):  
Andrew Zammit-Mangion ◽  
Michael Bertolacci ◽  
Jenny Fisher ◽  
Ann Stavert ◽  
Matthew Rigby ◽  
...  

Abstract. WOMBAT (the WOllongong Methodology for Bayesian Assimilation of Trace-gases) is a fully Bayesian hierarchical statistical framework for flux inversion of trace gases from flask, in situ, and remotely sensed data. WOMBAT extends the conventional Bayesian synthesis framework through the consideration of a correlated error term, the capacity for online bias correction, and the provision of uncertainty quantification on all unknowns that appear in the Bayesian statistical model. We show, in an observing system simulation experiment (OSSE), that these extensions are crucial when the data are indeed biased and have errors that are spatio-temporally correlated. Using the GEOS-Chem atmospheric transport model, we show that WOMBAT is able to obtain posterior means and variances on non-fossil-fuel CO2 fluxes from Orbiting Carbon Observatory-2 (OCO-2) data that are comparable to those from the Model Intercomparison Project (MIP) reported in Crowell et al. (2019). We also find that WOMBAT's predictions of out-of-sample retrievals obtained from the Total Column Carbon Observing Network (TCCON) are, for the most part, more accurate than those made by the MIP participants.


2022 ◽  
Vol 21 (1) ◽  
Author(s):  
Irene Kyomuhangi ◽  
Emanuele Giorgi

Abstract Background In malaria serology analysis, the standard approach to obtain seroprevalence, i.e the proportion of seropositive individuals in a population, is based on a threshold which is used to classify individuals as seropositive or seronegative. The choice of this threshold is often arbitrary and is based on methods that ignore the age-dependency of the antibody distribution. Methods Using cross-sectional antibody data from the Western Kenyan Highlands, this paper introduces a novel approach that has three main advantages over the current threshold-based approach: it avoids the use of thresholds; it accounts for the age dependency of malaria antibodies; and it allows us to propagate the uncertainty from the classification of individuals into seropositive and seronegative when estimating seroprevalence. The reversible catalytic model is used as an example for illustrating how to propagate this uncertainty into the parameter estimates of the model. Results This paper finds that accounting for age-dependency leads to a better fit to the data than the standard approach which uses a single threshold across all ages. Additionally, the paper also finds that the proposed threshold-free approach is more robust against the selection of different age-groups when estimating seroprevalence. Conclusion The novel threshold-free approach presented in this paper provides a statistically principled and more objective approach to estimating malaria seroprevalence. The introduced statistical framework also provides a means to compare results across studies which may use different age ranges for the estimation of seroprevalence.


2021 ◽  
pp. jnnp-2021-327211
Author(s):  
Anna K Bonkhoff ◽  
Tom Hope ◽  
Danilo Bzdok ◽  
Adrian G Guggisberg ◽  
Rachel L Hawe ◽  
...  

IntroductionStroke causes different levels of impairment and the degree of recovery varies greatly between patients. The majority of recovery studies are biased towards patients with mild-to-moderate impairments, challenging a unified recovery process framework. Our aim was to develop a statistical framework to analyse recovery patterns in patients with severe and non-severe initial impairment and concurrently investigate whether they recovered differently.MethodsWe designed a Bayesian hierarchical model to estimate 3–6 months upper limb Fugl-Meyer (FM) scores after stroke. When focusing on the explanation of recovery patterns, we addressed confounds affecting previous recovery studies and considered patients with FM-initial scores <45 only. We systematically explored different FM-breakpoints between severe/non-severe patients (FM-initial=5–30). In model comparisons, we evaluated whether impairment-level-specific recovery patterns indeed existed. Finally, we estimated the out-of-sample prediction performance for patients across the entire initial impairment range.ResultsRecovery data was assembled from eight patient cohorts (n=489). Data were best modelled by incorporating two subgroups (breakpoint: FM-initial=10). Both subgroups recovered a comparable constant amount, but with different proportional components: severely affected patients recovered more the smaller their impairment, while non-severely affected patients recovered more the larger their initial impairment. Prediction of 3–6 months outcomes could be done with an R2=63.5% (95% CI=51.4% to 75.5%).ConclusionsOur work highlights the benefit of simultaneously modelling recovery of severely-to-non-severely impaired patients and demonstrates both shared and distinct recovery patterns. Our findings provide evidence that the severe/non-severe subdivision in recovery modelling is not an artefact of previous confounds. The presented out-of-sample prediction performance may serve as benchmark to evaluate promising biomarkers of stroke recovery.


2021 ◽  
Author(s):  
François Blanquart ◽  
Nathanaël Hozé ◽  
Benjamin John Cowling ◽  
Florence Débarre ◽  
Simon Cauchemez

Evaluating the characteristics of emerging SARS-CoV-2 variants of concern is essential to inform pandemic risk assessment. A variant may grow faster if it produces a larger number of secondary infections (transmissibility advantage) or if the timing of secondary infections (generation time) is better. So far, assessments have largely focused on deriving the transmissibility advantage assuming the generation time was unchanged. Yet, knowledge of both is needed to anticipate impact. Here we develop an analytical framework to investigate the contribution of both the transmissibility advantage and generation time to the growth advantage of a variant. We find that the growth advantage depends on the epidemiological context (level of epidemic control). More specifically, variants conferring earlier transmission are more strongly favoured when the historical strains have fast epidemic growth, while variants conferring later transmission are more strongly favoured when historical strains have slow or negative growth. We develop these conceptual insights into a statistical framework to infer both the transmissibility advantage and generation time of a variant. On simulated data, our framework correctly estimates both parameters when it covers time periods characterized by different epidemiological contexts. Applied to data for the Alpha and Delta variants in England and in Europe, we find that Alpha confers a +54% [95% CI, 45-63%] transmissibility advantage compared to previous strains, and Delta +140% [98-182%] compared to Alpha, and mean generation times are similar to historical strains for both variants. This work helps interpret variant frequency and will strengthen risk assessment for future variants of concern.


2021 ◽  
Author(s):  
R. Sanjjey ◽  
S. Abisheak ◽  
T.R. Dineshkumar ◽  
M. Kirthan ◽  
S. Sivasaravanababu

This work advances the state-of-art secured WBAN system and QR pattern enabled authentication for privacy measures. An attempt was made to integrate all the above process to build high performance WBAN system. In this work, a comprehensive statistical framework is developed with randomized key generation and secured cipher transformation for secured sensor node communication. We create primary colour channels based on three different QR codes that are widely used for colour printing and complementary channels for capturing colour images. Last but not least, we produced a colour QR pattern.


2021 ◽  
Author(s):  
Pierre Bost ◽  
Daniel Schulz ◽  
Stefanie Engler ◽  
Clive Wasserfall ◽  
Bernd Bodenmiller

Recent advances in multiplexed imaging methods allow simultaneous detection of dozens of proteins or RNAs enabling deep spatial characterization of both healthy and tumor tissues. Parameters for design of optimal sequencing-based experiments have been established, but such parameters are lacking for multiplex imaging studies. Here, using a spatial transcriptomic atlas of healthy and tumor human tissues, we developed a new statistical framework that determines the number of fields of view necessary to accurately identify all cell types that are part of the tissue. Using this strategy on imaging mass cytometry data, we identified a measurement of tissue spatial segregation that enables optimal experimental design and that is technology invariant. This strategy will enable significantly improved design of multiplexed imaging studies.


2021 ◽  
Vol 2022 (1) ◽  
pp. 460-480
Author(s):  
Bogdan Kulynych ◽  
Mohammad Yaghini ◽  
Giovanni Cherubin ◽  
Michael Veale ◽  
Carmela Troncoso

Abstract A membership inference attack (MIA) against a machine-learning model enables an attacker to determine whether a given data record was part of the model’s training data or not. In this paper, we provide an in-depth study of the phenomenon of disparate vulnerability against MIAs: unequal success rate of MIAs against different population subgroups. We first establish necessary and sufficient conditions for MIAs to be prevented, both on average and for population subgroups, using a notion of distributional generalization. Second, we derive connections of disparate vulnerability to algorithmic fairness and to differential privacy. We show that fairness can only prevent disparate vulnerability against limited classes of adversaries. Differential privacy bounds disparate vulnerability but can significantly reduce the accuracy of the model. We show that estimating disparate vulnerability by naïvely applying existing attacks can lead to overestimation. We then establish which attacks are suitable for estimating disparate vulnerability, and provide a statistical framework for doing so reliably. We conduct experiments on synthetic and real-world data finding significant evidence of disparate vulnerability in realistic settings.


Sign in / Sign up

Export Citation Format

Share Document