sample weights
Recently Published Documents


TOTAL DOCUMENTS

75
(FIVE YEARS 13)

H-INDEX

12
(FIVE YEARS 2)

2021 ◽  
Vol 0 (0) ◽  
Author(s):  
Brian Erard

Abstract Although one often has detailed information about participants in a program, the lack of comparable information on non-participants precludes standard qualitative choice estimation. This challenge can be overcome by incorporating a supplementary sample of covariate values from the general population. This paper presents new estimators based on this sampling strategy, which perform comparably to the best existing supplementary sampling estimators. The key advantage of the new estimators is that they readily incorporate sample weights, so that they can be applied to Census surveys and other supplementary data sources that have been generated using complex sample designs. This substantially widens the range of problems that can be addressed under a supplementary sampling estimation framework. The potential for improving precision by incorporating imperfect knowledge of the population prevalence rate is also explored.


2021 ◽  
Vol 110 ◽  
pp. 107585
Author(s):  
Carlos Santiago ◽  
Catarina Barata ◽  
Michele Sasdelli ◽  
Gustavo Carneiro ◽  
Jacinto C. Nascimento

Author(s):  
Xiao Wang ◽  
Shaohua Fan ◽  
Kun Kuang ◽  
Chuan Shi ◽  
Jiawei Liu ◽  
...  

Most of existing clustering algorithms are proposed without considering the selection bias in data. In many real applications, however, one cannot guarantee the data is unbiased. Selection bias might bring the unexpected correlation between features and ignoring those unexpected correlations will hurt the performance of clustering algorithms. Therefore, how to remove those unexpected correlations induced by selection bias is extremely important yet largely unexplored for clustering. In this paper, we propose a novel Decorrelation regularized K-Means algorithm (DCKM) for clustering with data selection bias. Specifically, the decorrelation regularizer aims to learn the global sample weights which are capable of balancing the sample distribution, so as to remove unexpected correlations among features. Meanwhile, the learned weights are combined with k-means, which makes the reweighted k-means cluster on the inherent data distribution without unexpected correlation influence. Moreover, we derive the updating rules to effectively infer the parameters in DCKM. Extensive experiments results on real world datasets well demonstrate that our DCKM algorithm achieves significant performance gains, indicating the necessity of removing unexpected feature correlations induced by selection bias when clustering.


2020 ◽  
Vol 391 ◽  
pp. 325-333
Author(s):  
Manuel Castejón-Limas ◽  
Hector Alaiz-Moreton ◽  
Laura Fernández-Robles ◽  
Javier Alfonso-Cendón ◽  
Camino Fernández-Llamas ◽  
...  

2020 ◽  
Vol 34 (07) ◽  
pp. 11685-11692
Author(s):  
Zili Liu ◽  
Tu Zheng ◽  
Guodong Xu ◽  
Zheng Yang ◽  
Haifeng Liu ◽  
...  

Modern object detectors can rarely achieve short training time, fast inference speed, and high accuracy at the same time. To strike a balance among them, we propose the Training-Time-Friendly Network (TTFNet). In this work, we start with light-head, single-stage, and anchor-free designs, which enable fast inference speed. Then, we focus on shortening training time. We notice that encoding more training samples from annotated boxes plays a similar role as increasing batch size, which helps enlarge the learning rate and accelerate the training process. To this end, we introduce a novel approach using Gaussian kernels to encode training samples. Besides, we design the initiative sample weights for better information utilization. Experiments on MS COCO show that our TTFNet has great advantages in balancing training time, inference speed, and accuracy. It has reduced training time by more than seven times compared to previous real-time detectors while maintaining state-of-the-art performances. In addition, our super-fast version of TTFNet-18 and TTFNet-53 can outperform SSD300 and YOLOv3 by less than one-tenth of their training time, respectively. The code has been made available at https://github.com/ZJULearning/ttfnet.


2019 ◽  
Vol 31 (9) ◽  
pp. 1081-1091
Author(s):  
Megan E. O'Connell ◽  
Holly Tuokko ◽  
Helena Kadlec ◽  
Lauren E. Griffith ◽  
Martine Simard ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document