scholarly journals Deriving the tail distribution of the buffer contents in a statistical multiplexer with general heterogeneous on/off sources

Author(s):  
S. Wittevrongel ◽  
H. Bruneel
2003 ◽  
Vol 40 (02) ◽  
pp. 273-292
Author(s):  
J. Xue ◽  
Attahiru Sule Alfa

This paper considers the asymptotic tail distribution of the number of cells queued in a statistical multiplexer fed with homogeneous generalized binary Markov sources. As the asymptotic decay rate is easy to obtain, we focus our effort on bounding the asymptotic constant, which is dependent on the initial phase combination of the sources and is hard to compute even for a moderate number of sources. We derive upper and lower bounds for the asymptotic constant, taking the initial phase combination into account. Numerical experiments show the accuracy of these bounds. They also show that, while the asymptotic decay rates are the same, the variation of initial phase combination of the sources may significantly affect the asymptotic constants.


2003 ◽  
Vol 40 (02) ◽  
pp. 273-292
Author(s):  
J. Xue ◽  
Attahiru Sule Alfa

This paper considers the asymptotic tail distribution of the number of cells queued in a statistical multiplexer fed with homogeneous generalized binary Markov sources. As the asymptotic decay rate is easy to obtain, we focus our effort on bounding the asymptotic constant, which is dependent on the initial phase combination of the sources and is hard to compute even for a moderate number of sources. We derive upper and lower bounds for the asymptotic constant, taking the initial phase combination into account. Numerical experiments show the accuracy of these bounds. They also show that, while the asymptotic decay rates are the same, the variation of initial phase combination of the sources may significantly affect the asymptotic constants.


Symmetry ◽  
2021 ◽  
Vol 13 (4) ◽  
pp. 679
Author(s):  
Jimmy Reyes ◽  
Emilio Gómez-Déniz ◽  
Héctor W. Gómez ◽  
Enrique Calderín-Ojeda

There are some generalizations of the classical exponential distribution in the statistical literature that have proven to be helpful in numerous scenarios. Some of these distributions are the families of distributions that were proposed by Marshall and Olkin and Gupta. The disadvantage of these models is the impossibility of fitting data of a bimodal nature of incorporating covariates in the model in a simple way. Some empirical datasets with positive support, such as losses in insurance portfolios, show an excess of zero values and bimodality. For these cases, classical distributions, such as exponential, gamma, Weibull, or inverse Gaussian, to name a few, are unable to explain data of this nature. This paper attempts to fill this gap in the literature by introducing a family of distributions that can be unimodal or bimodal and nests the exponential distribution. Some of its more relevant properties, including moments, kurtosis, Fisher’s asymmetric coefficient, and several estimation methods, are illustrated. Different results that are related to finance and insurance, such as hazard rate function, limited expected value, and the integrated tail distribution, among other measures, are derived. Because of the simplicity of the mean of this distribution, a regression model is also derived. Finally, examples that are based on actuarial data are used to compare this new family with the exponential distribution.


Author(s):  
Tong Wei ◽  
Yu-Feng Li

Large-scale multi-label learning annotates relevant labels for unseen data from a huge number of candidate labels. It is well known that in large-scale multi-label learning, labels exhibit a long tail distribution in which a significant fraction of labels are tail labels. Nonetheless, how tail labels make impact on the performance metrics in large-scale multi-label learning was not explicitly quantified. In this paper, we disclose that whatever labels are randomly missing or misclassified, tail labels impact much less than common labels in terms of commonly used performance metrics (Top-$k$ precision and nDCG@$k$). With the observation above, we develop a low-complexity large-scale multi-label learning algorithm with the goal of facilitating fast prediction and compact models by trimming tail labels adaptively. Experiments clearly verify that both the prediction time and the model size are significantly reduced without sacrificing much predictive performance for state-of-the-art approaches.


Sign in / Sign up

Export Citation Format

Share Document