A note on bias optimality in controlled queueing systems

2000 ◽  
Vol 37 (1) ◽  
pp. 300-305 ◽  
Author(s):  
Mark E. Lewis ◽  
Martin L. Puterman

The use of bias optimality to distinguish among gain optimal policies was recently studied by Haviv and Puterman [1] and extended in Lewis et al. [2]. In [1], upon arrival to an M/M/1 queue, customers offer the gatekeeper a reward R. If accepted, the gatekeeper immediately receives the reward, but is charged a holding cost, c(s), depending on the number of customers in the system. The gatekeeper, whose objective is to ‘maximize’ rewards, must decide whether to admit the customer. If the customer is accepted, the customer joins the queue and awaits service. Haviv and Puterman [1] showed there can be only two Markovian, stationary, deterministic gain optimal policies and that only the policy which uses the larger control limit is bias optimal. This showed the usefulness of bias optimality to distinguish between gain optimal policies. In the same paper, they conjectured that if the gatekeeper receives the reward upon completion of a job instead of upon entry, the bias optimal policy will be the lower control limit. This note confirms that conjecture.

2000 ◽  
Vol 37 (01) ◽  
pp. 300-305 ◽  
Author(s):  
Mark E. Lewis ◽  
Martin L. Puterman

The use ofbias optimalityto distinguish among gain optimal policies was recently studied by Haviv and Puterman [1] and extended in Lewiset al.[2]. In [1], upon arrival to anM/M/1 queue, customers offer the gatekeeper a rewardR. If accepted, the gatekeeper immediately receives the reward, but is charged a holding cost,c(s), depending on the number of customers in the system. The gatekeeper, whose objective is to ‘maximize’ rewards, must decide whether to admit the customer. If the customer is accepted, the customer joins the queue and awaits service. Haviv and Puterman [1] showed there can be only two Markovian, stationary, deterministic gain optimal policies and that only the policy which uses thelargercontrol limit is bias optimal. This showed the usefulness of bias optimality to distinguish between gain optimal policies. In the same paper, they conjectured that if the gatekeeper receives the reward uponcompletionof a job instead of upon entry, the bias optimal policy will be the lower control limit. This note confirms that conjecture.


1998 ◽  
Vol 35 (1) ◽  
pp. 136-150 ◽  
Author(s):  
Moshe Haviv ◽  
Martin L. Puterman

This paper studies an admission control M/M/1 queueing system. It shows that the only gain (average) optimal stationary policies with gain and bias which satisfy the optimality equation are of control limit type, that there are at most two and, if there are two, they occur consecutively. Conditions are provided which ensure the existence of two gain optimal control limit policies and are illustrated with an example. The main result is that bias optimality distinguishes these two gain optimal policies and that the larger of the two control limits is the unique bias optimal stationary policy. Consequently it is also Blackwell optimal. This result is established by appealing to the third optimality equation of the Markov decision process and some observations concerning the structure of solutions of the second optimality equation.


1998 ◽  
Vol 35 (01) ◽  
pp. 136-150 ◽  
Author(s):  
Moshe Haviv ◽  
Martin L. Puterman

This paper studies an admission control M/M/1 queueing system. It shows that the only gain (average) optimal stationary policies with gain and bias which satisfy the optimality equation are of control limit type, that there are at most two and, if there are two, they occur consecutively. Conditions are provided which ensure the existence of two gain optimal control limit policies and are illustrated with an example. The main result is that bias optimality distinguishes these two gain optimal policies and that the larger of the two control limits is the unique bias optimal stationary policy. Consequently it is also Blackwell optimal. This result is established by appealing to the third optimality equation of the Markov decision process and some observations concerning the structure of solutions of the second optimality equation.


1999 ◽  
Vol 13 (1) ◽  
pp. 55-70 ◽  
Author(s):  
Zvi Benyamini ◽  
Uri Yechiali

Control limit type policies are widely discussed in the literature, particularly regarding the maintenance of deteriorating systems. Previous studies deal mainly with stationary deterioration processes, where costs and transition probabilities depend only on the state of the system, regardless of its cumulative age. In this paper, we consider a nonstationary deterioration process, in which operation and maintenance costs, as well as transition probabilities “deteriorate” with both the system's state and its cumulative age. We discuss conditions under which control limit policies are optimal for such processes and compare them with those used in the analysis of stationary models.Two maintenance models are examined: in the first (as in the majority of classic studies), the only maintenance action allowed is the replacement of the system by a new one. In this case, we show that the nonstationary results are direct generalizations of their counterparts in stationary models. We propose an efficient algorithm for finding the optimal policy, utilizing its control limit form. In the second model we also allow for repairs to better states (without changing the age). In this case, the optimal policy is shown to have the form of a 3-way control limit rule. However, conditions analogous to those used in the stationary problem do not suffice, so additional, more restrictive ones are suggested and discussed.


1967 ◽  
Vol 4 (01) ◽  
pp. 162-179 ◽  
Author(s):  
J. W. Cohen

The distribution of the maximum number of customers simultaneously present during a busy period is studied for the queueing systems M/G/1 and G/M/1. These distributions are obtained by using taboo probabilities. Some new relations for transition probabilities and entrance time distributions are derived.


1997 ◽  
Vol 29 (01) ◽  
pp. 114-137
Author(s):  
Linn I. Sennott

This paper studies the expected average cost control problem for discrete-time Markov decision processes with denumerably infinite state spaces. A sequence of finite state space truncations is defined such that the average costs and average optimal policies in the sequence converge to the optimal average cost and an optimal policy in the original process. The theory is illustrated with several examples from the control of discrete-time queueing systems. Numerical results are discussed.


2003 ◽  
Vol 16 (4) ◽  
pp. 311-326 ◽  
Author(s):  
Mykola Bratiychuk ◽  
Andrzej Chydzinski

This paper examines a new class of queueing systems and proves a theorem on the existence of the ergodic distribution of the number of customers in such a system. An ergodic distribution is computed explicitly for the special case of a G/M−M/1 system, where the interarrival distribution does not change and both service distributions are exponential. A numerical example is also given.


1990 ◽  
Vol 22 (03) ◽  
pp. 764-767 ◽  
Author(s):  
Ludolf E. Meester ◽  
J. George Shanthikumar

We consider a tandem queueing system with m stages and finite intermediate buffer storage spaces. Each stage has a single server and the service times are independent and exponentially distributed. There is an unlimited supply of customers in front of the first stage. For this system we show that the number of customers departing from each of the m stages during the time interval [0, t] for any t ≧ 0 is strongly stochastically increasing and concave in the buffer storage capacities. Consequently the throughput of this tandem queueing system is an increasing and concave function of the buffer storage capacities. We establish this result using a sample path recursion for the departure processes from the m stages of the tandem queueing system, that may be of independent interest. The concavity of the throughput is used along with the reversibility property of tandem queues to obtain the optimal buffer space allocation that maximizes the throughput for a three-stage tandem queue.


2002 ◽  
Vol 39 (01) ◽  
pp. 20-37 ◽  
Author(s):  
Mark E. Lewis ◽  
Hayriye Ayhan ◽  
Robert D. Foley

We consider a finite-capacity queueing system where arriving customers offer rewards which are paid upon acceptance into the system. The gatekeeper, whose objective is to ‘maximize’ rewards, decides if the reward offered is sufficient to accept or reject the arriving customer. Suppose the arrival rates, service rates, and system capacity are changing over time in a known manner. We show that all bias optimal (a refinement of long-run average reward optimal) policies are of threshold form. Furthermore, we give sufficient conditions for the bias optimal policy to be monotonic in time. We show, via a counterexample, that if these conditions are violated, the optimal policy may not be monotonic in time or of threshold form.


2018 ◽  
Vol 28 (1) ◽  
pp. 141-154 ◽  
Author(s):  
Alexander Zeifman ◽  
Rostislav Razumchik ◽  
Yacov Satin ◽  
Ksenia Kiseleva ◽  
Anna Korotysheva ◽  
...  

AbstractIn this paper we present a method for the computation of convergence bounds for four classes of multiserver queueing systems, described by inhomogeneous Markov chains. Specifically, we consider an inhomogeneous M/M/S queueing system with possible state-dependent arrival and service intensities, and additionally possible batch arrivals and batch service. A unified approach based on a logarithmic norm of linear operators for obtaining sharp upper and lower bounds on the rate of convergence and corresponding sharp perturbation bounds is described. As a side effect, we show, by virtue of numerical examples, that the approach based on a logarithmic norm can also be used to approximate limiting characteristics (the idle probability and the mean number of customers in the system) of the systems considered with a given approximation error.


Sign in / Sign up

Export Citation Format

Share Document