Loading…
A scheduling problem for several parallel servers
In this paper, we study a scheduling problem of jobs from two different queues on several parallel servers. Jobs have exponentially distributed processing times, and incur costs per unit of time, until they leave the system, and there are no arrivals to the system at any time. The objective is to fi...
Saved in:
Published in: | Mathematical methods of operations research (Heidelberg, Germany) Germany), 2007-08, Vol.66 (1), p.127-148 |
---|---|
Main Authors: | , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | In this paper, we study a scheduling problem of jobs from two different queues on several parallel servers. Jobs have exponentially distributed processing times, and incur costs per unit of time, until they leave the system, and there are no arrivals to the system at any time. The objective is to find the optimal strategy, i.e., to allocate the servers to the queues, such that the expected holding costs are minimized. We give a sufficient condition for which it is always optimal to allocate the servers only to jobs of a certain queue. Finally, the case of two servers is completely solved. [PUBLICATION ABSTRACT] |
---|---|
ISSN: | 1432-2994 1432-5217 |
DOI: | 10.1007/s00186-007-0150-4 |