Show simple item record

dc.rights.licenseCC-BY-NC-ND
dc.contributor.advisorExterne beoordelaar - External assesor,
dc.contributor.authorRietmole, Puck te
dc.date.accessioned2023-06-30T00:00:52Z
dc.date.available2023-06-30T00:00:52Z
dc.date.issued2023
dc.identifier.urihttps://studenttheses.uu.nl/handle/20.500.12932/44061
dc.description.abstractThe problem to be considered is non-preemptive stochastic single machine scheduling to minimize the total weighted completion time of a set of N jobs. This is a classical problem setting that has been studied in the literature on scheduling, with many applications in operations research. In this thesis, we try to understand algorithms and models where the underlying probability distributions are unknown, and processing times can only be assessed via samples. Specifically, the thesis asks what can be done when there is one single sample only. Whenever the expected processing times are known exactly, it is known since the 1960's that one must schedule jobs in decreasing order of weight over expected processing time. For the setting with access to only one sample of the processing time, we define SWOPS, the algorithm that Schedules by Weights Over Processing Time Samples. This algorithm is the most intuitive candidate for a better- than-random algorithm. We first formalize the concept of comparing algorithm performance. This allows us to formulate our main research question; in which scenarios does SWOPS perform at least as well as R, the uniformly random algorithm? We show by counter-example that in general, the performance of SWOPS can be arbitrarily worse than R, motivating that our research question is not trivial. We then show that the matter of comparing relative performance of algorithms can be simplified to a case where the number of jobs N equals 2. We also introduce a novel concept called the Relative Optimality Gap (ROG) of an algorithm. This scales the algorithm's cost to always lie in the interval [0,1], where a ROG of 0 corresponds to an optimal algorithm and a ROG of 1 means the expected performance is the worst possible. It is then proven that finding an upper ROG-bound of at most 1/2 on an algorithm A is equivalent to showing that A performs at least as well as R. We apply these methods to several scenarios, for different types of processing time distributions, and show that SWOPS performs well in these cases. The first scenario is when processing time distributions are symmetric. Second, when all processing time distributions are translated versions of the same underlying distribution. Third, when all processing time distributions are linearly scaled versions of the same under- lying distribution. Lastly, we consider several special sub-cases for exponentially distributed processing times. In the end, we discuss what perspective these findings give us about the performance of SWOPS. We also speculate about the broader implications of our results, in terms of what types of critical failure modes might exist when insufficient samples are available.
dc.description.sponsorshipUtrecht University
dc.language.isoEN
dc.subjectThe thesis explores how the single-machine stochastic scheduling problem changes when we introduce the additional challenge that processing time distributions are only accessible through a single sample.
dc.titleSampling-based Stochastic Single-machine Scheduling
dc.type.contentMaster Thesis
dc.rights.accessrightsOpen Access
dc.subject.keywordsScheduling, job-shop, algorithm, discrete optimization, one-shot learning, learning-augmented, sampling
dc.subject.courseuuMathematical Sciences
dc.thesis.id17887


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record