Author's School

School of Engineering & Applied Science

Author's Department/Program

Computer Science and Engineering

Language

English (en)

Date of Award

January 2009

Degree Type

Dissertation

Degree Name

Doctor of Philosophy (PhD)

Chair and Committee

William Smart

Abstract

Scheduling policies for open soft real-time systems must be able to balance the competing concerns of meeting their objectives under exceptional conditions while achieving good performance in the average case. Balancing these concerns requires modeling strategies that represent the range of possible task behaviors, and solution techniques that are capable of effectively managing uncertainty in order to discover scheduling policies that are effective across the range of system modes. We develop methods for solving a particular class of task scheduling problems in an open soft real-time setting involving repeating, non-preemptable tasks that contend for a single shared resource. We enforce timeliness by optimizing performance with respect to the proportional progress of tasks in the system. We model this scheduling problem as an infinite-state Markov decision process, and provide guarantees regarding the existence of optimal solutions to this problem. We derive several methods for approximating optimal scheduling policies and provide theoretical justification and empirical evidence that these solutions are good approximations to the optimal solution. We consider cases in which task models are known, and adapt reinforcement learning methods to learn task models when they are not available.

DOI

https://doi.org/10.7936/K7KD1VZX

Comments

Permanent URL: http://dx.doi.org/10.7936/K7KD1VZX

Share

COinS