Full metadata record
DC FieldValueLanguage
dc.contributor.advisorBuchholz, Peter-
dc.contributor.authorScheftelowitsch, Dimitri-
dc.date.accessioned2018-07-03T06:59:16Z-
dc.date.available2018-07-03T06:59:16Z-
dc.date.issued2018-
dc.identifier.urihttp://hdl.handle.net/2003/36946-
dc.identifier.urihttp://dx.doi.org/10.17877/DE290R-18945-
dc.description.abstractMarkov decision processes model stochastic uncertainty in systems and allow one to construct strategies which optimize the behaviour of a system with respect to some reward function. However, the parameters for this uncertainty, that is, the probabilities inside a Markov decision model, are derived from empirical or expert knowledge and are themselves subject to uncertainties such as measurement errors or limited expertise. This work considers second-order uncertainty models for Markov decision processes and derives theoretical and practical results. Among other models, this work considers two main forms of uncertainty. One form is a set of discrete scenarios with a prior probability distribution and the task to maximize the expected reward under the given probability distribution. Another form of uncertainty is a continuous uncertainty set of scenarios and the task to compute a policy that optimizes the rewards in the optimistic and pessimistic cases. The work provides two kinds of results. First, we establish complexity-theoretic hardness results for the considered optimization problems. Second, we design heuristics for some of the problems and evaluate them empirically. In the first class of results, we show that additional model uncertainty makes the optimization problems harder to solve, as they add an additional party with own optimization goals. In the second class of results, we show that even if the discussed problems are hard to solve in theory, we can come up with efficient heuristics that can solve them adequately well for practical applications.en
dc.language.isoende
dc.subjectOptimierungde
dc.subjectStochastische Prozessede
dc.subjectRobuste Optimierungde
dc.subjectUnsicherheitde
dc.subject.ddc004-
dc.titleMarkov decision processes with uncertain parametersen
dc.typeTextde
dc.contributor.refereeHermanns, Holger-
dc.date.accepted2018-05-03-
dc.type.publicationtypedoctoralThesisde
dc.subject.rswkOptimierungde
dc.subject.rswkStochastischer Prozessde
dc.subject.rswkRobuste Optimierungde
dc.subject.rswkUnsicheres Schließende
dcterms.accessRightsopen access-
eldorado.secondarypublicationfalsede
Appears in Collections:LS 04 Praktische Informatik

Files in This Item:
File Description SizeFormat 
Scheftelowitsch_Dissertation.pdfDNB1.1 MBAdobe PDFView/Open


This item is protected by original copyright



This item is protected by original copyright rightsstatements.org