Elicitation and Planning in Markov Decision Processes with Unknown Rewards

Elicitation and Planning in Markov Decision Processes with Unknown Rewards
Author :
Publisher :
Total Pages : 0
Release :
ISBN-10 : OCLC:1022562936
ISBN-13 :
Rating : 4/5 ( Downloads)

Book Synopsis Elicitation and Planning in Markov Decision Processes with Unknown Rewards by : Pegah Alizadeh

Download or read book Elicitation and Planning in Markov Decision Processes with Unknown Rewards written by Pegah Alizadeh and published by . This book was released on 2016 with total page 0 pages. Available in PDF, EPUB and Kindle. Book excerpt: Markov decision processes (MDPs) are models for solving sequential decision problemswhere a user interacts with the environment and adapts her policy by taking numericalreward signals into account. The solution of an MDP reduces to formulate the userbehavior in the environment with a policy function that specifies which action to choose ineach situation. In many real world decision problems, the users have various preferences,and therefore, the gain of actions on states are different and should be re-decoded foreach user. In this dissertation, we are interested in solving MDPs for users with differentpreferences.We use a model named Vector-valued MDP (VMDP) with vector rewards. We propose apropagation-search algorithm that allows to assign a vector-value function to each policyand identify each user with a preference vector on the existing set of preferences wherethe preference vector satisfies the user priorities. Since the user preference vector is notknown we present several methods for solving VMDPs while approximating the user'spreference vector.We introduce two algorithms that reduce the number of queries needed to find the optimalpolicy of a user: 1) A propagation-search algorithm, where we propagate a setof possible optimal policies for the given MDP without knowing the user's preferences.2) An interactive value iteration algorithm (IVI) on VMDPs, namely Advantage-basedValue Iteration (ABVI) algorithm that uses clustering and regrouping advantages. Wealso demonstrate how ABVI algorithm works properly for two different types of users:confident and uncertain.We finally work on a minimax regret approximation method as a method for findingthe optimal policy w.r.t the limited information about user's preferences. All possibleobjectives in the system are just bounded between two higher and lower bounds while thesystem is not aware of user's preferences among them. We propose an heuristic minimaxregret approximation method for solving MDPs with unknown rewards that is faster andless complex than the existing methods in the literature.

Elicitation and Planning in Markov Decision Processes with Unknown Rewards Related Books

Elicitation and Planning in Markov Decision Processes with Unknown Rewards
Language: en
Pages: 0
Authors: Pegah Alizadeh
Categories:
Type: BOOK - Published: 2016 - Publisher:

GET EBOOK

Markov decision processes (MDPs) are models for solving sequential decision problemswhere a user interacts with the environment and adapts her policy by taking
Regret-based Reward Elicitation for Markov Decision Processes
Language: en
Pages:
Authors: Regan Kevin
Categories:
Type: BOOK - Published: 2014 - Publisher:

GET EBOOK

Cognitive Electronic Warfare: An Artificial Intelligence Approach
Language: en
Pages: 288
Authors: Karen Haigh
Categories: Technology & Engineering
Type: BOOK - Published: 2021-07-31 - Publisher: Artech House

GET EBOOK

This comprehensive book gives an overview of how cognitive systems and artificial intelligence (AI) can be used in electronic warfare (EW). Readers will learn h
Algorithmic Decision Theory
Language: en
Pages: 593
Authors: Toby Walsh
Categories: Computers
Type: BOOK - Published: 2015-08-27 - Publisher: Springer

GET EBOOK

This book constitutes the thoroughly refereed conference proceedings of the 4th International Conference on Algorithmic Decision Theory , ADT 2015, held in Sept
Planning with Markov Decision Processes
Language: en
Pages: 212
Authors: Mausam
Categories: Computers
Type: BOOK - Published: 2012-06-01 - Publisher: Morgan & Claypool Publishers

GET EBOOK

Markov Decision Processes (MDPs) are widely popular in Artificial Intelligence for modeling sequential decision-making scenarios with probabilistic dynamics. Th