Partial order planning is an important approach that solves planning problems without completely specifying the orderings between the actions in the plan. This property provides greater flexibility in executing plans; hence making the partial order planners a preferred choice over other planning methodologies. However, in order to find partially ordered plans, partial order planners perform a search in plan space rather than in space of world states and an uninformed search in plan space leads to poor efficiency. In this thesis, I discuss applying a reinforcement learning method, called First-visit Monte Carlo method, to partial order planning in order to design agents which do not need any training data or heuristics but are still able to make informed decisions in plan space based on experience. Communicating effectively with the agent is crucial in reinforcement learning. I address how this task was accomplished in plan space and the results from an evaluation of a blocks world test bed.
Identifer | oai:union.ndltd.org:unt.edu/info:ark/67531/metadc5232 |
Date | 05 1900 |
Creators | Ceylan, Hakan |
Contributors | Swigger, Kathleen M., Brazile, Robert, Mihalcea, Rada, 1974- |
Publisher | University of North Texas |
Source Sets | University of North Texas |
Language | English |
Detected Language | English |
Type | Thesis or Dissertation |
Format | Text |
Rights | Use restricted to UNT Community, Copyright, Ceylan, Hakan, Copyright is held by the author, unless otherwise noted. All rights reserved. |
Page generated in 0.0022 seconds