Temporal Planning
- Automated Planning and Acting; Chapter “Deliberation with Temporal Models” Section 4.1 (Introduction) and Section 4.2 (Temporal Representation.
Ghallab, M., Nau, D., Traverso, P., (2016).
http://projects.laas.fr/planning/
Lab: Plan Synthesis
- The PDDL wiki, which includes a reference and guide:
https://planning.wiki - An Introduction to the Planning Domain Definition Language.
Haslum, P., Lipovetzky, N., Magazzeni, D., Muise, C., (2019). - PDDL – The Planning Domain Definition Language.
Ghallab, M., Knoblock, C., Wilkins, D., Barrett, A., Christianson, D., Friedman, M., Kwok, C., Golden, K., Penberthy, S., Smith, D., Sun, Y., & Weld, D. (1998). - Fox, M., & Long, D. (2003). PDDL2.1: An Extension to PDDL for Expressing Temporal Planning Domains. CoRR, abs/1106.4561.
http://arxiv.org/abs/1106.4561
Planning and Plan Execution for Human-Robot Interaction tasks
- Planning for Social Interaction in a Robot Bartender Domain.
Petrick, R. P. A., Foster, M. E., In Proc. of ICAPS 2013, pages 389-397.
http://www.dcs.gla.ac.uk/~mefoster/papers/petrick-foster-icaps2013.pdf - Planning with sensing, concurrency, and exogenous events: logical framework and implementation.
Iocchi, L., Nardi, D., Rosati, R., In Proc. of the 7th Int. Conf. on Principles of Knowledge Representation and Reasoning (KR’2000)
https://www.diag.uniroma1.it/iocchi/publications/iocchi-kr00.pdf - A knowledge-based approach to planning with incomplete information and sensing.
Petrick, R.P.A., Bacchus, F., In Proceedings of AIPS, pp. 212–221 (2002)
https://www.aaai.org/Library/AIPS/2002/aips02-022.php - Petri Net Plans – A framework for collaboration and coordination in multi-robot systems.
Ziparo, V., Iocchi, L., Lima, P., Nardi, D., Palamara, P., Autonomous Agents and Multi-Agent Systems, 23(3): 344-383 (2011)
http://www.diag.uniroma1.it/iocchi/publications/iocchi-jaamas11-draft.pdf - Plan Explanations as Model Reconciliation: Moving Beyond Explanation as Soliloquy.
Chakraborti, T., Sreedharan, S., Zhang, Y., Kambhampati, S., Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence (IJCAI) 2017.
https://www.ijcai.org/Proceedings/2017/0023.pdf - Short-Term Human Robot Interaction through Conditional Planning and Execution.
Sanelli, V., Cashmore, M., Magazzeni, D., Iocchi, L., In Proc. of International Conference on Automated Planning and Scheduling (ICAPS), 2017.
https://www.diag.uniroma1.it/iocchi/publications/sanelli-icaps2017.pdf
Integrated Planning and Acting using Operational Models
- Integrating acting, planning, and learning in hierarchical operational models.
Patra, S., Mason, J., Kumar, A., Ghallab, M., Traverso, P., Nau D., In ICAPS-2020. Best student paper honourable mention award.
Integrated Planning and Acting using Operational Models
Relevant Introduction
– https://www.morganclaypool.com/doi/abs/10.2200/S00426ED1V01Y201206AIM017
Sutton and Barto, RL Book
– edition 1 (can be read quickly): http://incompleteideas.net/book/first/the-book.html
– edition 2 (more modern update): http://incompleteideas.net/book/the-book-2nd.html
Monte Carlo Tree Search
– Thomas Keller had some great insights in his THTS paper:
http://www2.informatik.uni-freiburg.de/~ki/papers/keller-helmert-icaps2013.pdf
– His PROST planner cited in the above paper that has won many ICAPS probabilistic planning competitions and more recent versions of his MCTS work can be found on his web page:
https://ai.dmi.unibas.ch/people/tkeller/index.html
RTDP papers
– Real-time Dynamic Programming (discussed above in Thomas Keller’s THTS work) is a beautiful way to combine the benefits of dynamic programming and tree search:
https://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.43.6020&rep=rep1&type=pdf
– Bounded Real-time Dynamic Programming is a simple but beautiful RTDP algorithm that exploits lower and upper bound properties of dynamic programming:
http://www.cs.cmu.edu/~ggordon/mcmahan-likhachev-gordon.brtdp.pdf
AIXI
– I mentioned today that Marcus Hutter has worked on optimal decision making when you only have observations and you cannot easily define a state. He calls this model AIXI and he has a website and book here: http://www.hutter1.net/ai/
Latent State Models
– I mentioned a very interesting trend of recent work on learning and planning with latent models of state (rather than the explicit models we’re used to dealing with in ICAPS), here are some pointers:
Predictron (DeepMind): https://arxiv.org/abs/1612.08810
MuZero (DeepMind): https://deepmind.com/research/publications/Mastering-Atari-Go-Chess-and-Shogi-by-Planning-with-a-Learned-Model
Masataro Asai’s Research Publications: https://scholar.google.co.jp/citations?user=b4UzH5AAAAAJ&hl=en
Planning in Continuous State MDPs
– There is excellent work for planning with continuous states and actions in the deterministic setting, however, there is not nearly as much work as I would like to see in the stochastic setting (though perhaps I have missed recent work). Let me give you two groups of pointers based on the work of my own group:
… Symbolic Dynamic Programming (exact dynamic programming solutions to continuous state and action MDPs):
Continuous State MDPs: https://ssanner.github.io/papers/anu/cont_mdp.pdf
Continuous State and Action MDPs: https://ssanner.github.io/papers/anu/aaai12_sdp.pdf
(see also Related Work in these papers and many more papers with “Symbolic Dynamic Programming” in the title on my group’s web page: https://d3m.mie.utoronto.ca/publications/)
… Deriving Plans and Policies in Continuous State and Action MDPs specified in RDDL:
Policy Learning: https://ssanner.github.io/papers/aaai19_drp.pdf
(the above paper includes a Python version of a RDDL parser)
Planning in Deterministic Continuous RDDL Models: https://ssanner.github.io/papers/nips17_tfplan.pdf
RDDL – Relational Dynamic Influence Diagram Language
– A compact language for specifying MDPs used in the past decade of ICAPS probabilistic planning competitions. Overview, Slides, Exercises, the Source Code. etc:
https://github.com/ssanner/rddlsim