References

  1. Yossiri Adulyasak, Pradeep Varakantham, Asrar Ahmed & Patrick Jaillet (2015): Solving uncertain MDPs with objectives that are separable over instantiations of model uncertainty. In: AAAI Conference on Artificial Intelligence, pp. 3454–3460.
  2. Asrar Ahmed, Pradeep Varakantham, Yossiri Adulyasak & Patrick Jaillet (2013): Regret based robust solutions for uncertain Markov decision processes. In: Advances in Neural Information Processing Systems (NeurIPS), pp. 881–889, doi:10.1613/jai.5242.
  3. Asrar Ahmed, Pradeep Varakantham, Meghna Lowalekar, Yossiri Adulyasak & Patrick Jaillet (2017): Sampling based approaches for minimizing regret in uncertain Markov decision processes (MDPs). Journal of Artificial Intelligence Research (JAIR) 59, pp. 229–264, doi:10.1613/jair.5242.
  4. Sule Anjomshoae, Amro Najjar, Davide Calvaresi & Kary Främling (2019): Explainable agents and robots: Results from a systematic literature review. In: International Conference on Autonomous Agents and MultiAgent Systems (AAMAS), pp. 1078–1088.
  5. J. Andrew Bagnell, Andrew Y. Ng & Jeff G. Schneider (2001): Solving uncertain Markov decision processes. Technical Report. Carnegie Mellon University.
  6. Connor Basich, Justin Svegliato, Kyle Hollins Wray, Stefan Witwicki, Joydeep Biswas & Shlomo Zilberstein (2020): Learning to Optimize Autonomy in Competence-Aware Systems. In: International Joint Conference on Autonomous Agents and Multiagent Systems (AAMAS), pp. 123–131.
  7. Felix Berkenkamp, Matteo Turchetta, Angela Schoellig & Andreas Krause (2017): Safe model-based reinforcement learning with stability guarantees. In: Advances in Neural Information Processing Systems (NeurIPS), pp. 908–918.
  8. Alberto Broggi, Massimo Bertozzi, Alessandra Fascioli, C. Guarino Lo Bianco & Aurelio Piazzi (1999): The ARGO autonomous vehicle’s vision and control systems. International Journal of Intelligent Control and Systems 3(4), pp. 409–441.
  9. Alberto Broggi, Pietro Cerri, Mirko Felisa, Maria Chiara Laghi, Luca Mazzei & Pier Paolo Porta (2012): The VisLab Intercontinental Autonomous Challenge: An extensive test for a platoon of intelligent vehicles. International Journal of Vehicle Autonomous Systems 10(3), pp. 147–164, doi:10.1504/IJVAS.2012.051250.
  10. Katherine Chen & Michael Bowling (2012): Tractable objectives for robust policy optimization. In: Advances in Neural Information Processing Systems (NeurIPS), pp. 2069–2077.
  11. Sonia Chernova & Manuela Veloso (2009): Interactive policy learning through confidence-based autonomy. Journal of Artificial Intelligence Research (JAIR) 34, pp. 1–25, doi:10.1613/jair.2584.
  12. Jeffery A. Clouse (1996): On integrating apprentice learning and reinforcement learning. University of Massachusetts Amherst.
  13. Manoranjan Dash & Huan Liu (1997): Feature selection for classification. Intelligent Data Analysis 1(3), pp. 131–156, doi:10.1016/S1088-467X(97)00008-5.
  14. Francesco Del Duchetto, Ayse Kucukyilmaz, Luca Iocchi & Marc Hanheide (2018): Do not make the same mistakes again and again: Learning local recovery policies for navigation from human demonstrations. IEEE Robotics and Automation Letters (RA-L) 3(4), pp. 4084–4091, doi:10.1109/LRA.2018.2861080.
  15. Karina V. Delgado, Leliane N. De Barros, Daniel B. Dias & Scott Sanner (2016): Real-time dynamic programming for Markov decision processes with imprecise probabilities. Artificial Intelligence (AIJ) 230, pp. 192–223, doi:10.1016/j.artint.2015.09.005.
  16. Karina Valdivia Delgado, Scott Sanner & Leliane Nunes De Barros (2011): Efficient solutions to factored MDPs with imprecise transition probabilities. Artificial Intelligence (AIJ) 175(9-10), pp. 1498–1527, doi:10.1016/j.artint.2011.01.001.
  17. Ernst D. Dickmanns (2007): Dynamic vision for perception and control of motion. Springer Science & Business Media, doi:10.1007/978-1-84628-638-4.
  18. Carlos Diuk, Lihong Li & Bethany R. Leffler (2009): The adaptive k-meteorologists problem and its application to structure learning and feature selection in reinforcement learning. In: International Conference on Machine Learning (ICML), pp. 249–256, doi:10.1145/1553374.1553406.
  19. Liang Du & Yi-Dong Shen (2015): Unsupervised feature selection with adaptive structure learning. In: ACM International Conference on Knowledge Discovery and Data Mining (SIGKDD), pp. 209–218, doi:10.1145/2783258.2783345.
  20. Yang Gao & Steve Chien (2017): Review on space robotics: Toward top-level science through space exploration. Science Robotics 2(7), doi:10.1126/scirobotics.aan5074.
  21. Nick Hawes, Christopher Burbridge, Ferdian Jovan, Lars Kunze, Bruno Lacerda, Lenka Mudrova, Jay Young, Jeremy Wyatt, Denise Hebesberger & Tobias Kortner (2017): The STRANDS project: Long-term autonomy in everyday environments. IEEE Robotics & Automation Magazine 24(3), doi:10.1109/MRA.2016.2636359.
  22. Kevin Anthony Hoff & Masooda Bashir (2015): Trust in automation: Integrating empirical evidence on factors that influence trust. Human Factors 57(3), pp. 407–434, doi:10.1177/0018720814547570.
  23. Yin Lou, Rich Caruana, Johannes Gehrke & Giles Hooker (2013): Accurate intelligible models with pairwise interactions. In: ACM International Conference on Knowledge Discovery and Data Mining (SIGKDD), pp. 623–631, doi:10.1145/2487575.2487579.
  24. Wim Meeussen, Eitan Marder-Eppstein, Kevin Watts & Brian P. Gerkey (2011): Long term autonomy in office environments. In: Robotics: Science and Systems (RSS) ALONE Workshop.
  25. John F. Mustard, D. Beaty & D. Bass (2013): Mars 2020 science rover: Science goals and mission concept. In: AAS/Division for Planetary Sciences Meeting Abstracts 45.
  26. Feiping Nie, Wei Zhu & Xuelong Li (2016): Unsupervised feature selection with structured graph optimization. In: AAAI Conference on Artificial Intelligence, pp. 1302–1308.
  27. Luis Pineda, Takeshi Takahashi, Hee-Tae Jung, Shlomo Zilberstein & Roderic Grupen (2015): Continual planning for search and rescue robots. In: IEEE/RAS International Conference on Humanoid Robots (Humanoids). IEEE, pp. 243–248, doi:10.1109/HUMANOIDS.2015.7363542.
  28. Kevin Regan & Craig Boutilier (2010): Robust policy computation in reward-uncertain MDPs using nondominated policies. In: AAAI Conference on Artificial Intelligence, pp. 1127–1133.
  29. Kevin Regan & Craig Boutilier (2011): Robust online optimization of reward-uncertain MDPs. In: International Joint Conference on Artificial Intelligence (IJCAI), pp. 2165–2171, doi:10.5591/978-1-57735-516-8/IJCAI11-361.
  30. Marc Rigter, Bruno Lacerda & Nick Hawes (2020): A Framework for learning From demonstration with minimal human effort. IEEE Robotics and Automation Letters (RA-L) 5(2), pp. 2023–2030, doi:10.1109/LRA.2018.2861080.
  31. Michael T. Rosenstein & Andrew G. Barto (2004): Supervised actor-critic reinforcement learning. In: J. Si, A. G. Barto, W. B. Powell & D. Wunsch: Handbook of Learning and Approximate Dynamic Programming, chapter 7. IEEE Press, pp. 359–380.
  32. Sandhya Saisubramanian, Shlomo Zilberstein & Prashant Shenoy (2017): Optimizing electric vehicle charging through determinization. In: ICAPS Workshop on Scheduling and Planning Applications.
  33. Jay K. Satia & Roy E. Lave Jr. (1973): Markovian decision processes with uncertain transition probabilities. Operations Research 21(3), pp. 728–740, doi:10.1287/opre.21.3.728.
  34. Ricardo Shirota Filho, Fabio Gagliardi Cozman, Felipe W. Trevizan, Cassio Polpo de Campos & Leliane Nunes De Barros (2007): Multilinear and integer programming for Markov decision processes with imprecise probabilities. In: 5th International Symposium on Imprecise Probability: Theories and Applications, pp. 395–404.
  35. William D. Smart & Leslie Pack Kaelbling (2002): Effective reinforcement learning for mobile robots. In: IEEE International Conference on Robotics and Automation (ICRA) 4, pp. 3404–3410, doi:10.1109/ROBOT.2002.1014237.
  36. Halit Bener Suay & Sonia Chernova (2011): Effect of human guidance and state space size on interactive reinforcement learning. In: IEEE Conference on Robot and Human Interactive Communication (RO-MAN). IEEE, pp. 1–6, doi:10.1109/ROMAN.2011.6005223.
  37. Justin Svegliato, Kyle Hollins Wray, Stefan J. Witwicki, Joydeep Biswas & Shlomo Zilberstein (2019): Belief Space Metareasoning for Exception Recovery. In: IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 1224–1229, doi:10.1109/IROS40897.2019.8967676.
  38. Lisa Torrey & Matthew Taylor (2013): Teaching on a budget: Agents advising agents in reinforcement learning. In: International Conference on Autonomous Agents and Multi-Agent Systems (AAMAS), pp. 1053–1060.
  39. Felipe W. Trevizan, Fabio Gagliardi Cozman & Leliane Nunes de Barros (2007): Planning under risk and knightian uncertainty. In: International Joint Conference on Artificial Intelligence (IJCAI) 2007, pp. 2023–2028.
  40. Thomas J. Walsh, Sergiu Goschin & Michael L. Littman (2010): Integrating sample-based planning and model-based reinforcement learning. In: AAAI Conference on Artificial Intelligence, pp. 612–617.
  41. Chelsea C. White III & Hany K. El-Deib (1986): Parameter imprecision in finite state, finite action dynamic programs. Operations Research 34(1), pp. 120–129, doi:10.1287/opre.34.1.120.
  42. Chelsea C. White III & Hany K. Eldeib (1994): Markov decision processes with imprecise transition probabilities. Operations Research 42(4), pp. 739–749, doi:10.1287/opre.42.4.739.
  43. Grady Williams, Nolan Wagener, Brian Goldfain, Paul Drews, James M. Rehg, Byron Boots & Evangelos A. Theodorou (2017): Information theoretic MPC for model-based reinforcement learning. In: IEEE International Conference on Robotics and Automation (ICRA). IEEE, pp. 1714–1721, doi:10.1109/ICRA.2017.7989202.
  44. Eric M. Wolff, Ufuk Topcu & Richard M. Murray (2012): Robust control of uncertain Markov decision processes with temporal logic specifications. In: Conference on Decision and Control (CDC). IEEE, pp. 3372–3379, doi:10.1109/CDC.2012.6426174.
  45. Kyle Hollins Wray, Luis Enrique Pineda & Shlomo Zilberstein (2016): Hierarchical Approach to Transfer of Control in Semi-Autonomous Systems. In: International Joint Conference on Artificial Intelligence (IJCAI), pp. 517–523.
  46. Huan Xu & Shie Mannor (2009): Parametric regret in uncertain Markov decision processes. In: Conference on Decision and Control (CDC). IEEE, pp. 3606–3613, doi:10.1109/CDC.2009.5400796.

Comments and questions to: eptcs@eptcs.org
For website issues: webmaster@eptcs.org