Adaptive submodular inverse reinforcement learning for spatial search and map exploration

Ji Jie Wu, Kuo Shih Tseng

Research output: Contribution to journalArticlepeer-review

2 Scopus citations

Abstract

Finding optimal paths for spatial search and map exploration problems are NP-hard. Since spatial search and environmental exploration are parts of human central activities, learning human behavior from data is a way to solve these problems. Utilizing the adaptive submodularity of two problems, this research proposes an adaptive submodular inverse reinforcement learning (ASIRL) algorithm to learn human behavior. The ASIRL approach is to learn the reward functions in the Fourier domain and then recover it in the spatial domain. The near-optimal path can be computed through learned reward functions. The experiments demonstrate that the ASIRL outperforms state of the art approaches (e.g., REWARDAGG and QVALAGG).

Original languageEnglish
Pages (from-to)321-347
Number of pages27
JournalAutonomous Robots
Volume46
Issue number2
DOIs
StatePublished - Feb 2022

Keywords

  • Adaptive submodularity
  • Compressed sensing
  • Inverse reinforcement learning
  • Map exploration
  • Spatial search

Fingerprint

Dive into the research topics of 'Adaptive submodular inverse reinforcement learning for spatial search and map exploration'. Together they form a unique fingerprint.

Cite this