Abstract
Planning problems with large state spaces and unknown models can be solved by value-based reinforcement learning methods, using function approximation which abstracts states with feature vectors. Sample complexity increases when unnecessary features are selected. Conversely, omitting necessary features results in poor performances. A feature selection algorithm should add all necessary features while avoiding unnecessary ones. We introduce an online feature selection that considers the structure of relational MDPs (RMDPs) to reduce unnecessary features. Generalisation over different planning problems can be achieved by representing the approximated value-function with lifted representations of state variables and actions. Empirical results show that our method is practical for solving large RMDPs with reduced sample complexities and can generalise over different problems.
Original language | English |
---|---|
Publication status | Published - 16 Dec 2020 |
Event | 35th Workshop of the UK Planning and Scheduling Special Interest Group 2020 - Online Duration: 16 Dec 2020 → 16 Dec 2020 |
Workshop
Workshop | 35th Workshop of the UK Planning and Scheduling Special Interest Group 2020 |
---|---|
Abbreviated title | UK PlanSIG 2020 |
Period | 16/12/20 → 16/12/20 |