Skip to main content

2002 | OriginalPaper | Buchkapitel

Q-Cut—Dynamic Discovery of Sub-goals in Reinforcement Learning

verfasst von : Ishai Menache, Shie Mannor, Nahum Shimkin

Erschienen in: Machine Learning: ECML 2002

Verlag: Springer Berlin Heidelberg

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

We present the Q-Cut algorithm, a graph theoretic approach for automatic detection of sub-goals in a dynamic environment, which is used for acceleration of the Q-Learning algorithm. The learning agent creates an on-line map of the process history, and uses an efficient Max-Flow/Min-Cut algorithm for identifying bottlenecks. The policies for reaching bottlenecks are separately learned and added to the model in a form of options (macro-actions). We then extend the basic Q-Cut algorithm to the Segmented Q-Cut algorithm, which uses previously identified bottlenecks for state space partitioning, necessary for finding additional bottlenecks in complex environments. Experiments show significant performance improvements, particulary in the initial learning phase.

Metadaten
Titel
Q-Cut—Dynamic Discovery of Sub-goals in Reinforcement Learning
verfasst von
Ishai Menache
Shie Mannor
Nahum Shimkin
Copyright-Jahr
2002
Verlag
Springer Berlin Heidelberg
DOI
https://doi.org/10.1007/3-540-36755-1_25

Neuer Inhalt