Peter Stone's Selected Publications

Classified by TopicClassified by Publication TypeSorted by DateSorted by First Author Last NameClassified by Funding Source


Disentangled Unsupervised Skill Discovery for Efficient Hierarchical Reinforcement Learning

Disentangled Unsupervised Skill Discovery for Efficient Hierarchical Reinforcement Learning.
Jiaheng Hu, Zizhao Wang, Roberto Martín-Martín, and Peter Stone.
In Conference on Neural Information Parocessing Systems (NeurIPS), December 2024.

Download

[PDF]2.7MB  

Abstract

A hallmark of intelligent agents is the ability to learn reusable skills purelyfrom unsupervised interaction with the environment. However, existingunsupervised skill discovery methods often learn entangled skills where one skillvariable simultaneously influences many entities in the environment, makingdownstream skill chaining extremely challenging. We propose DisentangledUnsupervised Skill Discovery (DUSDi), a method for learning disentangled skillsthat can be efficiently reused to solve downstream tasks. DUSDi decomposes skillsinto disentangled components, where each skill component only affects one factorof the state space. Importantly, these skill components can be concurrentlycomposed to generate low-level actions, and efficiently chained to tackledownstream tasks through hierarchical Reinforcement Learning. DUSDi defines anovel mutual-information-based objective to enforce disentanglement between theinfluences of different skill components, and utilizes value factorization tooptimize this objective efficiently. Evaluated in a set of challengingenvironments, DUSDi successfully learns disentangled skills, and significantlyoutperforms previous skill discovery methods when it comes to applying thelearned skills to solve downstream tasks. Code and skills visualization atjiahenghu.github.io/DUSDi-site/.

BibTeX Entry

@InProceedings{hu_neurips2024,
  author   = {Jiaheng Hu and Zizhao Wang and Roberto Martín-Martín and Peter Stone},
  title    = {Disentangled Unsupervised Skill Discovery for Efficient Hierarchical Reinforcement Learning},
  booktitle = {Conference on Neural Information Parocessing Systems (NeurIPS)},
  year     = {2024},
  month    = {December},
  location = {Vancouver, Canada},
  abstract = {A hallmark of intelligent agents is the ability to learn reusable skills purely
from unsupervised interaction with the environment. However, existing
unsupervised skill discovery methods often learn entangled skills where one skill
variable simultaneously influences many entities in the environment, making
downstream skill chaining extremely challenging. We propose Disentangled
Unsupervised Skill Discovery (DUSDi), a method for learning disentangled skills
that can be efficiently reused to solve downstream tasks. DUSDi decomposes skills
into disentangled components, where each skill component only affects one factor
of the state space. Importantly, these skill components can be concurrently
composed to generate low-level actions, and efficiently chained to tackle
downstream tasks through hierarchical Reinforcement Learning. DUSDi defines a
novel mutual-information-based objective to enforce disentanglement between the
influences of different skill components, and utilizes value factorization to
optimize this objective efficiently. Evaluated in a set of challenging
environments, DUSDi successfully learns disentangled skills, and significantly
outperforms previous skill discovery methods when it comes to applying the
learned skills to solve downstream tasks. Code and skills visualization at
jiahenghu.github.io/DUSDi-site/.
  },
}

Generated by bib2html.pl (written by Patrick Riley ) on Tue Nov 19, 2024 10:24:40