Skill Disentanglement in Reproducing Kernel Hilbert Space

Research output: Contribution to journalConference articlepeer-review

Abstract

Unsupervised Skill Discovery aims at learning diverse skills without any extrinsic rewards and leverage them as prior for learning a variety of downstream tasks. Existing approaches to unsupervised reinforcement learning typically involve discovering skills through empowerment-driven techniques or by maximizing entropy to encourage exploration. However, this mutual information objective often results in either static skills that discourage exploration or maximise coverage at the expense of non-discriminable skills. Instead of focusing only on maximizing bounds on f-divergence, we combine it with Integral Probability Metrics to maximize the distance between distributions to promote behavioural diversity and enforce disentanglement. Our method, Hilbert Unsupervised Skill Discovery (HUSD), provides an additional objective that seeks to obtain exploration and separability of state-skill pairs by maximizing the Maximum Mean Discrepancy between the joint distribution of skills and states and the product of their marginals in Reproducing Kernel Hilbert Space. Our results on Unsupervised RL Benchmark show that HUSD outperforms previous exploration algorithms on state-based tasks.
Original languageEnglish
Pages (from-to)16153-62
Number of pages10
JournalProceedings of the AAAI Conference on Artificial Intelligence
VolumeVol. 39
DOIs
Publication statusPublished - 11 Apr 2025
EventThe 39th Annual AAAI Conference on Artificial Intelligence - Pennsylvania Convention Center, Philadelphia, United States
Duration: 27 Feb 20254 Mar 2025
https://aaai.org/conference/aaai/aaai-25/

Cite this