4:00 PM - 4:20 PM
[1B4-OS-41b-02] Learning Hierarchical State Space Models via Surprise- and Uncertainty-based Chunking
Keywords:World models, Hierarchical state space models, Temporal abstraction
Complex real-world tasks are often long-horizon, making world models that can accurately predict far into the future crucial for AI agents.
Hierarchical state-space models, which incorporate temporal hierarchies in latent states, have shown promise for long-term prediction by segmenting time series into subsequences and learning temporal abstraction.
However, existing methods relying on rigid subsequence length assumptions or significant changes in observation often perform poorly in environments where optimal subsequence lengths vary or environmental changes occur gradually.
This study proposes a method for learning hierarchical state-space models based on the discovery of frequently occurring, highly reusable patterns, drawing insights from chunking mechanisms in cognitive science.
Our method extracts frequent patterns by utilizing changes in surprise and uncertainty in low-level latent states.
Leveraging these patterns to learn high-level latent states reduces the complexity of transitions, enabling efficient long-term prediction.
Experiments on video prediction tasks show that our method outperforms the baselines, underscoring the effectiveness of hierarchical structures derived from frequent patterns for long-term prediction.
Hierarchical state-space models, which incorporate temporal hierarchies in latent states, have shown promise for long-term prediction by segmenting time series into subsequences and learning temporal abstraction.
However, existing methods relying on rigid subsequence length assumptions or significant changes in observation often perform poorly in environments where optimal subsequence lengths vary or environmental changes occur gradually.
This study proposes a method for learning hierarchical state-space models based on the discovery of frequently occurring, highly reusable patterns, drawing insights from chunking mechanisms in cognitive science.
Our method extracts frequent patterns by utilizing changes in surprise and uncertainty in low-level latent states.
Leveraging these patterns to learn high-level latent states reduces the complexity of transitions, enabling efficient long-term prediction.
Experiments on video prediction tasks show that our method outperforms the baselines, underscoring the effectiveness of hierarchical structures derived from frequent patterns for long-term prediction.
Authentication for paper PDF access
A password is required to view paper PDFs. If you are a registered participant, please log on the site from Participant Log In.
You could view the PDF with entering the PDF viewing password bellow.