Merging pixel-derived abstract world models with hierarchical option training
Develop an integrated approach that jointly learns abstract symbolic world models directly from pixel inputs (e.g., via methods such as VisualPredicator, ExoPredicator, or predicate-learning from vision) together with hierarchical neural option training, enabling end-to-end acquisition of neurosymbolic world models and compositional skills from raw visual observations.
References
There is ongoing effort to learn abstract symbolic world models directly from pixels , but merging that line of work with option training remains open.
— Joint Learning of Hierarchical Neural Options and Abstract World Model
(2602.02799 - Piriyakulkij et al., 2 Feb 2026) in Section: Limitations and Future Direction