Thanks to visit codestin.com
Credit goes to github.com

Skip to content

World-In-World/world-in-world

Repository files navigation

World-in-World logo

Website badge arXiv badge HF badge GitHub badge

World-in-World is a unified closed-loop benchmark and toolkit for evaluating visual world models (WMs) by their embodied utility rather than only image or video appearance. World-in-World provides: (1) a unified online planning strategy that works with different WMs, (2) a unified action API that adapts to text, viewpoint, and low‑level controls, and (3) a task suite covering Active Recognition (AR), Active Embodied QA (A‑EQA), Image‑Goal Navigation (IGNav), and Robotic Manipulation.


📰 News

  • 2025-10-22: Preprint released on arXiv. Landing page and repository initialized.

✨ Overview

Overview

In this work, we propose World-in-World, which wraps generative World models In a closed-loop World interface to measure their practical utility for embodied agents. We test whether generated worlds actually enhance embodied reasoning and task performance—for example, helping an agent perceive the environment, plan and execute actions, and re-plan based on new observations within such a closed loop. Establishing this evaluation framework is essential for tracking genuine progress across the rapidly expanding landscape of visual world models and embodied AI.


🚧 Repository Status

The release will follow the to‑do list below and will be updated continuously.

Under construction

  • Full documentation and tutorials for environment setup and task evaluation.
    • AR, IGNav, AEQA
    • Manipulation
  • WM post‑training instructions
  • Instructions to add a new WM to World‑in‑World
  • Additional tools and scripts

🚀 Getting Started

1) Documentation structure

2) Checklist for running an evaluation

For any task, complete the following steps in order.

  1. Set up environments.
  2. Download scene datasets.
  3. Download evaluation episodes.
  4. Deploy policies (VLM policy, heuristic policy, diffusion policy).
  5. Deploy other task‑related models if needed.
  6. Deploy the WM server.
  7. Run the evaluation script.
  8. Accumulate results.

After the first run, the environment and datasets are in place. For later runs, you usually only repeat steps 4–8. If you encounter any issue, please feel free to open an issue or contact us.


📝 Citation

If you find this work useful, please cite:

@misc{zhang2025worldinworld,
  title        = {World-in-World: World Models in a Closed-Loop World},
  author       = {Zhang, Jiahan and Jiang, Muqing and Dai, Nanru and Lu, Taiming and Uzunoglu, Arda and Zhang, Shunchi and Wei, Yana and Wang, Jiahao and Patel, Vishal M. and Liang, Paul Pu and Khashabi, Daniel and Peng, Cheng and Chellappa, Rama and Shu, Tianmin and Yuille, Alan and Du, Yilun and Chen, Jieneng},
  year         = {2025},
  eprint       = {2510.18135},
  archivePrefix= {arXiv},
}

About

Code implementation of the paper "World-in-World: World Models in a Closed-Loop World"

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published