Joseph0102's picture
Update README.md (#4)
cc86e06 verified
---
library_name: hunyuanworld-voyager
license: other
license_name: tencent-hunyuanworld-voyager-community
license_link: https://github.com/Tencent-Hunyuan/HunyuanWorld-Voyager/blob/main/LICENSE
language:
- en
- zh
tags:
- hunyuan3d
- worldmodel
- 3d-aigc
- 3d-generation
- 3d
- scene-generation
- image-to-video
pipeline_tag: image-to-video
extra_gated_eu_disallowed: true
---
<div align="center">
<a href=""><img src="https://img.shields.io/static/v1?label=Project%20Page&message=Web&color=green"></a> &ensp;
<a href="https://3d-models.hunyuan.tencent.com/voyager/voyager_en/assets/HYWorld_Voyager.pdf"><img src="https://img.shields.io/static/v1?label=Tech%20Report&message=Arxiv&color=red"></a> &ensp;
<a href="https://huggingface.co/tencent/HunyuanWorld-Voyager"><img src="https://img.shields.io/static/v1?label=HunyuanWorld-Voyager&message=HuggingFace&color=yellow"></a>
</div>
We introduce HunyuanWorld-Voyager, a novel video diffusion framework that generates world-consistent 3D point-cloud sequences from a single image with user-defined camera path. Voyager can generate 3D-consistent scene videos for world exploration following custom camera trajectories. It can also jointly generate aligned depth and RGB video for effective and direct 3D reconstruction.
![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/62e7c26236a8e8a827ff0891/ZVq46hyyfscgR8927wsq3.jpeg)
## ๐Ÿ”— BibTeX
If you find [Voyager](https://arxiv.org/abs/2506.04225) useful for your research and applications, please cite using this BibTeX:
```BibTeX
@article{huang2025voyager,
title={Voyager: Long-Range and World-Consistent Video Diffusion for Explorable 3D Scene Generation},
author={Huang, Tianyu and Zheng, Wangguandong and Wang, Tengfei and Liu, Yuhao and Wang, Zhenwei and Wu, Junta and Jiang, Jie and Li, Hui and Lau, Rynson WH and Zuo, Wangmeng and Guo, Chunchao},
journal={arXiv preprint arXiv:2506.04225},
year={2025}
}
```
## Acknowledgements
We would like to thank [HunyuanWorld](https://github.com/Tencent-Hunyuan/HunyuanWorld-1.0), [Hunyuan3D-2](https://github.com/Tencent-Hunyuan/Hunyuan3D-2), and [HunyuanVideo-I2V](https://github.com/Tencent-Hunyuan/HunyuanVideo-I2V). We also thank [VGGT](https://github.com/facebookresearch/vggt), [MoGE](https://github.com/microsoft/MoGe), [Metric3D](https://github.com/YvanYin/Metric3D), for their open research and exploration.