Update README.md
Browse files
README.md
CHANGED
|
@@ -1,14 +1,25 @@
|
|
| 1 |
# PKU-DyMVHumans
|
| 2 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3 |
## Overview
|
| 4 |
|
| 5 |
PKU-DyMVHumans is a versatile human-centric dataset designed for high-fidelity reconstruction and rendering of dynamic human performances in markerless multi-view capture settings.
|
| 6 |
|
| 7 |
It comprises 32 humans across 45 different dynamic scenarios, each featuring highly detailed appearances and complex human motions.
|
| 8 |
|
|
|
|
|
|
|
| 9 |
## Key Features
|
| 10 |
|
| 11 |
-
- **High-fidelity
|
| 12 |
|
| 13 |
- **High-detailed appearance**: It captures complex cloth deformation, and intricate texture details, like delicate satin ribbon and special headwear.
|
| 14 |
|
|
@@ -16,6 +27,7 @@ It comprises 32 humans across 45 different dynamic scenarios, each featuring hig
|
|
| 16 |
|
| 17 |
- **Human-object/scene interactions**: It includes human-object interactions, multi-person interactions and complex scene effects (like smoking).
|
| 18 |
|
|
|
|
| 19 |
## Dataset format
|
| 20 |
|
| 21 |
For each scene, we provide the multi-view images (`./case_name/per_view/cam_*/images/`), the coarse foreground with RGBA channels (`./case_name/per_view/cam_*/images/`), as well as the coarse foreground segmentation (`./case_name/per_view/cam_*/pha/`), which are obtained using [BackgroundMattingV2](https://github.com/PeterL1n/BackgroundMattingV2).
|
|
@@ -54,7 +66,7 @@ If you find this repo is helpful, please cite:
|
|
| 54 |
|
| 55 |
@article{zheng2024PKU-DyMVHumans,
|
| 56 |
title={PKU-DyMVHumans: A Multi-View Video Benchmark for High-Fidelity Dynamic Human Modeling},
|
| 57 |
-
author={Zheng, Xiaoyun and Liao, Liwei and Li,Xufeng and Jiao, Jianbo and Wang, Rongjie and Gao, Feng and Wang, Shiqi and Wang, Ronggang},
|
| 58 |
journal={IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
|
| 59 |
year={2024}
|
| 60 |
}
|
|
|
|
| 1 |
# PKU-DyMVHumans
|
| 2 |
|
| 3 |
+
## Sources
|
| 4 |
+
|
| 5 |
+
Project page:https://pku-dymvhumans.github.io/
|
| 6 |
+
|
| 7 |
+
Github: https://github.com/zhengxyun/PKU-DyMVHumans
|
| 8 |
+
|
| 9 |
+
Paper: https://arxiv.org/abs/2403.16080
|
| 10 |
+
|
| 11 |
+
|
| 12 |
## Overview
|
| 13 |
|
| 14 |
PKU-DyMVHumans is a versatile human-centric dataset designed for high-fidelity reconstruction and rendering of dynamic human performances in markerless multi-view capture settings.
|
| 15 |
|
| 16 |
It comprises 32 humans across 45 different dynamic scenarios, each featuring highly detailed appearances and complex human motions.
|
| 17 |
|
| 18 |
+
|
| 19 |
+
|
| 20 |
## Key Features
|
| 21 |
|
| 22 |
+
- **High-fidelity performance**:We construct a multi-view system to capture humans in motion, containing 56/60 synchronous cameras with 1080P or 4K resolution
|
| 23 |
|
| 24 |
- **High-detailed appearance**: It captures complex cloth deformation, and intricate texture details, like delicate satin ribbon and special headwear.
|
| 25 |
|
|
|
|
| 27 |
|
| 28 |
- **Human-object/scene interactions**: It includes human-object interactions, multi-person interactions and complex scene effects (like smoking).
|
| 29 |
|
| 30 |
+
|
| 31 |
## Dataset format
|
| 32 |
|
| 33 |
For each scene, we provide the multi-view images (`./case_name/per_view/cam_*/images/`), the coarse foreground with RGBA channels (`./case_name/per_view/cam_*/images/`), as well as the coarse foreground segmentation (`./case_name/per_view/cam_*/pha/`), which are obtained using [BackgroundMattingV2](https://github.com/PeterL1n/BackgroundMattingV2).
|
|
|
|
| 66 |
|
| 67 |
@article{zheng2024PKU-DyMVHumans,
|
| 68 |
title={PKU-DyMVHumans: A Multi-View Video Benchmark for High-Fidelity Dynamic Human Modeling},
|
| 69 |
+
author={Zheng, Xiaoyun and Liao, Liwei and Li, Xufeng and Jiao, Jianbo and Wang, Rongjie and Gao, Feng and Wang, Shiqi and Wang, Ronggang},
|
| 70 |
journal={IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
|
| 71 |
year={2024}
|
| 72 |
}
|