Update README.md
Browse files
README.md
CHANGED
|
@@ -93,4 +93,30 @@ pipeline_tag: image-to-video
|
|
| 93 |
<td style="vertical-align: middle;">Fine-tuned LoRA for <a href="https://huggingface.co/datasets/amagipeng/VR-Bench">TrapField</a> tasks (easy, medium, and hard) from base model <a href="https://huggingface.co/Wan-AI/Wan2.2-TI2V-5B">Wan2.2-TI2V-5B</a>.</td>
|
| 94 |
</tr>
|
| 95 |
</tbody>
|
| 96 |
-
</table>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 93 |
<td style="vertical-align: middle;">Fine-tuned LoRA for <a href="https://huggingface.co/datasets/amagipeng/VR-Bench">TrapField</a> tasks (easy, medium, and hard) from base model <a href="https://huggingface.co/Wan-AI/Wan2.2-TI2V-5B">Wan2.2-TI2V-5B</a>.</td>
|
| 94 |
</tr>
|
| 95 |
</tbody>
|
| 96 |
+
</table>
|
| 97 |
+
|
| 98 |
+
<h2 align="center">📑 Citation</h2>
|
| 99 |
+
|
| 100 |
+
<p align="center">
|
| 101 |
+
If you use this model or the VR-Bench dataset in your work, please cite:
|
| 102 |
+
</p>
|
| 103 |
+
|
| 104 |
+
<p align="center">
|
| 105 |
+
📄 <a href="https://arxiv.org/abs/2511.15065">
|
| 106 |
+
Reasoning via Video: The First Evaluation of Video Models' Reasoning Abilities through Maze-Solving Tasks
|
| 107 |
+
</a>
|
| 108 |
+
</p>
|
| 109 |
+
|
| 110 |
+
<pre>
|
| 111 |
+
<code>
|
| 112 |
+
@misc{yang2025reasoningvideoevaluationvideo,
|
| 113 |
+
title={Reasoning via Video: The First Evaluation of Video Models' Reasoning Abilities through Maze-Solving Tasks},
|
| 114 |
+
author={Cheng Yang and Haiyuan Wan and Yiran Peng and Xin Cheng and Zhaoyang Yu and Jiayi Zhang and Junchi Yu and Xinlei Yu and Xiawu Zheng and Dongzhan Zhou and Chenglin Wu},
|
| 115 |
+
year={2025},
|
| 116 |
+
eprint={2511.15065},
|
| 117 |
+
archivePrefix={arXiv},
|
| 118 |
+
primaryClass={cs.CV},
|
| 119 |
+
url={https://arxiv.org/abs/2511.15065},
|
| 120 |
+
}
|
| 121 |
+
</code>
|
| 122 |
+
</pre>
|