Image-to-Video
Diffusers
wruisi commited on
Commit
fdbd5d3
·
1 Parent(s): 2b16414

Update README

Browse files
Files changed (1) hide show
  1. README.md +5 -20
README.md CHANGED
@@ -55,10 +55,10 @@ The model was presented in the paper [A Very Big Video Reasoning Suite](https://
55
  | [VBVR-Wan2.2](https://huggingface.co/Video-Reason/VBVR-Wan2.2) | Wan2.2-I2V-A14B | Diffusers format |
56
  | [VBVR-Wan2.1-diffsynth](https://huggingface.co/Video-Reason/VBVR-Wan2.1-diffsynth) | Wan2.1-I2V-14B-720P | DiffSynth LoRA format |
57
  | [VBVR-Wan2.2-diffsynth](https://huggingface.co/Video-Reason/VBVR-Wan2.2-diffsynth) | Wan2.2-I2V-A14B | DiffSynth LoRA format |
58
- | [VBVR-LTX2.3-diffsynth](https://huggingface.co/Video-Reason/VBVR-LTX2.3-diffsynth) | LTX-Video-2.3 | DiffSynth LoRA format |
59
 
60
  ## Release Information
61
- VBVR-Wan2.1 is trained from Wan2.1-I2V-14B-720P without architectural modifications, as the goal of VBVR is to *investigate data scaling behavior* and provide *strong baseline models* for the video reasoning research community. Leveraging the VBVR-Dataset, which constitutes one of the largest video reasoning datasets to date, the VBVR model family achieved highest scores on VBVR-Bench.
62
 
63
  In this release, we present
64
  [**VBVR-Wan2.1**](https://huggingface.co/Video-Reason/VBVR-Wan2.1) (Diffusers format),
@@ -177,24 +177,9 @@ In this release, we present
177
 
178
  ## QuickStart
179
 
180
- ### Installation
181
-
182
- We recommend using [uv](https://docs.astral.sh/uv/) to manage the environment.
183
-
184
- > uv installation guide: <https://docs.astral.sh/uv/getting-started/installation/#installing-uv>
185
-
186
- ```bash
187
- pip install torch>=2.4.0 torchvision>=0.19.0 transformers Pillow huggingface_hub[cli]
188
- uv pip install git+https://github.com/huggingface/diffusers
189
- ```
190
-
191
- ### Example Code
192
-
193
- ```bash
194
- huggingface-cli download Video-Reason/VBVR-Wan2.1 --local-dir ./VBVR-Wan2.1
195
- python example.py \
196
- --model_path ./VBVR-Wan2.1
197
- ```
198
 
199
  ## Citation
200
 
 
55
  | [VBVR-Wan2.2](https://huggingface.co/Video-Reason/VBVR-Wan2.2) | Wan2.2-I2V-A14B | Diffusers format |
56
  | [VBVR-Wan2.1-diffsynth](https://huggingface.co/Video-Reason/VBVR-Wan2.1-diffsynth) | Wan2.1-I2V-14B-720P | DiffSynth LoRA format |
57
  | [VBVR-Wan2.2-diffsynth](https://huggingface.co/Video-Reason/VBVR-Wan2.2-diffsynth) | Wan2.2-I2V-A14B | DiffSynth LoRA format |
58
+ | [VBVR-LTX2.3-diffsynth](https://huggingface.co/Video-Reason/VBVR-LTX2.3-diffsynth) | LTX-2.3 | DiffSynth LoRA format |
59
 
60
  ## Release Information
61
+ VBVR-LTX2.3 is trained from LTX-2.3 without architectural modifications, as the goal of VBVR is to *investigate data scaling behavior* and provide *strong baseline models* for the video reasoning research community. Leveraging the VBVR-Dataset, which constitutes one of the largest video reasoning datasets to date, the VBVR model family achieved highest scores on VBVR-Bench.
62
 
63
  In this release, we present
64
  [**VBVR-Wan2.1**](https://huggingface.co/Video-Reason/VBVR-Wan2.1) (Diffusers format),
 
177
 
178
  ## QuickStart
179
 
180
+ ### Inference
181
+ For running inference, please refer to the [**official guide**](https://github.com/Video-Reason/VBVR-Wan2.2?tab=readme-ov-file#ltx-23-inference) in the VBVR-Wan2.2 GitHub repository.
182
+ This repository contains the latest instructions, configurations, and examples for performing inference with the VBVR family models.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
183
 
184
  ## Citation
185