TheDenk commited on
Commit
88caf8d
1 Parent(s): 7fb0fd6

Base version

Browse files
Files changed (3) hide show
  1. README.md +54 -0
  2. config.json +30 -0
  3. diffusion_pytorch_model.safetensors +3 -0
README.md CHANGED
@@ -1,3 +1,57 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ language:
4
+ - en
5
+ tags:
6
+ - cogvideox
7
+ - video-generation
8
+ - video-to-video
9
+ - controlnet
10
+ - diffusers
11
  ---
12
+
13
+ # CogvideoX-5b Controlnet Extention
14
+
15
+ ### How to
16
+ Clone repo
17
+ ```bash
18
+ git clone https://github.com/TheDenk/cogvideox-controlnet.git
19
+ cd cogvideox-controlnet
20
+ ```
21
+
22
+ Create venv
23
+ ```bash
24
+ python -m venv venv
25
+ source venv/bin/activate
26
+ ```
27
+
28
+ Install requirements
29
+ ```bash
30
+ pip install -r requirements.txt
31
+ ```
32
+
33
+ ### Inference examples
34
+ #### Inference with cli
35
+ ```bash
36
+ python -m inference.cli_demo \
37
+ --video_path "resources/car.mp4" \
38
+ --prompt "car is moving among mountains" \
39
+ --controlnet_type "hed" \
40
+ --base_model_path THUDM/CogVideoX-5b \
41
+ --controlnet_model_path TheDenk/cogvideox-5b-controlnet-hed-v1
42
+ ```
43
+
44
+ #### Inference with Gradio
45
+ ```bash
46
+ python -m inference.gradio_web_demo \
47
+ --controlnet_type "hed" \
48
+ --base_model_path THUDM/CogVideoX-5b \
49
+ --controlnet_model_path TheDenk/cogvideox-5b-controlnet-hed-v1
50
+ ```
51
+
52
+
53
+ ## Acknowledgements
54
+ Original code and models [CogVideoX](https://github.com/THUDM/CogVideo/tree/main).
55
+
56
+ ## Contacts
57
+ <p>Issues should be raised directly in the repository. For professional support and recommendations please <a>welcomedenk@gmail.com</a>.</p>
config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "CogVideoXControlnet",
3
+ "_diffusers_version": "0.31.0.dev0",
4
+ "activation_fn": "gelu-approximate",
5
+ "attention_bias": true,
6
+ "attention_head_dim": 64,
7
+ "downscale_coef": 8,
8
+ "dropout": 0.0,
9
+ "flip_sin_to_cos": true,
10
+ "freq_shift": 0,
11
+ "in_channels": 3,
12
+ "max_text_seq_length": 226,
13
+ "norm_elementwise_affine": true,
14
+ "norm_eps": 1e-05,
15
+ "num_attention_heads": 30,
16
+ "num_layers": 12,
17
+ "out_proj_dim": 3072,
18
+ "patch_size": 2,
19
+ "sample_frames": 49,
20
+ "sample_height": 60,
21
+ "sample_width": 90,
22
+ "spatial_interpolation_scale": 1.875,
23
+ "temporal_compression_ratio": 4,
24
+ "temporal_interpolation_scale": 1.0,
25
+ "time_embed_dim": 512,
26
+ "timestep_activation_fn": "silu",
27
+ "use_learned_positional_embeddings": false,
28
+ "use_rotary_positional_embeddings": false,
29
+ "vae_channels": 16
30
+ }
diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b953a7aeac1b882aa69300ae72e6c904e03b14ef3b1202fec1bc52eac527252
3
+ size 1506883000