asenella commited on
Commit
667715b
1 Parent(s): cb6337d

Uploading JNFCLIP in asenella/mhd128_embedding_clip_dim_64_beta_50_seed_0

Browse files
Files changed (7) hide show
  1. CLIP_networks.pkl +3 -0
  2. README.md +13 -0
  3. decoders.pkl +3 -0
  4. environment.json +1 -0
  5. joint_encoder.pkl +3 -0
  6. model.pt +3 -0
  7. model_config.json +1 -0
CLIP_networks.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:921dbb8786b463c765ea33d90b812963d4c83ceb2b8bae621f4fa437794e7414
3
+ size 49378774
README.md ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language: en
3
+ tags:
4
+ - multivae
5
+ license: apache-2.0
6
+ ---
7
+
8
+ ### Downloading this model from the Hub
9
+ This model was trained with multivae. It can be downloaded or reloaded using the method `load_from_hf_hub`
10
+ ```python
11
+ >>> from multivae.models import AutoModel
12
+ >>> model = AutoModel.load_from_hf_hub(hf_hub_path="your_hf_username/repo_name")
13
+ ```
decoders.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d91d8d95ba72c156582da59883536df66b8b3f6e57c08d5ad5342f787974eb44
3
+ size 36868419
environment.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"name": "EnvironmentConfig", "python_version": "3.10"}
joint_encoder.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:925bab2163cc318870f4998a9d533d23fb560a60ce4a0d51a9a79f8a7808275a
3
+ size 53581704
model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f34a848621ce7053753c558d41fad96c30283e2cf3c8ff1c160303e2e3fc530b
3
+ size 145694341
model_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"name": "JNFCLIPConfig", "n_modalities": 3, "latent_dim": 128, "input_dims": {"image": [3, 28, 28], "audio": [1, 32, 128], "trajectory": [200]}, "uses_likelihood_rescaling": true, "rescale_factors": null, "decoders_dist": {"image": "normal", "audio": "normal", "trajectory": "normal"}, "decoder_dist_params": {}, "custom_architectures": ["CLIP_networks", "decoders", "joint_encoder"], "warmup": 200, "nb_epochs_clip": 50, "embedding_clip_dim": 64, "weights_clip": null, "beta": 0.5}