Uploading JMVAE in asenella/reproduce_jmvae_seed_1
Browse files- README.md +13 -0
- decoders.pkl +3 -0
- encoders.pkl +3 -0
- joint_encoder.pkl +3 -0
- model.pt +3 -0
- model_config.json +1 -0
README.md
ADDED
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
language: en
|
3 |
+
tags:
|
4 |
+
- multivae
|
5 |
+
license: apache-2.0
|
6 |
+
---
|
7 |
+
|
8 |
+
### Downloading this model from the Hub
|
9 |
+
This model was trained with multivae. It can be downloaded or reloaded using the method `load_from_hf_hub`
|
10 |
+
```python
|
11 |
+
>>> from multivae.models import AutoModel
|
12 |
+
>>> model = AutoModel.load_from_hf_hub(hf_hub_path="your_hf_username/repo_name")
|
13 |
+
```
|
decoders.pkl
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7d2b9a282590f4baed7d8d894c0281905a20e53de93f5dc02299600b376a1494
|
3 |
+
size 4020770
|
encoders.pkl
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9b9e7613f8fb01447a69f135c9bd328d45df4e42970737cc276b234e5f5c3d9e
|
3 |
+
size 4283432
|
joint_encoder.pkl
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:696f5303f1226a298c23137eb92f3645b75f6ee4c27a2771fc89f3102206eb53
|
3 |
+
size 4008176
|
model.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ba6b3b66cbb8ee0a0d2720c38d48108e9922f227dbec3d373db86cdfceeb7b3e
|
3 |
+
size 12257571
|
model_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"name": "JMVAEConfig", "n_modalities": 2, "latent_dim": 64, "input_dims": {"images": [1, 28, 28], "labels": [1]}, "uses_likelihood_rescaling": false, "rescale_factors": null, "decoders_dist": {"images": "bernoulli", "labels": "categorical"}, "decoder_dist_params": {}, "custom_architectures": ["encoders", "decoders", "joint_encoder"], "alpha": 0.1, "warmup": 200}
|