MaziyarPanahi
commited on
Commit
•
dea7266
1
Parent(s):
62fe087
Upload 3 files
Browse files- prompt.txt +1 -0
- pytorch_lora_weights.safetensors +3 -0
- training_params.json +1 -0
prompt.txt
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
photo of a sks Vision Pro
|
pytorch_lora_weights.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:45240d932626581c8694f808ae805bec74f924fa70657e8448edb716edb4c5ef
|
3 |
+
size 23401064
|
training_params.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"model": "stabilityai/stable-diffusion-xl-base-1.0", "revision": null, "tokenizer": null, "image_path": "vision_pro_dreambooth_project/autotrain-data", "class_image_path": null, "prompt": "photo of a sks Vision Pro", "class_prompt": null, "num_class_images": 100, "class_labels_conditioning": null, "prior_preservation": false, "prior_loss_weight": 1.0, "project_name": "vision_pro_dreambooth_project", "seed": 42, "resolution": 1024, "center_crop": false, "train_text_encoder": false, "batch_size": 1, "sample_batch_size": 4, "epochs": 1, "num_steps": 500, "checkpointing_steps": 100000, "resume_from_checkpoint": null, "gradient_accumulation": 4, "disable_gradient_checkpointing": false, "lr": 0.0001, "scale_lr": false, "scheduler": "constant", "warmup_steps": 0, "num_cycles": 1, "lr_power": 1.0, "dataloader_num_workers": 0, "use_8bit_adam": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_weight_decay": 0.01, "adam_epsilon": 1e-08, "max_grad_norm": 1.0, "allow_tf32": false, "prior_generation_precision": null, "local_rank": -1, "xformers": false, "pre_compute_text_embeddings": false, "tokenizer_max_length": null, "text_encoder_use_attention_mask": false, "rank": 4, "xl": true, "fp16": true, "bf16": false, "repo_id": null, "push_to_hub": false, "username": null, "validation_prompt": null, "num_validation_images": 4, "validation_epochs": 50, "checkpoints_total_limit": null, "validation_images": null, "logging": false}
|