AlekseyCalvin
commited on
Commit
•
bd984e4
1
Parent(s):
75b9dd8
Update README.md
Browse files
README.md
CHANGED
@@ -44,13 +44,13 @@ widget:
|
|
44 |
---
|
45 |
<Gallery />
|
46 |
|
47 |
-
# Mayakovsky Style Soviet Constructivist Posters & Cartoons Flux LoRA
|
48 |
Trained via Ostris' [ai-toolkit](https://replicate.com/ostris/flux-dev-lora-trainer/train) on 50 high-resolution scans of 1910s/1920s posters & artworks by the great Soviet **poet, artist, & Marxist activist Vladimir Mayakovsky**. <br>
|
49 |
For this training experiment, we first spent many days rigorously translating the textual elements (slogans, captions, titles, inset poems, speech fragments, etc), with form/signification/rhymes intact, throughout every image subsequently used for training. <br>
|
50 |
These translated textographic elements were, furthermore, re-placed by us into their original visual contexts, using fonts matched up to the sources. <br>
|
51 |
We then manually composed highly detailed paragraph-long captions, wherein we detailed both the graphic and the textual content of each piece, its layout, as well as the most intuitive/intended apprehension of each composition. <br>
|
52 |
This version of the resultent LoRA was trained on our custom Schnell-based checkpoint (Historic Color 2), available [here in fp8 Safetensors](https://huggingface.co/AlekseyCalvin/HistoricColorSoonrFluxV2/tree/main) and [here in Diffusers format](https://huggingface.co/AlekseyCalvin/HistoricColorSoonr_v2_FluxSchnell_Diffusers). <br>
|
53 |
-
|
54 |
No synthetic data was used for the training, nor any auto-generated captions! Everything was manually and attentively pre-curated with a deep respect for the sources used. <br>
|
55 |
|
56 |
|
|
|
44 |
---
|
45 |
<Gallery />
|
46 |
|
47 |
+
# Mayakovsky Style Soviet Constructivist Posters & Cartoons Flux LoRA – Version 2 – by SOON®
|
48 |
Trained via Ostris' [ai-toolkit](https://replicate.com/ostris/flux-dev-lora-trainer/train) on 50 high-resolution scans of 1910s/1920s posters & artworks by the great Soviet **poet, artist, & Marxist activist Vladimir Mayakovsky**. <br>
|
49 |
For this training experiment, we first spent many days rigorously translating the textual elements (slogans, captions, titles, inset poems, speech fragments, etc), with form/signification/rhymes intact, throughout every image subsequently used for training. <br>
|
50 |
These translated textographic elements were, furthermore, re-placed by us into their original visual contexts, using fonts matched up to the sources. <br>
|
51 |
We then manually composed highly detailed paragraph-long captions, wherein we detailed both the graphic and the textual content of each piece, its layout, as well as the most intuitive/intended apprehension of each composition. <br>
|
52 |
This version of the resultent LoRA was trained on our custom Schnell-based checkpoint (Historic Color 2), available [here in fp8 Safetensors](https://huggingface.co/AlekseyCalvin/HistoricColorSoonrFluxV2/tree/main) and [here in Diffusers format](https://huggingface.co/AlekseyCalvin/HistoricColorSoonr_v2_FluxSchnell_Diffusers). <br>
|
53 |
+
To produce this earlier checkpoint (the full 5000 run is available in [another](https://huggingface.co/AlekseyCalvin/Mayakovsky_Posters_2_5kSt) repo), the training went on for 3600 steps at a DiT Learning Rate of .00002, batch 1, with the ademamix8bit optimizer, and both text encoders trained alongside the DiT!<br>
|
54 |
No synthetic data was used for the training, nor any auto-generated captions! Everything was manually and attentively pre-curated with a deep respect for the sources used. <br>
|
55 |
|
56 |
|