araminta_k PRO

alvdansen

AI & ML interests

Finetuning LoRA Models and ComfyUI Workflows. If you want to use my LoRAs commercially, please reach out!

Articles

Organizations

alvdansen's activity

posted an update about 1 month ago
view post
Post
1990
📸Photo LoRA Drop📸

I've been working on this one for a few days, but really I've had this dataset for a few years! I collected a bunch of open access photos online back in late 2022, but I was never happy enough with how they played with the base model!

I am so thrilled that they look so nice with Flux!

This for me is a version one of this model - I still see room for improvement and possibly expansion of it's 40 image dataset. For those who are curious:

40 Image
3200 Steps
Dim 32
3e-4

Enjoy! Create! Big thank you to Glif for sponsoring the model creation! :D

alvdansen/flux_film_foto
posted an update about 1 month ago
view post
Post
5416
Alright Ya'll

I know it's a Saturday, but I decided to release my first Flux Dev Lora.

A retrain of my "Frosting Lane" model and I am sure the styles will just keep improving.

Have fun! Link Below - Thanks again to @ostris for the trainer and Black Forest Labs for the awesome model!

alvdansen/frosting_lane_flux
posted an update 2 months ago
view post
Post
4478
New model drop...🥁

FROSTING LANE REDUX

The v1 of this model was released during a big model push, so I think it got lost in the shuffle. I revisited it for a project and realized it wasn't inventive enough around certain concepts, so I decided to retrain.

alvdansen/frosting_lane_redux

I think the original model was really strong on it's own, but because it was trained on fewer images I found that it was producing a very lackluster range of facial expressions, so I wanted to improve that.

The hardest part of creating models like this, I find, is maintaining the detailed linework without without overfitting. It takes a really balanced dataset and I repeat the data 12 times during the process, stopping at the last 10-20 epochs.

It is very difficult to predict the exact amount of time needed, so for me it is crucial to do epoch stops. Every model has a different threshold for ideal success.
replied to victor's post 2 months ago
view reply

Yeah that would be really helpful, I haven't had the time to try and do something like that.

replied to their post 2 months ago
posted an update 3 months ago
view post
Post
2932
I really like what the @jasperAITeam designed with Flash LoRA. It works really well for something that generates so quickly, and I'm excited to test it out with Animate Diff, because I recently was testing LCM on it's own for AD and the results were already promising.

I put together my own page of models using their code and LoRA. Enjoy!

alvdansen/flash-lora-araminta-k-styles
replied to their post 3 months ago
posted an update 3 months ago
replied to their post 3 months ago
view reply

I responded on X with the best way to contact me.

replied to their post 3 months ago
posted an update 3 months ago
view post
Post
5670
New LoRA Model!

I trained this model on a new spot I'm really excited to share (soon!)

This Monday I will be posting my first beginning to end blog showing the tool I've used, dataset, captioning techniques, and parameters to finetune this LoRA.

For now, check out the model in the link below.

alvdansen/m3lt
·
replied to louisbrulenaudet's post 3 months ago
posted an update 3 months ago
view post
Post
2466
Per popular request, I'm working on a beginning to end LoRA training workflow blog for a style.

It will focus on dataset curation through training on a pre-determined style to give a better insight on my process.

Curious what are some questions you might have that I can try to answer in it?
posted an update 3 months ago
view post
Post
2422
A few new styles added as SDXL LoRA:

Midsommar Cartoon
A playful cartoon style featuring bold colors and a retro aesthetic. Personal favorite at the moment.
alvdansen/midsommarcartoon
---
Wood Block XL
I've started training public domain styles to create some interesting datasets. In this case I found a group of images taken from really beautiful and colorful Japanese Blockprints.
alvdansen/wood-block-xl
--
Dimension W
For this model I did actually end up working on an SD 1.5 model as well as an SDXL. I prefer the SDXL version, and I am still looking for parameters I am really happy with for SD 1.5. That said, both have their merits. I trained this with the short film I am working on in mind.
alvdansen/dimension-w
alvdansen/dimension-w-sd15
replied to their post 3 months ago
view reply

I typically use Kohya, but I also test a lot of platform services for the right one because I am a creature of comfort :)

replied to their post 3 months ago
view reply

I need to double check the train_text_encoder_frac as I typically don't mess with that. For rank I'm usually at 32.

posted an update 3 months ago
view post
Post
951
Hey all!

Here I take a somewhat strong stance and am petitioning to revisit the default training parameters on the Diffusers LoRA page.

In my opinion and after observing and testing may training pipelines shared by startups and resources, I have found that many of them exhibit the same types of issues. Upon discussing with some of these founders and creators, the common theme has been working backwards from the Diffusers LoRA page.

In this article, I explain why the defaults in the Diffuser LoRA code produce some positive results, which can be initially misleading, and a suggestion on how that could be improved.

https://huggingface.co/blog/alvdansen/revisit-diffusers-default-params
  • 4 replies
·
replied to their post 3 months ago
view reply

No - I change them however it’s very case by case. I am trying to emphasize elements other than hyperparameters, because in my experience these concepts apply to a variety of hyperparameters.

replied to their post 3 months ago
replied to their post 3 months ago
view reply

Are you on X? You can contact me there @araminta_k

replied to their post 3 months ago
replied to their post 3 months ago
replied to their post 3 months ago
view reply

I will need to take a look at what the exact backend of face to all is. What is the result you’re getting ?

posted an update 3 months ago
view post
Post
3233
Hey All!

I've been asked a lot of share more on how I train LoRAs. The truth is I don't think my advice is very helpful without also including more contextual, theoretical commentary on how I **think** about training LoRAs for SDXL and other models.

I wrote a first article here about it - let me know what you think.

https://huggingface.co/blog/alvdansen/thoughts-on-lora-training-1

Edit: Also people kept asking where to start so I made a list of possible resources:
https://huggingface.co/blog/alvdansen/thoughts-on-lora-training-pt-2-training-services
·
replied to their post 3 months ago
posted an update 3 months ago
view post
Post
6754
I had a backlog of LoRA model weights for SDXL that I decided to prioritize this weekend and publish. I know many are using SD3 right now, however if you have the time to try them, I hope you enjoy them.

I intend to start writing more fully on the thought process behind my approach to curating and training style and subject finetuning, beginning this next week.

Thank you for reading this post! You can find the models on my page and I'll drop a few previews here.
·
replied to victor's post 3 months ago
replied to victor's post 3 months ago
view reply

🙌 Thank you so much for sharing! I will be sharing a training workflow in the coming week :D