DEV Community

Cover image for Huge FLUX LoRA vs Fine Tuning / DreamBooth Experiments Completed, Moreover Batch Size 1 vs 7
Furkan Gözükara
Furkan Gözükara

Posted on

1 1 1 1 1

Huge FLUX LoRA vs Fine Tuning / DreamBooth Experiments Completed, Moreover Batch Size 1 vs 7

Huge FLUX LoRA vs Fine Tuning / DreamBooth Experiments Completed, Moreover Batch Size 1 vs 7 Fully Tested as Well, Not Only for Realism But Also for Stylization — 15 vs 256 images having datasets compared as well (expressions / emotions tested too) — Used Kohya GUI for training

Full Article Link

Details

  • Download images in full resolution to see prompts and model names

  • All trainings are done with Kohya GUI, perfectly can be done locally on Windows, and all trainings were 1024x1024 pixels

  • Fine Tuning / DreamBooth works as low as 6 GB GPUs (0 quality degrade totally same as 48 GB config)

  • Best quality of LoRA requires 48 GB GPUs , 24 GB also works really good and minimum 8 GB GPU is necessary for LoRA (lots of quality degrade)

  • https://www.patreon.com/posts/112099700

  • Full size grids are also shared for the followings: https://www.patreon.com/posts/112099700

  • Training used 15 images dataset : 15_Images_Dataset.png

  • Training used 256 images dataset : 256_Images_Dataset.png

  • 15 Images Dataset, Batch Size 1 Fine Tuning Training : 15_imgs_BS_1_Realism_Epoch_Test.jpg , 15_imgs_BS_1_Style_Epoch_Test.jpg

  • 15 Images Dataset, Batch Size 7 Fine Tuning Training : 15_imgs_BS_7_Realism_Epoch_Test.jpg , 15_imgs_BS_7_Style_Epoch_Test.jpg

  • 256 Images Dataset, Batch Size 1 Fine Tuning Training : 256_imgs_BS_1_Realism_Epoch_Test.jpg , 256_imgs_BS_1_Stylized_Epoch_Test.jpg

  • 256 Images Dataset, Batch Size 7 Fine Tuning Training : 256_imgs_BS_7_Realism_Epoch_Test.jpg , 256_imgs_BS_7_Style_Epoch_Test.jpg

  • 15 Images Dataset, Batch Size 1 LoRA Training : 15_imgs_LORA_BS_1_Realism_Epoch_Test.jpg , 15_imgs_LORA_BS_1_Style_Epoch_Test.jpg

  • 15 Images Dataset, Batch Size 7 LoRA Training : 15_imgs_LORA_BS_7_Realism_Epoch_Test.jpg , 15_imgs_LORA_BS_7_Style_Epoch_Test.jpg

  • 256 Images Dataset, Batch Size 1 LoRA Training : 256_imgs_LORA_BS_1_Realism_Epoch_Test.jpg , 256_imgs_LORA_BS_1_Style_Epoch_Test.jpg

  • 256 Images Dataset, Batch Size 7 LoRA Training : 256_imgs_LORA_BS_7_Realism_Epoch_Test.jpg , 256_imgs_LORA_BS_7_Style_Epoch_Test.jpg

Comparisons

  • Fine Tuning / DreamBooth 15 vs 256 images and Batch Size 1 vs 7 for Realism : Fine_Tuning_15_vs_256_imgs_BS1_vs_BS7.jpg

  • Fine Tuning / DreamBooth 15 vs 256 images and Batch Size 1 vs 7 for Style : 15_vs_256_imgs_BS1_vs_BS7_Fine_Tuning_Style_Comparison.jpg

  • LoRA Training 15 vs 256 images vs Batch Size 1 vs 7 for Realism : LoRA_15_vs_256_imgs_BS1_vs_BS7.jpg

  • LoRA Training 15 vs 256 images vs Batch Size 1 vs 7 for Style : 15_vs_256_imgs_BS1_vs_BS7_LoRA_Style_Comparison.jpg

  • Testing smiling expression for LoRA Trainings : LoRA_Expression_Test_Grid.jpg

  • Testing smiling expression for Fine Tuning / DreamBooth Trainings : Fine_Tuning_Expression_Test_Grid.jpg

Fine Tuning / DreamBooth vs LoRA Comparisons

  • 15 Images Fine Tuning vs LoRA at Batch Size 1 : 15_imgs_BS1_LoRA_vs_Fine_Tuning.jpg

  • 15 Images Fine Tuning vs LoRA at Batch Size 7 : 15_imgs_BS7_LoRA_vs_Fine_Tuning.jpg

  • 256 Images Fine Tuning vs LoRA at Batch Size 1 : 256_imgs_BS1_LoRA_vs_Fine_Tuning.jpg

  • 256 Images Fine Tuning vs LoRA at Batch Size 7 : 256_imgs_BS7_LoRA_vs_Fine_Tuning.jpg

  • 15 vs 256 Images vs Batch Size 1 vs 7 vs LoRA vs Fine Tuning : 15_vs_256_imgs_BS1_vs_BS7_LoRA_vs_Fine_Tuning_Style_Comparison.jpg

  • Full conclusions and tips are also shared : https://www.patreon.com/posts/112099700

  • Additionally, I have shared full training entire logs that you can see each checkpoint took time. I have shared best checkpoints, their step count and took time according to being either LoRA, Fine Tuning or Batch size 1 or 7 or 15 images or 256 images, so a very detailed article regarding completed.

  • Check the images to see all shared files in the post.

  • Furthermore, a very very detailed analysis having article written and all latest DreamBooth / Fine Tuning configs and LoRA configs are shared with Kohya GUI installers for both Windows, Runpod and Massed Compute.

  • Moreover, I have shared new 28 realism and 37 stylization testing prompts.

Current tutorials are as below:

I have done the following trainings and thoroughly analyzed and compared all:

  • Fine Tuning / DreamBooth: 15 Training Images & Batch Size is 1

  • Fine Tuning / DreamBooth: 15 Training Images & Batch Size is 7

  • Fine Tuning / DreamBooth: 256 Training Images & Batch Size is 1

  • Fine Tuning / DreamBooth: 256 Training Images & Batch Size is 7

  • LoRA : 15 Training Images & Batch Size is 1

  • LoRA : 15 Training Images & Batch Size is 7

  • LoRA : 256 Training Images & Batch Size is 1

  • LoRA : 256 Training Images & Batch Size is 7

  • For each batch size 1 vs 7, a unique new learning rate (LR) is researched and best one used

  • Then compared all these checkpoints against each other very carefully and very thoroughly, and shared all findings and analysis

  • Huge FLUX LoRA vs Fine Tuning / DreamBooth Experiments Completed, Moreover Batch Size 1 vs 7 Fully Tested as Well, Not Only for Realism But Also for Stylization : https://www.patreon.com/posts/112099700

Some Part of Research as Images

API Trace View

How I Cut 22.3 Seconds Off an API Call with Sentry

Struggling with slow API calls? Dan Mindru walks through how he used Sentry's new Trace View feature to shave off 22.3 seconds from an API call.

Get a practical walkthrough of how to identify bottlenecks, split tasks into multiple parallel tasks, identify slow AI model calls, and more.

Read more →

Top comments (0)

Billboard image

The Next Generation Developer Platform

Coherence is the first Platform-as-a-Service you can control. Unlike "black-box" platforms that are opinionated about the infra you can deploy, Coherence is powered by CNC, the open-source IaC framework, which offers limitless customization.

Learn more

👋 Kindness is contagious

Please leave a ❤️ or a friendly comment on this post if you found it helpful!

Okay