dolphinium commited on
Commit
0d2cc9a
·
verified ·
1 Parent(s): 4030b56

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +57 -22
README.md CHANGED
@@ -10,12 +10,60 @@ tags:
10
  library_name: diffusers
11
  pipeline_tag: text-to-image
12
  base_model: black-forest-labs/FLUX.1-dev
 
 
 
 
 
 
 
 
13
  ---
14
 
15
  # FLUX.1-dev Impressionism fine-tuning with LoRA
16
 
17
  This is a LoRA fine-tuning of the FLUX.1 model trained on a curated dataset of impressionist paintings from WikiArt.
18
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  ## Dataset
20
  The model was trained on the [WikiArt Impressionism Curated Dataset](https://huggingface.co/datasets/dolphinium/wikiart-impressionism-curated), which contains 1,000 high-quality Impressionist paintings with the following distribution:
21
 
@@ -27,32 +75,19 @@ The model was trained on the [WikiArt Impressionism Curated Dataset](https://hug
27
  ## Model Details
28
  - Base Model: [FLUX.1](https://huggingface.co/black-forest-labs/FLUX.1-dev)
29
  - LoRA Rank: 16
30
- - Training Steps: 2000
31
- - Resolution: 512-1024px
32
 
33
- ## Usage
34
 
35
- ```python
36
- from diffusers import StableDiffusionPipeline
37
- import torch
38
-
39
- model_id = "black-forest-labs/FLUX.1-dev"
40
- lora_model_path = "dolphinium/FLUX.1-dev-wikiart-impressionism-v2"
41
-
42
- pipe = StableDiffusionPipeline.from_pretrained(
43
- model_id,
44
- torch_dtype=torch.float16
45
- ).to("cuda")
46
 
47
- # Load LoRA weights
48
- pipe.unet.load_attn_procs(lora_model_path)
49
 
50
- # Generate image
51
- prompt = "an impressionist style landscape with rolling hills and autumn trees"
52
- image = pipe(prompt).images[0]
53
- image.save("impressionist_landscape.png")
54
- ```
55
 
 
 
56
 
57
  ## License
58
- This model inherits the license of the base FLUX.1 model and the WikiArt dataset.
 
10
  library_name: diffusers
11
  pipeline_tag: text-to-image
12
  base_model: black-forest-labs/FLUX.1-dev
13
+ widget:
14
+ - text: >-
15
+ An impressionist painting portrays a vast landscape with gently rolling
16
+ hills under a radiant sky. Clusters of autumn trees dot the scene, rendered
17
+ with loose, expressive brushstrokes and a palette of warm oranges, deep
18
+ greens, and soft blues, creating a sense of tranquil, natural beauty
19
+ output:
20
+ url: images/example_jl6x0209w.png
21
  ---
22
 
23
  # FLUX.1-dev Impressionism fine-tuning with LoRA
24
 
25
  This is a LoRA fine-tuning of the FLUX.1 model trained on a curated dataset of impressionist paintings from WikiArt.
26
 
27
+ ## Training Process & Results
28
+
29
+ ### Training Environment
30
+ - GPU: NVIDIA A100
31
+ - Training Duration: ~1 hour for 1000 steps
32
+ - Training Notebook: [Google Colab Notebook](https://colab.research.google.com/drive/1G9k6iwSGKXmA32ok4zOPijFUFwBAZ9aB?usp=sharing)
33
+ - Training Framework: [AI-Toolkit](https://github.com/ostris/ai-toolkit)
34
+
35
+ ## Training Progress Visualization
36
+
37
+ ### Training Progress Grid
38
+ ![Training Progress Grid](sample_grid_annotated.png)
39
+ *4x6 grid showing model progression across different prompts (rows) at various training steps (columns: 0, 200, 400, 600, 800, 1000)*
40
+
41
+ ### Step-by-Step Evolution
42
+ ![Training Progress Animation](prompt_0.gif)
43
+ *Evolution of the model's output for the prompt: "An impressionist painting portrays a vast landscape with gently rolling hills under a radiant sky. Clusters of autumn trees dot the scene, rendered with loose, expressive brushstrokes and a palette of warm oranges, deep greens, and soft blues, creating a sense of tranquil, natural beauty" (Steps 0-1000, sampled every 100 steps)*
44
+
45
+
46
+ ### Base vs Fine-tuned
47
+ ![Base model vs Fine-tuned](base_vs_fine_tuned.png)
48
+ *Left side is the base model and right side is this fine-tuned model*
49
+
50
+
51
+ ### Current Results & Future Improvements
52
+ The most notable improvements are observed in landscape generation, which can be attributed to:
53
+ - Strong representation of landscapes (30%) in the training dataset
54
+ - Inherent structural similarities in impressionist landscape paintings
55
+ - Clear patterns in color usage and brushstroke techniques
56
+
57
+ Future improvements will focus on:
58
+ - Experimenting with different LoRA configurations and ranks
59
+ - Fine-tuning hyperparameters for better convergence
60
+ - Improving caption quality and specificity(current captions may be too complex that model can not capture spesific features)
61
+ - 'content_or_style' paramater on training config is currently set to 'balanced'. I also want to test 'style' parameter for model training.
62
+ - Extending training duration beyond 1000 steps
63
+ - Developing custom training scripts for more granular control
64
+
65
+ While the current implementation uses the [AI-Toolkit](https://github.com/ostris/ai-toolkit), future iterations will involve developing custom training scripts to gain deeper insights into model configuration and behavior.
66
+
67
  ## Dataset
68
  The model was trained on the [WikiArt Impressionism Curated Dataset](https://huggingface.co/datasets/dolphinium/wikiart-impressionism-curated), which contains 1,000 high-quality Impressionist paintings with the following distribution:
69
 
 
75
  ## Model Details
76
  - Base Model: [FLUX.1](https://huggingface.co/black-forest-labs/FLUX.1-dev)
77
  - LoRA Rank: 16
78
+ - Training Steps: 1000
79
+ - Resolution: 512-768-1024px
80
 
81
+ You can find detailed training configurations on [config.yaml](config.yaml)
82
 
83
+ ## Usage
 
 
 
 
 
 
 
 
 
 
84
 
85
+ To run code 4-bit with quantization check out this [Google Colab Notebook](https://colab.research.google.com/drive/1dnCeNGHQSuWACrG95rH4TXPgXwNNdTh-?usp=sharing).
 
86
 
87
+ On Google Colab the cheapest way to run code is acquiring a T4 with high-ram if I am not wrong :)
 
 
 
 
88
 
89
+ Also thanks to providers original notebook to run code 4-bit with quantization.
90
+ [Original Colab Notebook](https://github.com/NielsRogge/Transformers-Tutorials/blob/master/Flux/Run_Flux_on_an_8GB_machine.ipynb) :
91
 
92
  ## License
93
+ This model inherits the license of the base [FLUX.1 model](https://huggingface.co/black-forest-labs/FLUX.1-dev) and the [WikiArt](https://huggingface.co/datasets/huggan/wikiart) dataset.