FLUX.1-dev Impressionism fine-tuning with LoRA

This is a LoRA fine-tuning of the FLUX.1 model trained on a curated dataset of impressionist paintings from WikiArt.

Training Process & Results

Training Environment

Training Progress Visualization

Training Progress Grid

Training Progress Grid 4x6 grid showing model progression across different prompts (rows) at various training steps (columns: 0, 200, 400, 600, 800, 1000)

Step-by-Step Evolution

Training Progress Animation Evolution of the model's output for the prompt: "An impressionist painting portrays a vast landscape with gently rolling hills under a radiant sky. Clusters of autumn trees dot the scene, rendered with loose, expressive brushstrokes and a palette of warm oranges, deep greens, and soft blues, creating a sense of tranquil, natural beauty" (Steps 0-1000, sampled every 100 steps)

Base vs Fine-tuned

Base model vs Fine-tuned Left side is the base model and right side is this fine-tuned model

Current Results & Future Improvements

The most notable improvements are observed in landscape generation, which can be attributed to:

  • Strong representation of landscapes (30%) in the training dataset
  • Inherent structural similarities in impressionist landscape paintings
  • Clear patterns in color usage and brushstroke techniques

Future improvements will focus on:

  • Experimenting with different LoRA configurations and ranks
  • Fine-tuning hyperparameters for better convergence
  • Improving caption quality and specificity(current captions may be too complex that model can not capture spesific features)
  • 'content_or_style' paramater on training config is currently set to 'balanced'. I also want to test 'style' parameter for model training.
  • Extending training duration beyond 1000 steps
  • Developing custom training scripts for more granular control

While the current implementation uses the AI-Toolkit, future iterations will involve developing custom training scripts to gain deeper insights into model configuration and behavior.

Dataset

The model was trained on the WikiArt Impressionism Curated Dataset, which contains 1,000 high-quality Impressionist paintings with the following distribution:

  • Landscapes: 300 images (30%)
  • Portraits: 300 images (30%)
  • Urban Scenes: 200 images (20%)
  • Still Life: 200 images (20%)

Model Details

  • Base Model: FLUX.1
  • LoRA Rank: 16
  • Training Steps: 1000
  • Resolution: 512-768-1024px

You can find detailed training configurations on config.yaml

Usage

To run code 4-bit with quantization check out this Google Colab Notebook.

On Google Colab the cheapest way to run code is acquiring a T4 with high-ram if I am not wrong :)

Also thanks to providers original notebook to run code 4-bit with quantization. Original Colab Notebook :

License

This model inherits the license of the base FLUX.1 model and the WikiArt dataset.

Downloads last month
63
Inference Providers NEW
Examples

Model tree for dolphinium/FLUX.1-dev-wikiart-impressionism

Adapter
(24885)
this model