Humanlearning commited on
Commit
ec0a228
1 Parent(s): 1461149

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -2,37 +2,4 @@
2
  license: other
3
  license_name: flux-1-dev-non-commercial-license
4
  license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md
5
- language:
6
- - en
7
- tags:
8
- - flux
9
- - diffusers
10
- - lora
11
- base_model: "black-forest-labs/FLUX.1-dev"
12
- pipeline_tag: text-to-image
13
- instance_prompt: TOK
14
- ---
15
-
16
- # Flux Lora Me
17
-
18
- Trained on Replicate using:
19
-
20
- https://replicate.com/ostris/flux-dev-lora-trainer/train
21
-
22
-
23
- ## Trigger words
24
- You should use `TOK` to trigger the image generation.
25
-
26
-
27
- ## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers)
28
-
29
- ```py
30
- from diffusers import AutoPipelineForText2Image
31
- import torch
32
-
33
- pipeline = AutoPipelineForText2Image.from_pretrained('black-forest-labs/FLUX.1-dev', torch_dtype=torch.float16).to('cuda')
34
- pipeline.load_lora_weights('Humanlearning/flux-lora-me', weight_name='lora.safetensors')
35
- image = pipeline('your prompt').images[0]
36
- ```
37
-
38
- For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters)
 
2
  license: other
3
  license_name: flux-1-dev-non-commercial-license
4
  license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md
5
+ ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
config.yaml CHANGED
@@ -1,59 +1,60 @@
1
- job: custom_job
2
  config:
3
  name: flux_train_replicate
4
  process:
5
- - type: custom_sd_trainer
6
- training_folder: output
7
- device: cuda:0
8
- trigger_word: TOK
9
- network:
10
- type: lora
11
- linear: 16
12
- linear_alpha: 16
13
- save:
14
- dtype: float16
15
- save_every: 1501
16
- max_step_saves_to_keep: 1
17
- datasets:
18
- - folder_path: input_images
19
- caption_ext: txt
20
  caption_dropout_rate: 0.05
21
- shuffle_tokens: false
22
- cache_latents_to_disk: true
23
  resolution:
24
  - 512
25
  - 768
26
  - 1024
27
- train:
28
- batch_size: 1
29
- steps: 1500
30
- gradient_accumulation_steps: 1
31
- train_unet: true
32
- train_text_encoder: false
33
- content_or_style: balanced
34
- gradient_checkpointing: true
35
- noise_scheduler: flowmatch
36
- optimizer: adamw8bit
37
- lr: 0.0004
38
- ema_config:
39
- use_ema: true
40
- ema_decay: 0.99
41
- dtype: bf16
42
  model:
43
- name_or_path: FLUX.1-dev
44
  is_flux: true
 
45
  quantize: true
 
 
 
 
46
  sample:
47
- sampler: flowmatch
48
- sample_every: 1501
49
- width: 1024
50
  height: 1024
51
- prompts: []
52
  neg: ''
 
 
 
 
 
53
  seed: 42
54
  walk_seed: true
55
- guidance_scale: 4
56
- sample_steps: 20
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
57
  meta:
58
  name: flux_train_replicate
59
  version: '1.0'
 
 
1
  config:
2
  name: flux_train_replicate
3
  process:
4
+ - datasets:
5
+ - cache_latents_to_disk: true
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  caption_dropout_rate: 0.05
7
+ caption_ext: txt
8
+ folder_path: input_images
9
  resolution:
10
  - 512
11
  - 768
12
  - 1024
13
+ shuffle_tokens: false
14
+ device: cuda:0
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  model:
 
16
  is_flux: true
17
+ name_or_path: black-forest-labs/FLUX.1-dev
18
  quantize: true
19
+ network:
20
+ linear: 16
21
+ linear_alpha: 16
22
+ type: lora
23
  sample:
24
+ guidance_scale: 4
 
 
25
  height: 1024
 
26
  neg: ''
27
+ prompts:
28
+ - a sign that says 'I LOVE PROMPTS!' in the style of [trigger]
29
+ sample_every: 250
30
+ sample_steps: 20
31
+ sampler: flowmatch
32
  seed: 42
33
  walk_seed: true
34
+ width: 1024
35
+ save:
36
+ dtype: float16
37
+ max_step_saves_to_keep: 1
38
+ save_every: 1501
39
+ train:
40
+ batch_size: 1
41
+ content_or_style: balanced
42
+ dtype: bf16
43
+ ema_config:
44
+ ema_decay: 0.99
45
+ use_ema: true
46
+ gradient_accumulation_steps: 1
47
+ gradient_checkpointing: true
48
+ lr: 0.0004
49
+ noise_scheduler: flowmatch
50
+ optimizer: adamw8bit
51
+ steps: 1500
52
+ train_text_encoder: false
53
+ train_unet: true
54
+ training_folder: output
55
+ trigger_word: TOK
56
+ type: sd_trainer
57
+ job: extension
58
  meta:
59
  name: flux_train_replicate
60
  version: '1.0'
flux_train_replicate.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db3a16bb00f7200a8258007e0d87bf2e7a862c632b7e5f9212470ebe6c08e826
3
+ size 171969416
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57dbc887b63ed12ef225835513009838dfdda301a8ee43eb3aedb63d7e428724
3
+ size 173272836
samples/1724257435643__000000000_0.jpg ADDED
samples/1724257956439__000000250_0.jpg ADDED
samples/1724258475590__000000500_0.jpg ADDED
samples/1724258992120__000000750_0.jpg ADDED
samples/1724259509076__000001000_0.jpg ADDED
samples/1724260028706__000001250_0.jpg ADDED
samples/1724260548796__000001500_0.jpg ADDED