idle Stoev
commited on
Commit
•
1727b12
1
Parent(s):
e82394e
Config Variation
Browse files- README.md +10 -0
- scheduler/scheduler_config.json +1 -4
- text_encoder/config.json +2 -2
- tokenizer/tokenizer_config.json +2 -1
- unet/config.json +2 -24
- vae/config.json +1 -2
README.md
CHANGED
@@ -9,6 +9,8 @@ tags:
|
|
9 |
- di.ffusion.ai
|
10 |
- art
|
11 |
- base model
|
|
|
|
|
12 |
library_name: diffusers
|
13 |
widget:
|
14 |
- text: >-
|
@@ -52,6 +54,14 @@ widget:
|
|
52 |
![ffusion-basesm32.jpg](https://cdn-uploads.huggingface.co/production/uploads/6380cf05f496d57325c12194/FHidip3rT8mL1UdGDyfSd.jpeg)
|
53 |
|
54 |
## Model Overview: Unleashing the Power of Imagination!
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
55 |
|
56 |
FFUSION AI is a state-of-the-art image generation and transformation tool, developed around the leading Latent Diffusion Model. Leveraging Stable Diffusion 2.1, FFUSION AI converts your prompts into captivating artworks. Discover an imaginative landscape where ideas come to life in vibrant, surreal visuals.
|
57 |
|
|
|
9 |
- di.ffusion.ai
|
10 |
- art
|
11 |
- base model
|
12 |
+
- diffusers
|
13 |
+
inference: true
|
14 |
library_name: diffusers
|
15 |
widget:
|
16 |
- text: >-
|
|
|
54 |
![ffusion-basesm32.jpg](https://cdn-uploads.huggingface.co/production/uploads/6380cf05f496d57325c12194/FHidip3rT8mL1UdGDyfSd.jpeg)
|
55 |
|
56 |
## Model Overview: Unleashing the Power of Imagination!
|
57 |
+
<div style="display: flex; flex-wrap: wrap; gap: 2px;">
|
58 |
+
<a href="https://huggingface.co/FFusion/"><img src="https://img.shields.io/badge/🧠%20Model%20Type-Diffusion--based%20text--to--image%20generation%20model-blueviolet" alt="Model Type"></a>
|
59 |
+
<a href="https://huggingface.co/FFusion/"><img src="https://img.shields.io/badge/🔏%20License-CreativeML%20Open%20RAIL++--M%20License-orange" alt="License"></a>
|
60 |
+
<a href="https://huggingface.co/FFusion/"><img src="https://img.shields.io/badge/🖥️%20Hardware%20Type-A100%20PCIe%2040GB-green" alt="Hardware Type"></a>
|
61 |
+
<a href="https://huggingface.co/FFusion/"><img src="https://img.shields.io/badge/⏰%20Hours%20Used-1190-red" alt="Hours Used"></a>
|
62 |
+
<a href="https://huggingface.co/FFusion/"><img src="https://img.shields.io/badge/🌩️%20Cloud%20Provider-CoreWeave%20%26%20Runpod-blue" alt="Cloud Provider"></a>
|
63 |
+
<a href="https://huggingface.co/FFusion/"><img src="https://img.shields.io/badge/🍃%20Carbon%20Emitted-124.95%20kg%20of%20CO2-brightgreen" alt="Carbon Emitted"></a>
|
64 |
+
</div>
|
65 |
|
66 |
FFUSION AI is a state-of-the-art image generation and transformation tool, developed around the leading Latent Diffusion Model. Leveraging Stable Diffusion 2.1, FFUSION AI converts your prompts into captivating artworks. Discover an imaginative landscape where ideas come to life in vibrant, surreal visuals.
|
67 |
|
scheduler/scheduler_config.json
CHANGED
@@ -9,11 +9,8 @@
|
|
9 |
"dynamic_thresholding_ratio": 0.995,
|
10 |
"num_train_timesteps": 1000,
|
11 |
"prediction_type": "v_prediction",
|
12 |
-
"rescale_betas_zero_snr": false,
|
13 |
-
"sample_max_value": 1.0,
|
14 |
"set_alpha_to_one": false,
|
15 |
"steps_offset": 1,
|
16 |
-
"
|
17 |
-
"timestep_spacing": "leading",
|
18 |
"trained_betas": null
|
19 |
}
|
|
|
9 |
"dynamic_thresholding_ratio": 0.995,
|
10 |
"num_train_timesteps": 1000,
|
11 |
"prediction_type": "v_prediction",
|
|
|
|
|
12 |
"set_alpha_to_one": false,
|
13 |
"steps_offset": 1,
|
14 |
+
"skip_prk_steps": true,
|
|
|
15 |
"trained_betas": null
|
16 |
}
|
text_encoder/config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
|
3 |
"architectures": [
|
4 |
"CLIPTextModel"
|
5 |
],
|
@@ -19,7 +19,7 @@
|
|
19 |
"num_hidden_layers": 23,
|
20 |
"pad_token_id": 1,
|
21 |
"projection_dim": 512,
|
22 |
-
"torch_dtype": "
|
23 |
"transformers_version": "4.31.0",
|
24 |
"vocab_size": 49408
|
25 |
}
|
|
|
1 |
{
|
2 |
+
|
3 |
"architectures": [
|
4 |
"CLIPTextModel"
|
5 |
],
|
|
|
19 |
"num_hidden_layers": 23,
|
20 |
"pad_token_id": 1,
|
21 |
"projection_dim": 512,
|
22 |
+
"torch_dtype": "float32",
|
23 |
"transformers_version": "4.31.0",
|
24 |
"vocab_size": 49408
|
25 |
}
|
tokenizer/tokenizer_config.json
CHANGED
@@ -8,7 +8,6 @@
|
|
8 |
"rstrip": false,
|
9 |
"single_word": false
|
10 |
},
|
11 |
-
"clean_up_tokenization_spaces": true,
|
12 |
"do_lower_case": true,
|
13 |
"eos_token": {
|
14 |
"__type": "AddedToken",
|
@@ -20,7 +19,9 @@
|
|
20 |
},
|
21 |
"errors": "replace",
|
22 |
"model_max_length": 77,
|
|
|
23 |
"pad_token": "<|endoftext|>",
|
|
|
24 |
"tokenizer_class": "CLIPTokenizer",
|
25 |
"unk_token": {
|
26 |
"__type": "AddedToken",
|
|
|
8 |
"rstrip": false,
|
9 |
"single_word": false
|
10 |
},
|
|
|
11 |
"do_lower_case": true,
|
12 |
"eos_token": {
|
13 |
"__type": "AddedToken",
|
|
|
19 |
},
|
20 |
"errors": "replace",
|
21 |
"model_max_length": 77,
|
22 |
+
"name_or_path": "stabilityai/stable-diffusion-2-1",
|
23 |
"pad_token": "<|endoftext|>",
|
24 |
+
"special_tokens_map_file": "./special_tokens_map.json",
|
25 |
"tokenizer_class": "CLIPTokenizer",
|
26 |
"unk_token": {
|
27 |
"__type": "AddedToken",
|
unet/config.json
CHANGED
@@ -2,9 +2,6 @@
|
|
2 |
"_class_name": "UNet2DConditionModel",
|
3 |
"_diffusers_version": "0.19.0.dev0",
|
4 |
"act_fn": "silu",
|
5 |
-
"addition_embed_type": null,
|
6 |
-
"addition_embed_type_num_heads": 64,
|
7 |
-
"addition_time_embed_dim": null,
|
8 |
"attention_head_dim": [
|
9 |
5,
|
10 |
10,
|
@@ -18,12 +15,7 @@
|
|
18 |
1280
|
19 |
],
|
20 |
"center_input_sample": false,
|
21 |
-
"class_embed_type": null,
|
22 |
-
"class_embeddings_concat": false,
|
23 |
-
"conv_in_kernel": 3,
|
24 |
-
"conv_out_kernel": 3,
|
25 |
"cross_attention_dim": 1024,
|
26 |
-
"cross_attention_norm": null,
|
27 |
"down_block_types": [
|
28 |
"CrossAttnDownBlock2D",
|
29 |
"CrossAttnDownBlock2D",
|
@@ -32,38 +24,24 @@
|
|
32 |
],
|
33 |
"downsample_padding": 1,
|
34 |
"dual_cross_attention": false,
|
35 |
-
"encoder_hid_dim": null,
|
36 |
-
"encoder_hid_dim_type": null,
|
37 |
"flip_sin_to_cos": true,
|
38 |
"freq_shift": 0,
|
39 |
"in_channels": 4,
|
40 |
"layers_per_block": 2,
|
41 |
-
"mid_block_only_cross_attention": null,
|
42 |
"mid_block_scale_factor": 1,
|
43 |
-
"mid_block_type": "UNetMidBlock2DCrossAttn",
|
44 |
"norm_eps": 1e-05,
|
45 |
"norm_num_groups": 32,
|
46 |
"num_attention_heads": null,
|
47 |
"num_class_embeds": null,
|
48 |
"only_cross_attention": false,
|
49 |
"out_channels": 4,
|
50 |
-
"
|
51 |
-
"resnet_out_scale_factor": 1.0,
|
52 |
-
"resnet_skip_time_act": false,
|
53 |
-
"resnet_time_scale_shift": "default",
|
54 |
-
"sample_size": 96,
|
55 |
-
"time_cond_proj_dim": null,
|
56 |
-
"time_embedding_act_fn": null,
|
57 |
-
"time_embedding_dim": null,
|
58 |
-
"time_embedding_type": "positional",
|
59 |
-
"timestep_post_act": null,
|
60 |
-
"transformer_layers_per_block": 1,
|
61 |
"up_block_types": [
|
62 |
"UpBlock2D",
|
63 |
"CrossAttnUpBlock2D",
|
64 |
"CrossAttnUpBlock2D",
|
65 |
"CrossAttnUpBlock2D"
|
66 |
],
|
67 |
-
"upcast_attention":
|
68 |
"use_linear_projection": true
|
69 |
}
|
|
|
2 |
"_class_name": "UNet2DConditionModel",
|
3 |
"_diffusers_version": "0.19.0.dev0",
|
4 |
"act_fn": "silu",
|
|
|
|
|
|
|
5 |
"attention_head_dim": [
|
6 |
5,
|
7 |
10,
|
|
|
15 |
1280
|
16 |
],
|
17 |
"center_input_sample": false,
|
|
|
|
|
|
|
|
|
18 |
"cross_attention_dim": 1024,
|
|
|
19 |
"down_block_types": [
|
20 |
"CrossAttnDownBlock2D",
|
21 |
"CrossAttnDownBlock2D",
|
|
|
24 |
],
|
25 |
"downsample_padding": 1,
|
26 |
"dual_cross_attention": false,
|
|
|
|
|
27 |
"flip_sin_to_cos": true,
|
28 |
"freq_shift": 0,
|
29 |
"in_channels": 4,
|
30 |
"layers_per_block": 2,
|
|
|
31 |
"mid_block_scale_factor": 1,
|
|
|
32 |
"norm_eps": 1e-05,
|
33 |
"norm_num_groups": 32,
|
34 |
"num_attention_heads": null,
|
35 |
"num_class_embeds": null,
|
36 |
"only_cross_attention": false,
|
37 |
"out_channels": 4,
|
38 |
+
"sample_size": 64,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
39 |
"up_block_types": [
|
40 |
"UpBlock2D",
|
41 |
"CrossAttnUpBlock2D",
|
42 |
"CrossAttnUpBlock2D",
|
43 |
"CrossAttnUpBlock2D"
|
44 |
],
|
45 |
+
"upcast_attention": false,
|
46 |
"use_linear_projection": true
|
47 |
}
|
vae/config.json
CHANGED
@@ -20,8 +20,7 @@
|
|
20 |
"layers_per_block": 2,
|
21 |
"norm_num_groups": 32,
|
22 |
"out_channels": 3,
|
23 |
-
"sample_size":
|
24 |
-
"scaling_factor": 0.18215,
|
25 |
"up_block_types": [
|
26 |
"UpDecoderBlock2D",
|
27 |
"UpDecoderBlock2D",
|
|
|
20 |
"layers_per_block": 2,
|
21 |
"norm_num_groups": 32,
|
22 |
"out_channels": 3,
|
23 |
+
"sample_size": 256,
|
|
|
24 |
"up_block_types": [
|
25 |
"UpDecoderBlock2D",
|
26 |
"UpDecoderBlock2D",
|