File size: 2,774 Bytes
0cfed93
 
 
 
 
 
 
 
 
 
86ac7d9
0cfed93
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
630e419
0cfed93
 
 
 
 
 
9712020
00525ec
9712020
00525ec
 
0cfed93
 
9712020
 
0cfed93
 
 
 
7a1608f
0cfed93
 
 
 
ed91aa0
5f5853f
0cfed93
 
ed91aa0
86ac7d9
0cfed93
 
ed91aa0
0cfed93
 
 
 
 
9712020
 
0cfed93
 
 
9712020
 
0cfed93
 
 
 
 
ed91aa0
 
0cfed93
 
 
 
 
 
 
 
 
 
 
c187e3d
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
{
    "prior": {
        "clip": {
            "make": "openai",
            "model": "ViT-L/14"
        },
        "net": {
            "dim": 768,
            "depth": 12,
            "num_timesteps": 1000,
            "max_text_len": 77,
            "num_time_embeds": 1,
            "num_image_embeds": 1,
            "num_text_embeds": 1,
            "dim_head": 64,
            "heads": 12,
            "ff_mult": 4,
            "norm_out": true,
            "attn_dropout": 0.05,
            "ff_dropout": 0.05,
            "final_proj": true,
            "normformer": true,
            "rotary_emb": true
        },
        "image_embed_dim": 768,
        "image_size": 224,
        "image_channels": 3,
        "timesteps": 1000,
        "sample_timesteps": 64,
        "cond_drop_prob": 0.1,
        "loss_type": "l2",
        "predict_x_start": true,
        "beta_schedule": "cosine",
        "condition_on_text_encodings": true
    },
    "data": {
        "batch_size": 225,
        "num_data_points": 500000000,
        "eval_every_seconds": 1600,
        "image_url": "s3://s-laion5b/embeddings/vit-l-14/laion2B-en/img_emb/",
        "meta_url": "s3://s-laion5b/embeddings/vit-l-14/laion2B-en/metadata/",
        "splits": {
            "train": 0.9,
            "val":  1.2e-5,
            "test": 0.09998799999999998
        }
    },
    "train": {
        "epochs": 5,
        "lr": 0.5e-5,
        "wd": 6.02e-2,
        "max_grad_norm": 0.5,
        "use_ema": true,
        "ema_beta": 0.9999,
        "ema_update_after_step": 50,
        "warmup_steps": 50,
        "amp": false,
        "save_every_seconds": 3600,
        "eval_timesteps": [64, 1000],
        "random_seed": 84513
    },
    "tracker": {
        "data_path": ".prior",
        "overwrite_data_path": true,
        "log": {
            "log_type": "wandb",
            "wandb_entity": "nousr_laion",
            "wandb_project": "dalle2_diffusion_prior",
            "wandb_resume": true,
            "wandb_run_id": "20jdn8i2",
            "verbose": true
        },
        "load": {
            "load_from": "local",
            "file_path": "/fsx/nousr/.prior/latest_checkpoint.pth"
        },
        "save": [
            {
                "save_to": "local",
                "save_type": "checkpoint",
                "save_latest_to": ".prior/latest_checkpoint.pth",
                "save_best_to": ".prior/best_checkpoint.pth"
            },
            {
                "save_to": "huggingface",
                "huggingface_repo": "laion/DALLE2-PyTorch",
                "save_meta_to": "prior/",
                "save_latest_to": "prior/latest.pth",
                "save_best_to": "prior/best.pth",
                "save_type": "model"
            }
        ]
    }
}