Upload 11 files
Browse files- 32k-ookura aeon/D_12000.pth +3 -0
- 32k-ookura aeon/D_18000.pth +3 -0
- 32k-ookura aeon/D_24000.pth +3 -0
- 32k-ookura aeon/D_6000.pth +3 -0
- 32k-ookura aeon/G_12000.pth +3 -0
- 32k-ookura aeon/G_18000.pth +3 -0
- 32k-ookura aeon/G_24000.pth +3 -0
- 32k-ookura aeon/G_6000.pth +3 -0
- 32k-ookura aeon/config.json +90 -0
- 32k-ookura aeon/eval/events.out.tfevents.1675947309.DESKTOP-P582Q00.36808.1 +3 -0
- 32k-ookura aeon/train.log +612 -0
32k-ookura aeon/D_12000.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4fb1a2f4bc3f7d7805e1cfb2f025f934179a969ab31189d7b92b0a2a7cd3eeeb
|
3 |
+
size 561098185
|
32k-ookura aeon/D_18000.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:94fd1b3cee44a0a172dbec9f73b6d3843fcc5b451ff4b4ce94fe9367ff7dd2dd
|
3 |
+
size 561098185
|
32k-ookura aeon/D_24000.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:138f7ca1cb84b9ec64c407b8a0c7e015927aca2b8161dc87491f72d67f0949c9
|
3 |
+
size 561098185
|
32k-ookura aeon/D_6000.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:21460785b6945cc96bdbae223830a00d00630d888ed9e44c32e21ecd08b4e714
|
3 |
+
size 561098185
|
32k-ookura aeon/G_12000.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a96b59146242180f2ed5731a3de2c11ce7a8700d9a2e08c9cafcfed79252ad37
|
3 |
+
size 699505437
|
32k-ookura aeon/G_18000.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:62cd3b69f265941c2ecc833f2b668dd10daf08613187d9c8688e7f6032a3116b
|
3 |
+
size 699505437
|
32k-ookura aeon/G_24000.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4214d3fa11861441b5df25b85a5afdcac3c31b4284ea9b1afb42a84111338b41
|
3 |
+
size 699505437
|
32k-ookura aeon/G_6000.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f0878abe02ff01b91f7c8211d4a161907c49dcc046aadf3be5d8190e586149af
|
3 |
+
size 699505437
|
32k-ookura aeon/config.json
ADDED
@@ -0,0 +1,90 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"train": {
|
3 |
+
"log_interval": 200,
|
4 |
+
"eval_interval": 1000,
|
5 |
+
"seed": 1234,
|
6 |
+
"epochs": 10000,
|
7 |
+
"learning_rate": 0.0001,
|
8 |
+
"betas": [
|
9 |
+
0.8,
|
10 |
+
0.99
|
11 |
+
],
|
12 |
+
"eps": 1e-09,
|
13 |
+
"batch_size": 6,
|
14 |
+
"fp16_run": false,
|
15 |
+
"lr_decay": 0.999875,
|
16 |
+
"segment_size": 17920,
|
17 |
+
"init_lr_ratio": 1,
|
18 |
+
"warmup_epochs": 0,
|
19 |
+
"c_mel": 45,
|
20 |
+
"c_kl": 1.0,
|
21 |
+
"use_sr": true,
|
22 |
+
"max_speclen": 384,
|
23 |
+
"port": "8001"
|
24 |
+
},
|
25 |
+
"data": {
|
26 |
+
"training_files": "filelists/train.txt",
|
27 |
+
"validation_files": "filelists/val.txt",
|
28 |
+
"max_wav_value": 32768.0,
|
29 |
+
"sampling_rate": 32000,
|
30 |
+
"filter_length": 1280,
|
31 |
+
"hop_length": 320,
|
32 |
+
"win_length": 1280,
|
33 |
+
"n_mel_channels": 80,
|
34 |
+
"mel_fmin": 0.0,
|
35 |
+
"mel_fmax": null
|
36 |
+
},
|
37 |
+
"model": {
|
38 |
+
"inter_channels": 192,
|
39 |
+
"hidden_channels": 192,
|
40 |
+
"filter_channels": 768,
|
41 |
+
"n_heads": 2,
|
42 |
+
"n_layers": 6,
|
43 |
+
"kernel_size": 3,
|
44 |
+
"p_dropout": 0.1,
|
45 |
+
"resblock": "1",
|
46 |
+
"resblock_kernel_sizes": [
|
47 |
+
3,
|
48 |
+
7,
|
49 |
+
11
|
50 |
+
],
|
51 |
+
"resblock_dilation_sizes": [
|
52 |
+
[
|
53 |
+
1,
|
54 |
+
3,
|
55 |
+
5
|
56 |
+
],
|
57 |
+
[
|
58 |
+
1,
|
59 |
+
3,
|
60 |
+
5
|
61 |
+
],
|
62 |
+
[
|
63 |
+
1,
|
64 |
+
3,
|
65 |
+
5
|
66 |
+
]
|
67 |
+
],
|
68 |
+
"upsample_rates": [
|
69 |
+
10,
|
70 |
+
8,
|
71 |
+
2,
|
72 |
+
2
|
73 |
+
],
|
74 |
+
"upsample_initial_channel": 512,
|
75 |
+
"upsample_kernel_sizes": [
|
76 |
+
16,
|
77 |
+
16,
|
78 |
+
4,
|
79 |
+
4
|
80 |
+
],
|
81 |
+
"n_layers_q": 3,
|
82 |
+
"use_spectral_norm": false,
|
83 |
+
"gin_channels": 256,
|
84 |
+
"ssl_dim": 256,
|
85 |
+
"n_speakers": 2
|
86 |
+
},
|
87 |
+
"spk": {
|
88 |
+
"aeon": 0
|
89 |
+
}
|
90 |
+
}
|
32k-ookura aeon/eval/events.out.tfevents.1675947309.DESKTOP-P582Q00.36808.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:58d5947e7524adecdb198323d53d55751ec96d0c0bbda5343b0d39106ac0525c
|
3 |
+
size 65057694
|
32k-ookura aeon/train.log
ADDED
@@ -0,0 +1,612 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
2023-02-09 23:54:56,860 32k INFO {'train': {'log_interval': 200, 'eval_interval': 1000, 'seed': 1234, 'epochs': 10000, 'learning_rate': 0.0001, 'betas': [0.8, 0.99], 'eps': 1e-09, 'batch_size': 6, 'fp16_run': False, 'lr_decay': 0.999875, 'segment_size': 17920, 'init_lr_ratio': 1, 'warmup_epochs': 0, 'c_mel': 45, 'c_kl': 1.0, 'use_sr': True, 'max_speclen': 384, 'port': '8001'}, 'data': {'training_files': 'filelists/train.txt', 'validation_files': 'filelists/val.txt', 'max_wav_value': 32768.0, 'sampling_rate': 32000, 'filter_length': 1280, 'hop_length': 320, 'win_length': 1280, 'n_mel_channels': 80, 'mel_fmin': 0.0, 'mel_fmax': None}, 'model': {'inter_channels': 192, 'hidden_channels': 192, 'filter_channels': 768, 'n_heads': 2, 'n_layers': 6, 'kernel_size': 3, 'p_dropout': 0.1, 'resblock': '1', 'resblock_kernel_sizes': [3, 7, 11], 'resblock_dilation_sizes': [[1, 3, 5], [1, 3, 5], [1, 3, 5]], 'upsample_rates': [10, 8, 2, 2], 'upsample_initial_channel': 512, 'upsample_kernel_sizes': [16, 16, 4, 4], 'n_layers_q': 3, 'use_spectral_norm': False, 'gin_channels': 256, 'ssl_dim': 256, 'n_speakers': 2}, 'spk': {'aeon': 0}, 'model_dir': './logs\\32k'}
|
2 |
+
2023-02-09 23:55:16,700 32k INFO Loaded checkpoint './logs\32k\G_0.pth' (iteration 1)
|
3 |
+
2023-02-09 23:55:20,130 32k INFO Loaded checkpoint './logs\32k\D_0.pth' (iteration 1)
|
4 |
+
2023-02-09 23:55:47,616 32k INFO Train Epoch: 1 [0%]
|
5 |
+
2023-02-09 23:55:47,617 32k INFO [2.4760148525238037, 2.3857452869415283, 12.13435173034668, 38.74782943725586, 12.577811241149902, 0, 0.0001]
|
6 |
+
2023-02-09 23:55:53,842 32k INFO Saving model and optimizer state at iteration 1 to ./logs\32k\G_0.pth
|
7 |
+
2023-02-09 23:56:10,616 32k INFO Saving model and optimizer state at iteration 1 to ./logs\32k\D_0.pth
|
8 |
+
2023-02-09 23:57:17,312 32k INFO ====> Epoch: 1
|
9 |
+
2023-02-09 23:58:41,881 32k INFO ====> Epoch: 2
|
10 |
+
2023-02-09 23:59:43,565 32k INFO Train Epoch: 3 [63%]
|
11 |
+
2023-02-09 23:59:43,565 32k INFO [3.0033481121063232, 2.0346643924713135, 9.557120323181152, 17.325883865356445, 1.1444225311279297, 200, 9.99750015625e-05]
|
12 |
+
2023-02-10 00:00:06,584 32k INFO ====> Epoch: 3
|
13 |
+
2023-02-10 00:01:30,796 32k INFO ====> Epoch: 4
|
14 |
+
2023-02-10 00:02:55,178 32k INFO ====> Epoch: 5
|
15 |
+
2023-02-10 00:03:33,965 32k INFO Train Epoch: 6 [26%]
|
16 |
+
2023-02-10 00:03:33,965 32k INFO [2.6587367057800293, 2.2601664066314697, 11.51845932006836, 20.874536514282227, 1.2725728750228882, 400, 9.993751562304699e-05]
|
17 |
+
2023-02-10 00:04:19,925 32k INFO ====> Epoch: 6
|
18 |
+
2023-02-10 00:05:44,080 32k INFO ====> Epoch: 7
|
19 |
+
2023-02-10 00:07:02,071 32k INFO Train Epoch: 8 [89%]
|
20 |
+
2023-02-10 00:07:02,071 32k INFO [2.6835384368896484, 2.130737543106079, 6.842741966247559, 16.306140899658203, 1.2471444606781006, 600, 9.991253280566489e-05]
|
21 |
+
2023-02-10 00:07:08,670 32k INFO ====> Epoch: 8
|
22 |
+
2023-02-10 00:08:32,953 32k INFO ====> Epoch: 9
|
23 |
+
2023-02-10 00:09:58,155 32k INFO ====> Epoch: 10
|
24 |
+
2023-02-10 00:10:53,121 32k INFO Train Epoch: 11 [53%]
|
25 |
+
2023-02-10 00:10:53,121 32k INFO [2.508908271789551, 2.226515769958496, 10.416086196899414, 16.404212951660156, 0.9339253306388855, 800, 9.987507028906759e-05]
|
26 |
+
2023-02-10 00:11:22,660 32k INFO ====> Epoch: 11
|
27 |
+
2023-02-10 00:12:46,946 32k INFO ====> Epoch: 12
|
28 |
+
2023-02-10 00:14:11,203 32k INFO ====> Epoch: 13
|
29 |
+
2023-02-10 00:14:43,324 32k INFO Train Epoch: 14 [16%]
|
30 |
+
2023-02-10 00:14:43,324 32k INFO [2.2750678062438965, 2.470716714859009, 11.370126724243164, 16.513538360595703, 1.350426197052002, 1000, 9.983762181915804e-05]
|
31 |
+
2023-02-10 00:14:47,828 32k INFO Saving model and optimizer state at iteration 14 to ./logs\32k\G_1000.pth
|
32 |
+
2023-02-10 00:15:06,487 32k INFO Saving model and optimizer state at iteration 14 to ./logs\32k\D_1000.pth
|
33 |
+
2023-02-10 00:16:02,684 32k INFO ====> Epoch: 14
|
34 |
+
2023-02-10 00:17:27,130 32k INFO ====> Epoch: 15
|
35 |
+
2023-02-10 00:18:38,765 32k INFO Train Epoch: 16 [79%]
|
36 |
+
2023-02-10 00:18:38,766 32k INFO [2.3250410556793213, 2.5265045166015625, 12.642312049865723, 18.580968856811523, 1.06544029712677, 1200, 9.981266397366609e-05]
|
37 |
+
2023-02-10 00:18:51,825 32k INFO ====> Epoch: 16
|
38 |
+
2023-02-10 00:20:16,238 32k INFO ====> Epoch: 17
|
39 |
+
2023-02-10 00:21:40,479 32k INFO ====> Epoch: 18
|
40 |
+
2023-02-10 00:22:29,060 32k INFO Train Epoch: 19 [42%]
|
41 |
+
2023-02-10 00:22:29,061 32k INFO [2.718174457550049, 2.224525213241577, 9.797636985778809, 13.11706256866455, 0.9313941597938538, 1400, 9.977523890319963e-05]
|
42 |
+
2023-02-10 00:23:05,152 32k INFO ====> Epoch: 19
|
43 |
+
2023-02-10 00:24:29,510 32k INFO ====> Epoch: 20
|
44 |
+
2023-02-10 00:25:53,829 32k INFO ====> Epoch: 21
|
45 |
+
2023-02-10 00:26:19,301 32k INFO Train Epoch: 22 [5%]
|
46 |
+
2023-02-10 00:26:19,301 32k INFO [2.4931561946868896, 2.2891364097595215, 5.325507164001465, 14.156054496765137, 0.7020604610443115, 1600, 9.973782786538036e-05]
|
47 |
+
2023-02-10 00:27:18,459 32k INFO ====> Epoch: 22
|
48 |
+
2023-02-10 00:28:42,594 32k INFO ====> Epoch: 23
|
49 |
+
2023-02-10 00:29:47,519 32k INFO Train Epoch: 24 [68%]
|
50 |
+
2023-02-10 00:29:47,519 32k INFO [2.5088696479797363, 2.6265249252319336, 7.155318737030029, 11.276984214782715, 0.6565865278244019, 1800, 9.971289496681757e-05]
|
51 |
+
2023-02-10 00:30:07,171 32k INFO ====> Epoch: 24
|
52 |
+
2023-02-10 00:31:31,461 32k INFO ====> Epoch: 25
|
53 |
+
2023-02-10 00:32:55,740 32k INFO ====> Epoch: 26
|
54 |
+
2023-02-10 00:33:37,761 32k INFO Train Epoch: 27 [32%]
|
55 |
+
2023-02-10 00:33:37,761 32k INFO [2.3180291652679443, 2.524212598800659, 11.95003890991211, 18.545923233032227, 0.6431781053543091, 2000, 9.967550730505221e-05]
|
56 |
+
2023-02-10 00:33:42,280 32k INFO Saving model and optimizer state at iteration 27 to ./logs\32k\G_2000.pth
|
57 |
+
2023-02-10 00:33:59,554 32k INFO Saving model and optimizer state at iteration 27 to ./logs\32k\D_2000.pth
|
58 |
+
2023-02-10 00:34:45,750 32k INFO ====> Epoch: 27
|
59 |
+
2023-02-10 00:36:10,183 32k INFO ====> Epoch: 28
|
60 |
+
2023-02-10 00:37:32,059 32k INFO Train Epoch: 29 [95%]
|
61 |
+
2023-02-10 00:37:32,059 32k INFO [2.533724784851074, 2.297114133834839, 9.978306770324707, 15.051481246948242, 0.5844590067863464, 2200, 9.965058998565574e-05]
|
62 |
+
2023-02-10 00:37:35,283 32k INFO ====> Epoch: 29
|
63 |
+
2023-02-10 00:39:00,123 32k INFO ====> Epoch: 30
|
64 |
+
2023-02-10 00:40:24,839 32k INFO ====> Epoch: 31
|
65 |
+
2023-02-10 00:41:23,329 32k INFO Train Epoch: 32 [58%]
|
66 |
+
2023-02-10 00:41:23,329 32k INFO [2.198890447616577, 2.618730068206787, 10.582822799682617, 15.83022689819336, 0.8572500348091125, 2400, 9.961322568533789e-05]
|
67 |
+
2023-02-10 00:41:49,677 32k INFO ====> Epoch: 32
|
68 |
+
2023-02-10 00:43:14,179 32k INFO ====> Epoch: 33
|
69 |
+
2023-02-10 00:44:38,626 32k INFO ====> Epoch: 34
|
70 |
+
2023-02-10 00:45:14,097 32k INFO Train Epoch: 35 [21%]
|
71 |
+
2023-02-10 00:45:14,098 32k INFO [2.8067591190338135, 2.2793397903442383, 7.9753804206848145, 11.641331672668457, 1.0224435329437256, 2600, 9.957587539488128e-05]
|
72 |
+
2023-02-10 00:46:03,586 32k INFO ====> Epoch: 35
|
73 |
+
2023-02-10 00:47:27,999 32k INFO ====> Epoch: 36
|
74 |
+
2023-02-10 00:48:42,929 32k INFO Train Epoch: 37 [84%]
|
75 |
+
2023-02-10 00:48:42,930 32k INFO [2.4459195137023926, 2.123002767562866, 10.907198905944824, 16.264312744140625, 0.6869288682937622, 2800, 9.95509829819056e-05]
|
76 |
+
2023-02-10 00:48:52,746 32k INFO ====> Epoch: 37
|
77 |
+
2023-02-10 00:50:17,223 32k INFO ====> Epoch: 38
|
78 |
+
2023-02-10 00:51:41,864 32k INFO ====> Epoch: 39
|
79 |
+
2023-02-10 00:52:33,774 32k INFO Train Epoch: 40 [47%]
|
80 |
+
2023-02-10 00:52:33,774 32k INFO [2.385266065597534, 2.274216651916504, 10.997411727905273, 16.773902893066406, 0.8327657580375671, 3000, 9.951365602954526e-05]
|
81 |
+
2023-02-10 00:52:38,348 32k INFO Saving model and optimizer state at iteration 40 to ./logs\32k\G_3000.pth
|
82 |
+
2023-02-10 00:52:57,291 32k INFO Saving model and optimizer state at iteration 40 to ./logs\32k\D_3000.pth
|
83 |
+
2023-02-10 00:53:33,533 32k INFO ====> Epoch: 40
|
84 |
+
2023-02-10 00:54:58,115 32k INFO ====> Epoch: 41
|
85 |
+
2023-02-10 00:56:22,645 32k INFO ====> Epoch: 42
|
86 |
+
2023-02-10 00:56:51,489 32k INFO Train Epoch: 43 [11%]
|
87 |
+
2023-02-10 00:56:51,490 32k INFO [2.287508249282837, 2.692063331604004, 9.70651626586914, 16.967220306396484, 0.798807680606842, 3200, 9.947634307304244e-05]
|
88 |
+
2023-02-10 00:57:47,480 32k INFO ====> Epoch: 43
|
89 |
+
2023-02-10 00:59:12,119 32k INFO ====> Epoch: 44
|
90 |
+
2023-02-10 01:00:20,508 32k INFO Train Epoch: 45 [74%]
|
91 |
+
2023-02-10 01:00:20,509 32k INFO [2.249314785003662, 2.683621883392334, 9.802401542663574, 17.789817810058594, 0.4752405881881714, 3400, 9.945147554159202e-05]
|
92 |
+
2023-02-10 01:00:36,926 32k INFO ====> Epoch: 45
|
93 |
+
2023-02-10 01:02:01,330 32k INFO ====> Epoch: 46
|
94 |
+
2023-02-10 01:03:25,782 32k INFO ====> Epoch: 47
|
95 |
+
2023-02-10 01:04:11,005 32k INFO Train Epoch: 48 [37%]
|
96 |
+
2023-02-10 01:04:11,006 32k INFO [2.2856054306030273, 2.5306761264801025, 9.405101776123047, 14.461358070373535, 0.6637941002845764, 3600, 9.941418589985758e-05]
|
97 |
+
2023-02-10 01:04:50,618 32k INFO ====> Epoch: 48
|
98 |
+
2023-02-10 01:06:16,153 32k INFO ====> Epoch: 49
|
99 |
+
2023-02-10 01:07:40,668 32k INFO ====> Epoch: 50
|
100 |
+
2023-02-10 01:08:02,827 32k INFO Train Epoch: 51 [0%]
|
101 |
+
2023-02-10 01:08:02,827 32k INFO [2.449678659439087, 2.3578238487243652, 10.553171157836914, 16.373308181762695, 1.1569361686706543, 3800, 9.937691023999092e-05]
|
102 |
+
2023-02-10 01:09:05,476 32k INFO ====> Epoch: 51
|
103 |
+
2023-02-10 01:10:29,887 32k INFO ====> Epoch: 52
|
104 |
+
2023-02-10 01:11:31,703 32k INFO Train Epoch: 53 [63%]
|
105 |
+
2023-02-10 01:11:31,703 32k INFO [2.4566311836242676, 2.3078484535217285, 9.267095565795898, 14.384586334228516, 0.9955094456672668, 4000, 9.935206756519513e-05]
|
106 |
+
2023-02-10 01:11:36,151 32k INFO Saving model and optimizer state at iteration 53 to ./logs\32k\G_4000.pth
|
107 |
+
2023-02-10 01:11:53,677 32k INFO Saving model and optimizer state at iteration 53 to ./logs\32k\D_4000.pth
|
108 |
+
2023-02-10 01:12:20,483 32k INFO ====> Epoch: 53
|
109 |
+
2023-02-10 01:13:45,174 32k INFO ====> Epoch: 54
|
110 |
+
2023-02-10 01:15:09,860 32k INFO ====> Epoch: 55
|
111 |
+
2023-02-10 01:15:48,458 32k INFO Train Epoch: 56 [26%]
|
112 |
+
2023-02-10 01:15:48,458 32k INFO [2.2449021339416504, 2.2987923622131348, 10.592050552368164, 14.990753173828125, 0.2585275173187256, 4200, 9.931481519679228e-05]
|
113 |
+
2023-02-10 01:16:34,690 32k INFO ====> Epoch: 56
|
114 |
+
2023-02-10 01:17:59,199 32k INFO ====> Epoch: 57
|
115 |
+
2023-02-10 01:19:17,517 32k INFO Train Epoch: 58 [89%]
|
116 |
+
2023-02-10 01:19:17,518 32k INFO [2.7950096130371094, 2.033355236053467, 6.349589824676514, 13.051651000976562, 0.41371816396713257, 4400, 9.928998804478705e-05]
|
117 |
+
2023-02-10 01:19:24,069 32k INFO ====> Epoch: 58
|
118 |
+
2023-02-10 01:20:48,652 32k INFO ====> Epoch: 59
|
119 |
+
2023-02-10 01:22:13,434 32k INFO ====> Epoch: 60
|
120 |
+
2023-02-10 01:23:08,515 32k INFO Train Epoch: 61 [53%]
|
121 |
+
2023-02-10 01:23:08,516 32k INFO [2.509904146194458, 2.3118839263916016, 10.964993476867676, 18.20663833618164, 0.8472650647163391, 4600, 9.92527589532945e-05]
|
122 |
+
2023-02-10 01:23:38,264 32k INFO ====> Epoch: 61
|
123 |
+
2023-02-10 01:25:02,945 32k INFO ====> Epoch: 62
|
124 |
+
2023-02-10 01:26:27,685 32k INFO ====> Epoch: 63
|
125 |
+
2023-02-10 01:26:59,803 32k INFO Train Epoch: 64 [16%]
|
126 |
+
2023-02-10 01:26:59,803 32k INFO [2.302783489227295, 2.5982413291931152, 10.80894660949707, 14.60811996459961, 1.0399219989776611, 4800, 9.921554382096622e-05]
|
127 |
+
2023-02-10 01:27:52,584 32k INFO ====> Epoch: 64
|
128 |
+
2023-02-10 01:29:17,080 32k INFO ====> Epoch: 65
|
129 |
+
2023-02-10 01:30:28,770 32k INFO Train Epoch: 66 [79%]
|
130 |
+
2023-02-10 01:30:28,770 32k INFO [2.7284963130950928, 2.3790788650512695, 9.036128044128418, 13.677116394042969, 0.6173059940338135, 5000, 9.919074148525384e-05]
|
131 |
+
2023-02-10 01:30:33,203 32k INFO Saving model and optimizer state at iteration 66 to ./logs\32k\G_5000.pth
|
132 |
+
2023-02-10 01:30:50,085 32k INFO Saving model and optimizer state at iteration 66 to ./logs\32k\D_5000.pth
|
133 |
+
2023-02-10 01:31:06,739 32k INFO ====> Epoch: 66
|
134 |
+
2023-02-10 01:32:31,353 32k INFO ====> Epoch: 67
|
135 |
+
2023-02-10 01:33:57,529 32k INFO ====> Epoch: 68
|
136 |
+
2023-02-10 01:34:46,261 32k INFO Train Epoch: 69 [42%]
|
137 |
+
2023-02-10 01:34:46,262 32k INFO [2.5467963218688965, 2.0986404418945312, 10.101814270019531, 14.508784294128418, 1.3656277656555176, 5200, 9.915354960656915e-05]
|
138 |
+
2023-02-10 01:35:22,514 32k INFO ====> Epoch: 69
|
139 |
+
2023-02-10 01:36:47,135 32k INFO ====> Epoch: 70
|
140 |
+
2023-02-10 01:38:11,576 32k INFO ====> Epoch: 71
|
141 |
+
2023-02-10 01:38:37,123 32k INFO Train Epoch: 72 [5%]
|
142 |
+
2023-02-10 01:38:37,123 32k INFO [1.857373833656311, 3.1599011421203613, 8.379103660583496, 9.244300842285156, 1.2350279092788696, 5400, 9.911637167309565e-05]
|
143 |
+
2023-02-10 01:39:36,388 32k INFO ====> Epoch: 72
|
144 |
+
2023-02-10 01:41:00,864 32k INFO ====> Epoch: 73
|
145 |
+
2023-02-10 01:42:05,899 32k INFO Train Epoch: 74 [68%]
|
146 |
+
2023-02-10 01:42:05,900 32k INFO [2.3246712684631348, 2.638263463973999, 7.793229579925537, 13.017571449279785, 0.9624058604240417, 5600, 9.909159412887068e-05]
|
147 |
+
2023-02-10 01:42:25,722 32k INFO ====> Epoch: 74
|
148 |
+
2023-02-10 01:43:50,155 32k INFO ====> Epoch: 75
|
149 |
+
2023-02-10 01:45:14,883 32k INFO ====> Epoch: 76
|
150 |
+
2023-02-10 01:45:56,818 32k INFO Train Epoch: 77 [32%]
|
151 |
+
2023-02-10 01:45:56,818 32k INFO [2.2799293994903564, 2.503884792327881, 12.849569320678711, 18.74350929260254, 0.9742995500564575, 5800, 9.905443942579728e-05]
|
152 |
+
2023-02-10 01:46:39,621 32k INFO ====> Epoch: 77
|
153 |
+
2023-02-10 01:48:04,267 32k INFO ====> Epoch: 78
|
154 |
+
2023-02-10 01:49:25,876 32k INFO Train Epoch: 79 [95%]
|
155 |
+
2023-02-10 01:49:25,876 32k INFO [2.3028907775878906, 2.451934337615967, 11.012730598449707, 17.174062728881836, 0.7448775768280029, 6000, 9.902967736366644e-05]
|
156 |
+
2023-02-10 01:49:30,404 32k INFO Saving model and optimizer state at iteration 79 to ./logs\32k\G_6000.pth
|
157 |
+
2023-02-10 01:49:47,910 32k INFO Saving model and optimizer state at iteration 79 to ./logs\32k\D_6000.pth
|
158 |
+
2023-02-10 01:49:54,757 32k INFO ====> Epoch: 79
|
159 |
+
2023-02-10 01:51:19,458 32k INFO ====> Epoch: 80
|
160 |
+
2023-02-10 01:52:44,119 32k INFO ====> Epoch: 81
|
161 |
+
2023-02-10 01:53:42,673 32k INFO Train Epoch: 82 [58%]
|
162 |
+
2023-02-10 01:53:42,674 32k INFO [2.279362201690674, 2.260122776031494, 8.51861572265625, 12.269641876220703, 0.7776111960411072, 6200, 9.899254587647776e-05]
|
163 |
+
2023-02-10 01:54:09,077 32k INFO ====> Epoch: 82
|
164 |
+
2023-02-10 01:55:33,754 32k INFO ====> Epoch: 83
|
165 |
+
2023-02-10 01:56:58,300 32k INFO ====> Epoch: 84
|
166 |
+
2023-02-10 01:57:33,593 32k INFO Train Epoch: 85 [21%]
|
167 |
+
2023-02-10 01:57:33,594 32k INFO [2.6448638439178467, 2.2926855087280273, 11.688180923461914, 18.19353675842285, 0.9365038275718689, 6400, 9.895542831185631e-05]
|
168 |
+
2023-02-10 01:58:23,117 32k INFO ====> Epoch: 85
|
169 |
+
2023-02-10 01:59:47,507 32k INFO ====> Epoch: 86
|
170 |
+
2023-02-10 02:01:02,283 32k INFO Train Epoch: 87 [84%]
|
171 |
+
2023-02-10 02:01:02,283 32k INFO [2.427564859390259, 2.242896556854248, 11.131178855895996, 16.051637649536133, 0.914648711681366, 6600, 9.89306910009569e-05]
|
172 |
+
2023-02-10 02:01:12,207 32k INFO ====> Epoch: 87
|
173 |
+
2023-02-10 02:02:36,685 32k INFO ====> Epoch: 88
|
174 |
+
2023-02-10 02:04:01,181 32k INFO ====> Epoch: 89
|
175 |
+
2023-02-10 02:04:52,920 32k INFO Train Epoch: 90 [47%]
|
176 |
+
2023-02-10 02:04:52,921 32k INFO [2.4379117488861084, 2.36348557472229, 8.932344436645508, 16.195209503173828, 0.8471890091896057, 6800, 9.889359662901445e-05]
|
177 |
+
2023-02-10 02:05:25,823 32k INFO ====> Epoch: 90
|
178 |
+
2023-02-10 02:06:50,173 32k INFO ====> Epoch: 91
|
179 |
+
2023-02-10 02:08:14,749 32k INFO ====> Epoch: 92
|
180 |
+
2023-02-10 02:08:43,328 32k INFO Train Epoch: 93 [11%]
|
181 |
+
2023-02-10 02:08:43,329 32k INFO [2.467284917831421, 2.144986391067505, 11.449904441833496, 16.176807403564453, 0.4319636821746826, 7000, 9.885651616572276e-05]
|
182 |
+
2023-02-10 02:08:47,768 32k INFO Saving model and optimizer state at iteration 93 to ./logs\32k\G_7000.pth
|
183 |
+
2023-02-10 02:09:04,812 32k INFO Saving model and optimizer state at iteration 93 to ./logs\32k\D_7000.pth
|
184 |
+
2023-02-10 02:10:04,543 32k INFO ====> Epoch: 93
|
185 |
+
2023-02-10 02:11:29,241 32k INFO ====> Epoch: 94
|
186 |
+
2023-02-10 02:12:37,720 32k INFO Train Epoch: 95 [74%]
|
187 |
+
2023-02-10 02:12:37,720 32k INFO [2.2731475830078125, 2.425633430480957, 10.031761169433594, 14.7035493850708, 0.8044810891151428, 7200, 9.883180358131438e-05]
|
188 |
+
2023-02-10 02:12:54,162 32k INFO ====> Epoch: 95
|
189 |
+
2023-02-10 02:14:18,638 32k INFO ====> Epoch: 96
|
190 |
+
2023-02-10 02:15:42,921 32k INFO ====> Epoch: 97
|
191 |
+
2023-02-10 02:16:28,193 32k INFO Train Epoch: 98 [37%]
|
192 |
+
2023-02-10 02:16:28,194 32k INFO [2.3223936557769775, 2.573565721511841, 10.411147117614746, 19.68339729309082, 0.5216219425201416, 7400, 9.879474628751914e-05]
|
193 |
+
2023-02-10 02:17:07,685 32k INFO ====> Epoch: 98
|
194 |
+
2023-02-10 02:18:32,253 32k INFO ====> Epoch: 99
|
195 |
+
2023-02-10 02:19:56,642 32k INFO ====> Epoch: 100
|
196 |
+
2023-02-10 02:20:18,741 32k INFO Train Epoch: 101 [0%]
|
197 |
+
2023-02-10 02:20:18,742 32k INFO [2.397695541381836, 2.1934003829956055, 8.183572769165039, 12.770599365234375, 1.0327764749526978, 7600, 9.875770288847208e-05]
|
198 |
+
2023-02-10 02:21:21,459 32k INFO ====> Epoch: 101
|
199 |
+
2023-02-10 02:22:46,182 32k INFO ====> Epoch: 102
|
200 |
+
2023-02-10 02:23:47,848 32k INFO Train Epoch: 103 [63%]
|
201 |
+
2023-02-10 02:23:47,848 32k INFO [2.461080312728882, 2.3870487213134766, 6.649186134338379, 10.037254333496094, 0.8956831693649292, 7800, 9.873301500583906e-05]
|
202 |
+
2023-02-10 02:24:10,873 32k INFO ====> Epoch: 103
|
203 |
+
2023-02-10 02:25:35,259 32k INFO ====> Epoch: 104
|
204 |
+
2023-02-10 02:26:59,649 32k INFO ====> Epoch: 105
|
205 |
+
2023-02-10 02:27:38,326 32k INFO Train Epoch: 106 [26%]
|
206 |
+
2023-02-10 02:27:38,327 32k INFO [1.846121072769165, 3.3351776599884033, 8.606220245361328, 9.130136489868164, 0.8224015831947327, 8000, 9.86959947531291e-05]
|
207 |
+
2023-02-10 02:27:42,852 32k INFO Saving model and optimizer state at iteration 106 to ./logs\32k\G_8000.pth
|
208 |
+
2023-02-10 02:28:02,191 32k INFO Saving model and optimizer state at iteration 106 to ./logs\32k\D_8000.pth
|
209 |
+
2023-02-10 02:28:52,049 32k INFO ====> Epoch: 106
|
210 |
+
2023-02-10 02:30:17,599 32k INFO ====> Epoch: 107
|
211 |
+
2023-02-10 02:31:35,925 32k INFO Train Epoch: 108 [89%]
|
212 |
+
2023-02-10 02:31:35,925 32k INFO [2.4512813091278076, 2.255336046218872, 7.023709297180176, 11.943727493286133, 0.9954293370246887, 8200, 9.867132229656573e-05]
|
213 |
+
2023-02-10 02:31:42,471 32k INFO ====> Epoch: 108
|
214 |
+
2023-02-10 02:33:07,180 32k INFO ====> Epoch: 109
|
215 |
+
2023-02-10 02:34:31,690 32k INFO ====> Epoch: 110
|
216 |
+
2023-02-10 02:35:26,947 32k INFO Train Epoch: 111 [53%]
|
217 |
+
2023-02-10 02:35:26,947 32k INFO [2.391657590866089, 2.3984644412994385, 13.551353454589844, 14.703734397888184, 0.9091495871543884, 8400, 9.863432517573002e-05]
|
218 |
+
2023-02-10 02:35:56,571 32k INFO ====> Epoch: 111
|
219 |
+
2023-02-10 02:37:21,163 32k INFO ====> Epoch: 112
|
220 |
+
2023-02-10 02:38:45,688 32k INFO ====> Epoch: 113
|
221 |
+
2023-02-10 02:39:17,654 32k INFO Train Epoch: 114 [16%]
|
222 |
+
2023-02-10 02:39:17,655 32k INFO [2.39654278755188, 2.245805263519287, 10.06233024597168, 14.09300708770752, 0.6480247974395752, 8600, 9.859734192708044e-05]
|
223 |
+
2023-02-10 02:40:10,405 32k INFO ====> Epoch: 114
|
224 |
+
2023-02-10 02:41:34,622 32k INFO ====> Epoch: 115
|
225 |
+
2023-02-10 02:42:46,274 32k INFO Train Epoch: 116 [79%]
|
226 |
+
2023-02-10 02:42:46,274 32k INFO [2.5749049186706543, 2.414797306060791, 9.960501670837402, 16.4396915435791, 0.789290189743042, 8800, 9.857269413218213e-05]
|
227 |
+
2023-02-10 02:42:59,471 32k INFO ====> Epoch: 116
|
228 |
+
2023-02-10 02:44:24,046 32k INFO ====> Epoch: 117
|
229 |
+
2023-02-10 02:45:48,517 32k INFO ====> Epoch: 118
|
230 |
+
2023-02-10 02:46:37,109 32k INFO Train Epoch: 119 [42%]
|
231 |
+
2023-02-10 02:46:37,109 32k INFO [2.4546942710876465, 2.1338768005371094, 12.899450302124023, 15.397659301757812, 0.8034476637840271, 9000, 9.853573399228505e-05]
|
232 |
+
2023-02-10 02:46:41,546 32k INFO Saving model and optimizer state at iteration 119 to ./logs\32k\G_9000.pth
|
233 |
+
2023-02-10 02:47:01,129 32k INFO Saving model and optimizer state at iteration 119 to ./logs\32k\D_9000.pth
|
234 |
+
2023-02-10 02:47:40,886 32k INFO ====> Epoch: 119
|
235 |
+
2023-02-10 02:49:05,371 32k INFO ====> Epoch: 120
|
236 |
+
2023-02-10 02:50:29,859 32k INFO ====> Epoch: 121
|
237 |
+
2023-02-10 02:50:55,463 32k INFO Train Epoch: 122 [5%]
|
238 |
+
2023-02-10 02:50:55,463 32k INFO [2.207282543182373, 2.6213908195495605, 7.647693634033203, 14.46692180633545, 0.9529262185096741, 9200, 9.8498787710708e-05]
|
239 |
+
2023-02-10 02:51:55,210 32k INFO ====> Epoch: 122
|
240 |
+
2023-02-10 02:53:19,953 32k INFO ====> Epoch: 123
|
241 |
+
2023-02-10 02:54:25,092 32k INFO Train Epoch: 124 [68%]
|
242 |
+
2023-02-10 02:54:25,093 32k INFO [2.332580089569092, 2.3436594009399414, 6.139850616455078, 10.373167037963867, 0.9625253677368164, 9400, 9.847416455282387e-05]
|
243 |
+
2023-02-10 02:54:44,808 32k INFO ====> Epoch: 124
|
244 |
+
2023-02-10 02:56:09,403 32k INFO ====> Epoch: 125
|
245 |
+
2023-02-10 02:57:35,700 32k INFO ====> Epoch: 126
|
246 |
+
2023-02-10 02:58:17,551 32k INFO Train Epoch: 127 [32%]
|
247 |
+
2023-02-10 02:58:17,552 32k INFO [2.467034339904785, 2.2089598178863525, 7.952523708343506, 16.058887481689453, 0.8524291515350342, 9600, 9.84372413569007e-05]
|
248 |
+
2023-02-10 02:59:00,413 32k INFO ====> Epoch: 127
|
249 |
+
2023-02-10 03:00:24,954 32k INFO ====> Epoch: 128
|
250 |
+
2023-02-10 03:01:46,396 32k INFO Train Epoch: 129 [95%]
|
251 |
+
2023-02-10 03:01:46,397 32k INFO [2.367544174194336, 2.7189576625823975, 6.78631591796875, 12.748686790466309, 0.4685216546058655, 9800, 9.841263358464336e-05]
|
252 |
+
2023-02-10 03:01:49,616 32k INFO ====> Epoch: 129
|
253 |
+
2023-02-10 03:03:14,271 32k INFO ====> Epoch: 130
|
254 |
+
2023-02-10 03:04:38,789 32k INFO ====> Epoch: 131
|
255 |
+
2023-02-10 03:05:37,230 32k INFO Train Epoch: 132 [58%]
|
256 |
+
2023-02-10 03:05:37,231 32k INFO [2.4509665966033936, 2.7042665481567383, 10.243302345275879, 13.67796802520752, 0.6009257435798645, 10000, 9.837573345994909e-05]
|
257 |
+
2023-02-10 03:05:41,749 32k INFO Saving model and optimizer state at iteration 132 to ./logs\32k\G_10000.pth
|
258 |
+
2023-02-10 03:05:58,430 32k INFO Saving model and optimizer state at iteration 132 to ./logs\32k\D_10000.pth
|
259 |
+
2023-02-10 03:06:28,668 32k INFO ====> Epoch: 132
|
260 |
+
2023-02-10 03:07:54,369 32k INFO ====> Epoch: 133
|
261 |
+
2023-02-10 03:09:19,084 32k INFO ====> Epoch: 134
|
262 |
+
2023-02-10 03:09:54,319 32k INFO Train Epoch: 135 [21%]
|
263 |
+
2023-02-10 03:09:54,319 32k INFO [2.5057411193847656, 2.4407787322998047, 9.817587852478027, 13.728426933288574, 0.6413305997848511, 10200, 9.833884717107196e-05]
|
264 |
+
2023-02-10 03:10:43,813 32k INFO ====> Epoch: 135
|
265 |
+
2023-02-10 03:12:08,539 32k INFO ====> Epoch: 136
|
266 |
+
2023-02-10 03:13:23,629 32k INFO Train Epoch: 137 [84%]
|
267 |
+
2023-02-10 03:13:23,629 32k INFO [2.5296790599823, 2.208813190460205, 6.762031078338623, 10.408514976501465, 0.754698634147644, 10400, 9.831426399582366e-05]
|
268 |
+
2023-02-10 03:13:33,495 32k INFO ====> Epoch: 137
|
269 |
+
2023-02-10 03:14:58,219 32k INFO ====> Epoch: 138
|
270 |
+
2023-02-10 03:16:22,869 32k INFO ====> Epoch: 139
|
271 |
+
2023-02-10 03:17:14,809 32k INFO Train Epoch: 140 [47%]
|
272 |
+
2023-02-10 03:17:14,809 32k INFO [2.5353844165802, 2.2106950283050537, 10.004817008972168, 13.992541313171387, 0.6150087714195251, 10600, 9.827740075511432e-05]
|
273 |
+
2023-02-10 03:17:47,789 32k INFO ====> Epoch: 140
|
274 |
+
2023-02-10 03:19:12,269 32k INFO ====> Epoch: 141
|
275 |
+
2023-02-10 03:20:36,797 32k INFO ====> Epoch: 142
|
276 |
+
2023-02-10 03:21:05,474 32k INFO Train Epoch: 143 [11%]
|
277 |
+
2023-02-10 03:21:05,475 32k INFO [2.335813522338867, 2.381714344024658, 9.432477951049805, 14.522370338439941, 0.697417140007019, 10800, 9.824055133639235e-05]
|
278 |
+
2023-02-10 03:22:01,624 32k INFO ====> Epoch: 143
|
279 |
+
2023-02-10 03:23:26,232 32k INFO ====> Epoch: 144
|
280 |
+
2023-02-10 03:24:34,537 32k INFO Train Epoch: 145 [74%]
|
281 |
+
2023-02-10 03:24:34,538 32k INFO [1.924493670463562, 3.025908946990967, 10.635231018066406, 12.46295166015625, 0.3505359888076782, 11000, 9.821599273356685e-05]
|
282 |
+
2023-02-10 03:24:39,057 32k INFO Saving model and optimizer state at iteration 145 to ./logs\32k\G_11000.pth
|
283 |
+
2023-02-10 03:24:57,871 32k INFO Saving model and optimizer state at iteration 145 to ./logs\32k\D_11000.pth
|
284 |
+
2023-02-10 03:25:18,230 32k INFO ====> Epoch: 145
|
285 |
+
2023-02-10 03:26:42,920 32k INFO ====> Epoch: 146
|
286 |
+
2023-02-10 03:28:07,571 32k INFO ====> Epoch: 147
|
287 |
+
2023-02-10 03:28:52,871 32k INFO Train Epoch: 148 [37%]
|
288 |
+
2023-02-10 03:28:52,872 32k INFO [2.568720579147339, 2.2338619232177734, 6.399820327758789, 12.950323104858398, 0.9168809056282043, 11200, 9.817916633997459e-05]
|
289 |
+
2023-02-10 03:29:32,512 32k INFO ====> Epoch: 148
|
290 |
+
2023-02-10 03:30:57,311 32k INFO ====> Epoch: 149
|
291 |
+
2023-02-10 03:32:22,084 32k INFO ====> Epoch: 150
|
292 |
+
2023-02-10 03:32:44,293 32k INFO Train Epoch: 151 [0%]
|
293 |
+
2023-02-10 03:32:44,293 32k INFO [2.3033745288848877, 2.415085792541504, 10.928938865661621, 16.1735897064209, 0.3847680985927582, 11400, 9.814235375455375e-05]
|
294 |
+
2023-02-10 03:33:47,215 32k INFO ====> Epoch: 151
|
295 |
+
2023-02-10 03:35:11,867 32k INFO ====> Epoch: 152
|
296 |
+
2023-02-10 03:36:13,752 32k INFO Train Epoch: 153 [63%]
|
297 |
+
2023-02-10 03:36:13,753 32k INFO [2.173736333847046, 2.650914192199707, 12.767165184020996, 18.795696258544922, 0.4320840835571289, 11600, 9.811781969958938e-05]
|
298 |
+
2023-02-10 03:36:36,873 32k INFO ====> Epoch: 153
|
299 |
+
2023-02-10 03:38:01,486 32k INFO ====> Epoch: 154
|
300 |
+
2023-02-10 03:39:26,006 32k INFO ====> Epoch: 155
|
301 |
+
2023-02-10 03:40:04,581 32k INFO Train Epoch: 156 [26%]
|
302 |
+
2023-02-10 03:40:04,582 32k INFO [2.574475049972534, 2.363281488418579, 8.882105827331543, 10.567209243774414, 0.11711447685956955, 11800, 9.808103011628319e-05]
|
303 |
+
2023-02-10 03:40:50,861 32k INFO ====> Epoch: 156
|
304 |
+
2023-02-10 03:42:15,437 32k INFO ====> Epoch: 157
|
305 |
+
2023-02-10 03:43:33,696 32k INFO Train Epoch: 158 [89%]
|
306 |
+
2023-02-10 03:43:33,696 32k INFO [2.343777894973755, 2.4922895431518555, 10.256402969360352, 16.83711051940918, 0.874286413192749, 12000, 9.80565113912702e-05]
|
307 |
+
2023-02-10 03:43:38,306 32k INFO Saving model and optimizer state at iteration 158 to ./logs\32k\G_12000.pth
|
308 |
+
2023-02-10 03:43:55,136 32k INFO Saving model and optimizer state at iteration 158 to ./logs\32k\D_12000.pth
|
309 |
+
2023-02-10 03:44:05,409 32k INFO ====> Epoch: 158
|
310 |
+
2023-02-10 03:45:30,201 32k INFO ====> Epoch: 159
|
311 |
+
2023-02-10 03:46:54,691 32k INFO ====> Epoch: 160
|
312 |
+
2023-02-10 03:47:49,976 32k INFO Train Epoch: 161 [53%]
|
313 |
+
2023-02-10 03:47:49,976 32k INFO [2.596492290496826, 2.6639647483825684, 7.916460037231445, 15.857110977172852, 0.8963372707366943, 12200, 9.801974479570593e-05]
|
314 |
+
2023-02-10 03:48:19,643 32k INFO ====> Epoch: 161
|
315 |
+
2023-02-10 03:49:44,342 32k INFO ====> Epoch: 162
|
316 |
+
2023-02-10 03:51:09,047 32k INFO ====> Epoch: 163
|
317 |
+
2023-02-10 03:51:41,212 32k INFO Train Epoch: 164 [16%]
|
318 |
+
2023-02-10 03:51:41,212 32k INFO [2.252506971359253, 2.4695920944213867, 11.395182609558105, 16.45693016052246, 0.94123774766922, 12400, 9.798299198589162e-05]
|
319 |
+
2023-02-10 03:52:34,132 32k INFO ====> Epoch: 164
|
320 |
+
2023-02-10 03:53:59,347 32k INFO ====> Epoch: 165
|
321 |
+
2023-02-10 03:55:11,032 32k INFO Train Epoch: 166 [79%]
|
322 |
+
2023-02-10 03:55:11,032 32k INFO [2.5481297969818115, 2.5002939701080322, 10.718299865722656, 17.71586036682129, 0.3888394236564636, 12600, 9.795849776887939e-05]
|
323 |
+
2023-02-10 03:55:24,184 32k INFO ====> Epoch: 166
|
324 |
+
2023-02-10 03:56:48,762 32k INFO ====> Epoch: 167
|
325 |
+
2023-02-10 03:58:13,310 32k INFO ====> Epoch: 168
|
326 |
+
2023-02-10 03:59:01,779 32k INFO Train Epoch: 169 [42%]
|
327 |
+
2023-02-10 03:59:01,779 32k INFO [2.531505584716797, 1.924519658088684, 13.503902435302734, 16.02785301208496, 0.9448908567428589, 12800, 9.792176792382932e-05]
|
328 |
+
2023-02-10 03:59:38,066 32k INFO ====> Epoch: 169
|
329 |
+
2023-02-10 04:01:02,587 32k INFO ====> Epoch: 170
|
330 |
+
2023-02-10 04:02:27,100 32k INFO ====> Epoch: 171
|
331 |
+
2023-02-10 04:02:52,429 32k INFO Train Epoch: 172 [5%]
|
332 |
+
2023-02-10 04:02:52,430 32k INFO [2.5668065547943115, 2.419389247894287, 8.094529151916504, 14.7818603515625, 0.7867487072944641, 13000, 9.78850518507495e-05]
|
333 |
+
2023-02-10 04:02:56,916 32k INFO Saving model and optimizer state at iteration 172 to ./logs\32k\G_13000.pth
|
334 |
+
2023-02-10 04:03:14,363 32k INFO Saving model and optimizer state at iteration 172 to ./logs\32k\D_13000.pth
|
335 |
+
2023-02-10 04:04:17,632 32k INFO ====> Epoch: 172
|
336 |
+
2023-02-10 04:05:42,306 32k INFO ====> Epoch: 173
|
337 |
+
2023-02-10 04:06:47,480 32k INFO Train Epoch: 174 [68%]
|
338 |
+
2023-02-10 04:06:47,480 32k INFO [2.1723153591156006, 2.5889978408813477, 10.076593399047852, 12.853192329406738, 0.3051690459251404, 13200, 9.786058211724074e-05]
|
339 |
+
2023-02-10 04:07:07,239 32k INFO ====> Epoch: 174
|
340 |
+
2023-02-10 04:08:31,852 32k INFO ====> Epoch: 175
|
341 |
+
2023-02-10 04:09:56,580 32k INFO ====> Epoch: 176
|
342 |
+
2023-02-10 04:10:38,557 32k INFO Train Epoch: 177 [32%]
|
343 |
+
2023-02-10 04:10:38,558 32k INFO [2.395698070526123, 2.4369287490844727, 11.675952911376953, 15.961197853088379, 0.7354399561882019, 13400, 9.782388898597041e-05]
|
344 |
+
2023-02-10 04:11:21,538 32k INFO ====> Epoch: 177
|
345 |
+
2023-02-10 04:12:46,203 32k INFO ====> Epoch: 178
|
346 |
+
2023-02-10 04:14:07,776 32k INFO Train Epoch: 179 [95%]
|
347 |
+
2023-02-10 04:14:07,777 32k INFO [2.7345943450927734, 1.9628326892852783, 6.206479072570801, 7.970531940460205, 0.8745248913764954, 13600, 9.779943454222217e-05]
|
348 |
+
2023-02-10 04:14:10,989 32k INFO ====> Epoch: 179
|
349 |
+
2023-02-10 04:15:35,842 32k INFO ====> Epoch: 180
|
350 |
+
2023-02-10 04:17:00,411 32k INFO ====> Epoch: 181
|
351 |
+
2023-02-10 04:17:58,909 32k INFO Train Epoch: 182 [58%]
|
352 |
+
2023-02-10 04:17:58,909 32k INFO [2.2180511951446533, 2.461686372756958, 7.78315544128418, 12.147871971130371, 0.7892422080039978, 13800, 9.776276433842631e-05]
|
353 |
+
2023-02-10 04:18:25,267 32k INFO ====> Epoch: 182
|
354 |
+
2023-02-10 04:19:49,727 32k INFO ====> Epoch: 183
|
355 |
+
2023-02-10 04:21:14,382 32k INFO ====> Epoch: 184
|
356 |
+
2023-02-10 04:21:49,851 32k INFO Train Epoch: 185 [21%]
|
357 |
+
2023-02-10 04:21:49,851 32k INFO [2.5905978679656982, 2.213409185409546, 9.200039863586426, 12.909902572631836, 0.1616513580083847, 14000, 9.772610788423802e-05]
|
358 |
+
2023-02-10 04:21:54,470 32k INFO Saving model and optimizer state at iteration 185 to ./logs\32k\G_14000.pth
|
359 |
+
2023-02-10 04:22:11,052 32k INFO Saving model and optimizer state at iteration 185 to ./logs\32k\D_14000.pth
|
360 |
+
2023-02-10 04:23:03,696 32k INFO ====> Epoch: 185
|
361 |
+
2023-02-10 04:24:28,374 32k INFO ====> Epoch: 186
|
362 |
+
2023-02-10 04:25:43,503 32k INFO Train Epoch: 187 [84%]
|
363 |
+
2023-02-10 04:25:43,504 32k INFO [2.115373134613037, 2.6405253410339355, 13.0419340133667, 17.158859252929688, 0.6063501238822937, 14200, 9.77016778842374e-05]
|
364 |
+
2023-02-10 04:25:53,337 32k INFO ====> Epoch: 187
|
365 |
+
2023-02-10 04:27:18,024 32k INFO ====> Epoch: 188
|
366 |
+
2023-02-10 04:28:42,529 32k INFO ====> Epoch: 189
|
367 |
+
2023-02-10 04:29:34,368 32k INFO Train Epoch: 190 [47%]
|
368 |
+
2023-02-10 04:29:34,368 32k INFO [2.0680086612701416, 3.0300402641296387, 8.614965438842773, 12.649006843566895, 0.691765546798706, 14400, 9.766504433460612e-05]
|
369 |
+
2023-02-10 04:30:07,450 32k INFO ====> Epoch: 190
|
370 |
+
2023-02-10 04:31:32,088 32k INFO ====> Epoch: 191
|
371 |
+
2023-02-10 04:32:56,573 32k INFO ====> Epoch: 192
|
372 |
+
2023-02-10 04:33:25,364 32k INFO Train Epoch: 193 [11%]
|
373 |
+
2023-02-10 04:33:25,365 32k INFO [2.3075990676879883, 2.3698666095733643, 14.287137985229492, 16.29629135131836, 0.5351936221122742, 14600, 9.762842452083883e-05]
|
374 |
+
2023-02-10 04:34:21,439 32k INFO ====> Epoch: 193
|
375 |
+
2023-02-10 04:35:46,041 32k INFO ====> Epoch: 194
|
376 |
+
2023-02-10 04:36:54,340 32k INFO Train Epoch: 195 [74%]
|
377 |
+
2023-02-10 04:36:54,341 32k INFO [2.3796615600585938, 2.2257769107818604, 9.005014419555664, 12.171199798583984, 0.7831335067749023, 14800, 9.760401894015275e-05]
|
378 |
+
2023-02-10 04:37:10,784 32k INFO ====> Epoch: 195
|
379 |
+
2023-02-10 04:38:35,380 32k INFO ====> Epoch: 196
|
380 |
+
2023-02-10 04:39:59,882 32k INFO ====> Epoch: 197
|
381 |
+
2023-02-10 04:40:45,368 32k INFO Train Epoch: 198 [37%]
|
382 |
+
2023-02-10 04:40:45,368 32k INFO [2.5116686820983887, 2.4769821166992188, 9.495701789855957, 15.055347442626953, 0.6151214241981506, 15000, 9.756742200804793e-05]
|
383 |
+
2023-02-10 04:40:49,881 32k INFO Saving model and optimizer state at iteration 198 to ./logs\32k\G_15000.pth
|
384 |
+
2023-02-10 04:41:08,000 32k INFO Saving model and optimizer state at iteration 198 to ./logs\32k\D_15000.pth
|
385 |
+
2023-02-10 04:41:51,223 32k INFO ====> Epoch: 198
|
386 |
+
2023-02-10 04:43:15,797 32k INFO ====> Epoch: 199
|
387 |
+
2023-02-10 04:44:40,469 32k INFO ====> Epoch: 200
|
388 |
+
2023-02-10 04:45:02,531 32k INFO Train Epoch: 201 [0%]
|
389 |
+
2023-02-10 04:45:02,531 32k INFO [2.66977858543396, 2.327565908432007, 8.785197257995605, 12.054095268249512, 0.33490169048309326, 15200, 9.753083879807726e-05]
|
390 |
+
2023-02-10 04:46:05,243 32k INFO ====> Epoch: 201
|
391 |
+
2023-02-10 04:47:29,831 32k INFO ====> Epoch: 202
|
392 |
+
2023-02-10 04:48:31,669 32k INFO Train Epoch: 203 [63%]
|
393 |
+
2023-02-10 04:48:31,669 32k INFO [2.4113712310791016, 2.208491325378418, 7.582664966583252, 10.762511253356934, 0.24080340564250946, 15400, 9.750645761229709e-05]
|
394 |
+
2023-02-10 04:48:54,843 32k INFO ====> Epoch: 203
|
395 |
+
2023-02-10 04:50:20,441 32k INFO ====> Epoch: 204
|
396 |
+
2023-02-10 04:51:44,905 32k INFO ====> Epoch: 205
|
397 |
+
2023-02-10 04:52:23,595 32k INFO Train Epoch: 206 [26%]
|
398 |
+
2023-02-10 04:52:23,596 32k INFO [2.406938076019287, 2.329049587249756, 13.670656204223633, 15.57223129272461, 0.5972563028335571, 15600, 9.746989726111722e-05]
|
399 |
+
2023-02-10 04:53:09,786 32k INFO ====> Epoch: 206
|
400 |
+
2023-02-10 04:54:34,352 32k INFO ====> Epoch: 207
|
401 |
+
2023-02-10 04:55:52,644 32k INFO Train Epoch: 208 [89%]
|
402 |
+
2023-02-10 04:55:52,645 32k INFO [2.4009199142456055, 2.4792189598083496, 8.647868156433105, 13.368301391601562, 0.6690515875816345, 15800, 9.744553130976908e-05]
|
403 |
+
2023-02-10 04:55:59,209 32k INFO ====> Epoch: 208
|
404 |
+
2023-02-10 04:57:23,732 32k INFO ====> Epoch: 209
|
405 |
+
2023-02-10 04:58:48,239 32k INFO ====> Epoch: 210
|
406 |
+
2023-02-10 04:59:43,567 32k INFO Train Epoch: 211 [53%]
|
407 |
+
2023-02-10 04:59:43,567 32k INFO [2.491450309753418, 2.5477209091186523, 7.471154689788818, 13.8217191696167, 0.5033202767372131, 16000, 9.740899380309685e-05]
|
408 |
+
2023-02-10 04:59:47,998 32k INFO Saving model and optimizer state at iteration 211 to ./logs\32k\G_16000.pth
|
409 |
+
2023-02-10 05:00:05,780 32k INFO Saving model and optimizer state at iteration 211 to ./logs\32k\D_16000.pth
|
410 |
+
2023-02-10 05:00:39,185 32k INFO ====> Epoch: 211
|
411 |
+
2023-02-10 05:02:04,575 32k INFO ====> Epoch: 212
|
412 |
+
2023-02-10 05:03:29,926 32k INFO ====> Epoch: 213
|
413 |
+
2023-02-10 05:04:02,744 32k INFO Train Epoch: 214 [16%]
|
414 |
+
2023-02-10 05:04:02,744 32k INFO [2.5403432846069336, 2.3642735481262207, 10.938908576965332, 14.273392677307129, 0.46957525610923767, 16200, 9.7372469996277e-05]
|
415 |
+
2023-02-10 05:04:55,587 32k INFO ====> Epoch: 214
|
416 |
+
2023-02-10 05:06:20,137 32k INFO ====> Epoch: 215
|
417 |
+
2023-02-10 05:07:31,842 32k INFO Train Epoch: 216 [79%]
|
418 |
+
2023-02-10 05:07:31,842 32k INFO [2.339552879333496, 2.709596872329712, 11.353394508361816, 18.01609992980957, 0.5840747356414795, 16400, 9.734812840022278e-05]
|
419 |
+
2023-02-10 05:07:45,003 32k INFO ====> Epoch: 216
|
420 |
+
2023-02-10 05:09:10,461 32k INFO ====> Epoch: 217
|
421 |
+
2023-02-10 05:10:35,839 32k INFO ====> Epoch: 218
|
422 |
+
2023-02-10 05:11:24,387 32k INFO Train Epoch: 219 [42%]
|
423 |
+
2023-02-10 05:11:24,387 32k INFO [2.4308414459228516, 2.6758008003234863, 9.025983810424805, 12.856972694396973, 0.7895928621292114, 16600, 9.731162741507607e-05]
|
424 |
+
2023-02-10 05:12:00,773 32k INFO ====> Epoch: 219
|
425 |
+
2023-02-10 05:13:26,131 32k INFO ====> Epoch: 220
|
426 |
+
2023-02-10 05:14:50,815 32k INFO ====> Epoch: 221
|
427 |
+
2023-02-10 05:15:16,163 32k INFO Train Epoch: 222 [5%]
|
428 |
+
2023-02-10 05:15:16,163 32k INFO [2.519866704940796, 2.3976869583129883, 6.224302768707275, 9.703218460083008, 0.8612720966339111, 16800, 9.727514011608789e-05]
|
429 |
+
2023-02-10 05:16:15,647 32k INFO ====> Epoch: 222
|
430 |
+
2023-02-10 05:17:41,078 32k INFO ====> Epoch: 223
|
431 |
+
2023-02-10 05:18:46,053 32k INFO Train Epoch: 224 [68%]
|
432 |
+
2023-02-10 05:18:46,053 32k INFO [2.6223254203796387, 1.9520517587661743, 4.571777820587158, 8.507990837097168, 0.792948305606842, 17000, 9.725082285098293e-05]
|
433 |
+
2023-02-10 05:18:50,420 32k INFO Saving model and optimizer state at iteration 224 to ./logs\32k\G_17000.pth
|
434 |
+
2023-02-10 05:19:09,845 32k INFO Saving model and optimizer state at iteration 224 to ./logs\32k\D_17000.pth
|
435 |
+
2023-02-10 05:19:33,204 32k INFO ====> Epoch: 224
|
436 |
+
2023-02-10 05:20:58,534 32k INFO ====> Epoch: 225
|
437 |
+
2023-02-10 05:22:23,931 32k INFO ====> Epoch: 226
|
438 |
+
2023-02-10 05:23:05,991 32k INFO Train Epoch: 227 [32%]
|
439 |
+
2023-02-10 05:23:05,991 32k INFO [2.3721673488616943, 2.349473476409912, 10.199991226196289, 15.970855712890625, 0.8750467300415039, 17200, 9.721435835085619e-05]
|
440 |
+
2023-02-10 05:23:48,977 32k INFO ====> Epoch: 227
|
441 |
+
2023-02-10 05:25:14,316 32k INFO ====> Epoch: 228
|
442 |
+
2023-02-10 05:26:35,734 32k INFO Train Epoch: 229 [95%]
|
443 |
+
2023-02-10 05:26:35,735 32k INFO [2.3928985595703125, 2.939208507537842, 17.157743453979492, 19.45735740661621, 0.6444438695907593, 17400, 9.719005628024282e-05]
|
444 |
+
2023-02-10 05:26:38,950 32k INFO ====> Epoch: 229
|
445 |
+
2023-02-10 05:28:04,250 32k INFO ====> Epoch: 230
|
446 |
+
2023-02-10 05:29:28,728 32k INFO ====> Epoch: 231
|
447 |
+
2023-02-10 05:30:28,074 32k INFO Train Epoch: 232 [58%]
|
448 |
+
2023-02-10 05:30:28,075 32k INFO [2.3620455265045166, 2.3724076747894287, 9.916801452636719, 15.793679237365723, 0.8138560652732849, 17600, 9.715361456473177e-05]
|
449 |
+
2023-02-10 05:30:54,438 32k INFO ====> Epoch: 232
|
450 |
+
2023-02-10 05:32:18,834 32k INFO ====> Epoch: 233
|
451 |
+
2023-02-10 05:33:43,436 32k INFO ====> Epoch: 234
|
452 |
+
2023-02-10 05:34:18,636 32k INFO Train Epoch: 235 [21%]
|
453 |
+
2023-02-10 05:34:18,636 32k INFO [2.421670913696289, 2.613611936569214, 10.162355422973633, 15.634803771972656, 0.8601481318473816, 17800, 9.711718651315591e-05]
|
454 |
+
2023-02-10 05:35:08,234 32k INFO ====> Epoch: 235
|
455 |
+
2023-02-10 05:36:32,671 32k INFO ====> Epoch: 236
|
456 |
+
2023-02-10 05:37:47,763 32k INFO Train Epoch: 237 [84%]
|
457 |
+
2023-02-10 05:37:47,764 32k INFO [2.3902554512023926, 2.546750783920288, 13.740601539611816, 15.754182815551758, 0.7283089756965637, 18000, 9.709290873398365e-05]
|
458 |
+
2023-02-10 05:37:52,211 32k INFO Saving model and optimizer state at iteration 237 to ./logs\32k\G_18000.pth
|
459 |
+
2023-02-10 05:38:05,355 32k INFO Saving model and optimizer state at iteration 237 to ./logs\32k\D_18000.pth
|
460 |
+
2023-02-10 05:38:18,717 32k INFO ====> Epoch: 237
|
461 |
+
2023-02-10 05:39:44,053 32k INFO ====> Epoch: 238
|
462 |
+
2023-02-10 05:41:09,330 32k INFO ====> Epoch: 239
|
463 |
+
2023-02-10 05:42:01,142 32k INFO Train Epoch: 240 [47%]
|
464 |
+
2023-02-10 05:42:01,143 32k INFO [2.312722682952881, 2.6106648445129395, 9.877479553222656, 15.368383407592773, 0.6470428109169006, 18200, 9.705650344424885e-05]
|
465 |
+
2023-02-10 05:42:34,235 32k INFO ====> Epoch: 240
|
466 |
+
2023-02-10 05:43:58,685 32k INFO ====> Epoch: 241
|
467 |
+
2023-02-10 05:45:23,999 32k INFO ====> Epoch: 242
|
468 |
+
2023-02-10 05:45:53,547 32k INFO Train Epoch: 243 [11%]
|
469 |
+
2023-02-10 05:45:53,547 32k INFO [2.3766026496887207, 2.4300293922424316, 12.744778633117676, 14.884074211120605, 0.6547593474388123, 18400, 9.702011180479129e-05]
|
470 |
+
2023-02-10 05:46:49,612 32k INFO ====> Epoch: 243
|
471 |
+
2023-02-10 05:48:15,106 32k INFO ====> Epoch: 244
|
472 |
+
2023-02-10 05:49:23,380 32k INFO Train Epoch: 245 [74%]
|
473 |
+
2023-02-10 05:49:23,380 32k INFO [2.580474376678467, 2.082796812057495, 9.039981842041016, 12.80046558380127, 0.5056825280189514, 18600, 9.699585829277933e-05]
|
474 |
+
2023-02-10 05:49:39,801 32k INFO ====> Epoch: 245
|
475 |
+
2023-02-10 05:51:04,233 32k INFO ====> Epoch: 246
|
476 |
+
2023-02-10 05:52:28,838 32k INFO ====> Epoch: 247
|
477 |
+
2023-02-10 05:53:14,110 32k INFO Train Epoch: 248 [37%]
|
478 |
+
2023-02-10 05:53:14,110 32k INFO [2.426253318786621, 2.482189416885376, 7.620111465454102, 7.9587178230285645, 0.5158562064170837, 18800, 9.695948939241093e-05]
|
479 |
+
2023-02-10 05:53:53,852 32k INFO ====> Epoch: 248
|
480 |
+
2023-02-10 05:55:18,546 32k INFO ====> Epoch: 249
|
481 |
+
2023-02-10 05:56:42,949 32k INFO ====> Epoch: 250
|
482 |
+
2023-02-10 05:57:05,033 32k INFO Train Epoch: 251 [0%]
|
483 |
+
2023-02-10 05:57:05,034 32k INFO [2.527033805847168, 2.064669370651245, 9.806694984436035, 12.288697242736816, 0.5984607934951782, 19000, 9.692313412867544e-05]
|
484 |
+
2023-02-10 05:57:09,464 32k INFO Saving model and optimizer state at iteration 251 to ./logs\32k\G_19000.pth
|
485 |
+
2023-02-10 05:57:24,997 32k INFO Saving model and optimizer state at iteration 251 to ./logs\32k\D_19000.pth
|
486 |
+
2023-02-10 05:58:31,373 32k INFO ====> Epoch: 251
|
487 |
+
2023-02-10 05:59:56,772 32k INFO ====> Epoch: 252
|
488 |
+
2023-02-10 06:00:59,380 32k INFO Train Epoch: 253 [63%]
|
489 |
+
2023-02-10 06:00:59,381 32k INFO [2.33793568611145, 2.2299203872680664, 7.934765815734863, 14.959351539611816, 0.5853436589241028, 19200, 9.689890485956725e-05]
|
490 |
+
2023-02-10 06:01:22,563 32k INFO ====> Epoch: 253
|
491 |
+
2023-02-10 06:02:47,113 32k INFO ====> Epoch: 254
|
492 |
+
2023-02-10 06:04:12,445 32k INFO ====> Epoch: 255
|
493 |
+
2023-02-10 06:04:51,929 32k INFO Train Epoch: 256 [26%]
|
494 |
+
2023-02-10 06:04:51,929 32k INFO [2.5420777797698975, 2.1400043964385986, 7.56562614440918, 11.221778869628906, 0.6662202477455139, 19400, 9.68625723121918e-05]
|
495 |
+
2023-02-10 06:05:38,131 32k INFO ====> Epoch: 256
|
496 |
+
2023-02-10 06:07:03,640 32k INFO ====> Epoch: 257
|
497 |
+
2023-02-10 06:08:22,686 32k INFO Train Epoch: 258 [89%]
|
498 |
+
2023-02-10 06:08:22,686 32k INFO [2.2824349403381348, 2.342818260192871, 11.228169441223145, 16.061767578125, 0.9010499119758606, 19600, 9.683835818259144e-05]
|
499 |
+
2023-02-10 06:08:29,233 32k INFO ====> Epoch: 258
|
500 |
+
2023-02-10 06:09:54,606 32k INFO ====> Epoch: 259
|
501 |
+
2023-02-10 06:11:19,916 32k INFO ====> Epoch: 260
|
502 |
+
2023-02-10 06:12:16,024 32k INFO Train Epoch: 261 [53%]
|
503 |
+
2023-02-10 06:12:16,024 32k INFO [2.2774453163146973, 2.862053394317627, 11.609835624694824, 18.45319366455078, 0.8751690983772278, 19800, 9.680204833738185e-05]
|
504 |
+
2023-02-10 06:12:45,792 32k INFO ====> Epoch: 261
|
505 |
+
2023-02-10 06:14:11,129 32k INFO ====> Epoch: 262
|
506 |
+
2023-02-10 06:15:35,580 32k INFO ====> Epoch: 263
|
507 |
+
2023-02-10 06:16:08,423 32k INFO Train Epoch: 264 [16%]
|
508 |
+
2023-02-10 06:16:08,423 32k INFO [2.32112455368042, 2.333401918411255, 9.849132537841797, 13.351881980895996, 0.8110747933387756, 20000, 9.676575210666227e-05]
|
509 |
+
2023-02-10 06:16:13,685 32k INFO Saving model and optimizer state at iteration 264 to ./logs\32k\G_20000.pth
|
510 |
+
2023-02-10 06:16:30,849 32k INFO Saving model and optimizer state at iteration 264 to ./logs\32k\D_20000.pth
|
511 |
+
2023-02-10 06:17:26,961 32k INFO ====> Epoch: 264
|
512 |
+
2023-02-10 06:18:52,371 32k INFO ====> Epoch: 265
|
513 |
+
2023-02-10 06:20:05,039 32k INFO Train Epoch: 266 [79%]
|
514 |
+
2023-02-10 06:20:05,039 32k INFO [2.4292171001434326, 2.405426025390625, 10.425787925720215, 15.707857131958008, 0.6626737117767334, 20200, 9.674156218060047e-05]
|
515 |
+
2023-02-10 06:20:18,267 32k INFO ====> Epoch: 266
|
516 |
+
2023-02-10 06:21:43,774 32k INFO ====> Epoch: 267
|
517 |
+
2023-02-10 06:23:08,283 32k INFO ====> Epoch: 268
|
518 |
+
2023-02-10 06:23:56,720 32k INFO Train Epoch: 269 [42%]
|
519 |
+
2023-02-10 06:23:56,720 32k INFO [2.2364232540130615, 2.6233410835266113, 12.21335506439209, 15.925817489624023, 0.2746255695819855, 20400, 9.670528862935451e-05]
|
520 |
+
2023-02-10 06:24:33,096 32k INFO ====> Epoch: 269
|
521 |
+
2023-02-10 06:25:57,635 32k INFO ====> Epoch: 270
|
522 |
+
2023-02-10 06:27:22,159 32k INFO ====> Epoch: 271
|
523 |
+
2023-02-10 06:27:47,572 32k INFO Train Epoch: 272 [5%]
|
524 |
+
2023-02-10 06:27:47,572 32k INFO [2.279311180114746, 2.69355845451355, 11.382664680480957, 15.993112564086914, 0.7590271830558777, 20600, 9.666902867899003e-05]
|
525 |
+
2023-02-10 06:28:46,950 32k INFO ====> Epoch: 272
|
526 |
+
2023-02-10 06:30:11,421 32k INFO ====> Epoch: 273
|
527 |
+
2023-02-10 06:31:16,437 32k INFO Train Epoch: 274 [68%]
|
528 |
+
2023-02-10 06:31:16,437 32k INFO [2.2094767093658447, 2.471757411956787, 11.495461463928223, 15.169349670410156, 0.6064764857292175, 20800, 9.664486293227385e-05]
|
529 |
+
2023-02-10 06:31:36,272 32k INFO ====> Epoch: 274
|
530 |
+
2023-02-10 06:33:00,711 32k INFO ====> Epoch: 275
|
531 |
+
2023-02-10 06:34:25,096 32k INFO ====> Epoch: 276
|
532 |
+
2023-02-10 06:35:06,908 32k INFO Train Epoch: 277 [32%]
|
533 |
+
2023-02-10 06:35:06,909 32k INFO [2.3740768432617188, 2.466357707977295, 10.599915504455566, 11.676641464233398, 0.681009829044342, 21000, 9.660862563871342e-05]
|
534 |
+
2023-02-10 06:35:11,320 32k INFO Saving model and optimizer state at iteration 277 to ./logs\32k\G_21000.pth
|
535 |
+
2023-02-10 06:35:28,883 32k INFO Saving model and optimizer state at iteration 277 to ./logs\32k\D_21000.pth
|
536 |
+
2023-02-10 06:36:15,635 32k INFO ====> Epoch: 277
|
537 |
+
2023-02-10 06:37:40,991 32k INFO ====> Epoch: 278
|
538 |
+
2023-02-10 06:39:03,277 32k INFO Train Epoch: 279 [95%]
|
539 |
+
2023-02-10 06:39:03,277 32k INFO [2.7325069904327393, 2.2234275341033936, 7.8317461013793945, 10.93569564819336, 0.5372100472450256, 21200, 9.658447499181352e-05]
|
540 |
+
2023-02-10 06:39:06,507 32k INFO ====> Epoch: 279
|
541 |
+
2023-02-10 06:40:31,921 32k INFO ====> Epoch: 280
|
542 |
+
2023-02-10 06:41:57,223 32k INFO ====> Epoch: 281
|
543 |
+
2023-02-10 06:42:55,547 32k INFO Train Epoch: 282 [58%]
|
544 |
+
2023-02-10 06:42:55,548 32k INFO [2.078672409057617, 3.186306953430176, 10.578144073486328, 14.06082534790039, 0.670340895652771, 21400, 9.65482603409002e-05]
|
545 |
+
2023-02-10 06:43:21,901 32k INFO ====> Epoch: 282
|
546 |
+
2023-02-10 06:44:46,353 32k INFO ====> Epoch: 283
|
547 |
+
2023-02-10 06:46:10,896 32k INFO ====> Epoch: 284
|
548 |
+
2023-02-10 06:46:46,078 32k INFO Train Epoch: 285 [21%]
|
549 |
+
2023-02-10 06:46:46,078 32k INFO [2.4770002365112305, 2.4508018493652344, 9.283089637756348, 15.895116806030273, 0.6614298224449158, 21600, 9.651205926878348e-05]
|
550 |
+
2023-02-10 06:47:35,591 32k INFO ====> Epoch: 285
|
551 |
+
2023-02-10 06:49:00,001 32k INFO ====> Epoch: 286
|
552 |
+
2023-02-10 06:50:14,886 32k INFO Train Epoch: 287 [84%]
|
553 |
+
2023-02-10 06:50:14,887 32k INFO [2.2746620178222656, 2.5889053344726562, 10.645721435546875, 13.369743347167969, 0.817758321762085, 21800, 9.64879327619672e-05]
|
554 |
+
2023-02-10 06:50:24,830 32k INFO ====> Epoch: 287
|
555 |
+
2023-02-10 06:51:50,242 32k INFO ====> Epoch: 288
|
556 |
+
2023-02-10 06:53:14,658 32k INFO ====> Epoch: 289
|
557 |
+
2023-02-10 06:54:06,355 32k INFO Train Epoch: 290 [47%]
|
558 |
+
2023-02-10 06:54:06,355 32k INFO [2.408446788787842, 2.3364787101745605, 10.12728214263916, 14.59122085571289, 0.7084507942199707, 22000, 9.645175430986486e-05]
|
559 |
+
2023-02-10 06:54:10,737 32k INFO Saving model and optimizer state at iteration 290 to ./logs\32k\G_22000.pth
|
560 |
+
2023-02-10 06:54:28,834 32k INFO Saving model and optimizer state at iteration 290 to ./logs\32k\D_22000.pth
|
561 |
+
2023-02-10 06:55:05,411 32k INFO ====> Epoch: 290
|
562 |
+
2023-02-10 06:56:30,627 32k INFO ====> Epoch: 291
|
563 |
+
2023-02-10 06:57:55,820 32k INFO ====> Epoch: 292
|
564 |
+
2023-02-10 06:58:24,516 32k INFO Train Epoch: 293 [11%]
|
565 |
+
2023-02-10 06:58:24,516 32k INFO [2.3272194862365723, 2.602001190185547, 9.372489929199219, 11.096156120300293, 0.27660641074180603, 22200, 9.641558942298625e-05]
|
566 |
+
2023-02-10 06:59:20,776 32k INFO ====> Epoch: 293
|
567 |
+
2023-02-10 07:00:45,983 32k INFO ====> Epoch: 294
|
568 |
+
2023-02-10 07:01:55,037 32k INFO Train Epoch: 295 [74%]
|
569 |
+
2023-02-10 07:01:55,037 32k INFO [2.390826463699341, 2.554483413696289, 8.697030067443848, 14.283463478088379, 0.6869368553161621, 22400, 9.639148703212408e-05]
|
570 |
+
2023-02-10 07:02:11,478 32k INFO ====> Epoch: 295
|
571 |
+
2023-02-10 07:03:36,836 32k INFO ====> Epoch: 296
|
572 |
+
2023-02-10 07:05:02,243 32k INFO ====> Epoch: 297
|
573 |
+
2023-02-10 07:05:48,212 32k INFO Train Epoch: 298 [37%]
|
574 |
+
2023-02-10 07:05:48,213 32k INFO [2.413754940032959, 2.9149270057678223, 9.995668411254883, 12.166794776916504, 0.653400182723999, 22600, 9.635534474264972e-05]
|
575 |
+
2023-02-10 07:06:27,766 32k INFO ====> Epoch: 298
|
576 |
+
2023-02-10 07:07:52,162 32k INFO ====> Epoch: 299
|
577 |
+
2023-02-10 07:09:16,689 32k INFO ====> Epoch: 300
|
578 |
+
2023-02-10 07:09:38,796 32k INFO Train Epoch: 301 [0%]
|
579 |
+
2023-02-10 07:09:38,796 32k INFO [2.2665956020355225, 2.7774581909179688, 9.704061508178711, 13.352103233337402, 0.712449312210083, 22800, 9.631921600483981e-05]
|
580 |
+
2023-02-10 07:10:41,520 32k INFO ====> Epoch: 301
|
581 |
+
2023-02-10 07:12:06,795 32k INFO ====> Epoch: 302
|
582 |
+
2023-02-10 07:13:08,471 32k INFO Train Epoch: 303 [63%]
|
583 |
+
2023-02-10 07:13:08,471 32k INFO [2.1530864238739014, 2.413203477859497, 11.726449012756348, 18.033466339111328, 0.6967122554779053, 23000, 9.629513770582634e-05]
|
584 |
+
2023-02-10 07:13:12,906 32k INFO Saving model and optimizer state at iteration 303 to ./logs\32k\G_23000.pth
|
585 |
+
2023-02-10 07:13:28,702 32k INFO Saving model and optimizer state at iteration 303 to ./logs\32k\D_23000.pth
|
586 |
+
2023-02-10 07:13:55,451 32k INFO ====> Epoch: 303
|
587 |
+
2023-02-10 07:15:20,800 32k INFO ====> Epoch: 304
|
588 |
+
2023-02-10 07:16:46,057 32k INFO ====> Epoch: 305
|
589 |
+
2023-02-10 07:17:25,604 32k INFO Train Epoch: 306 [26%]
|
590 |
+
2023-02-10 07:17:25,605 32k INFO [2.3469924926757812, 2.375382661819458, 9.387177467346191, 12.057092666625977, -0.1277291625738144, 23200, 9.625903154283315e-05]
|
591 |
+
2023-02-10 07:18:11,964 32k INFO ====> Epoch: 306
|
592 |
+
2023-02-10 07:19:36,400 32k INFO ====> Epoch: 307
|
593 |
+
2023-02-10 07:20:54,517 32k INFO Train Epoch: 308 [89%]
|
594 |
+
2023-02-10 07:20:54,517 32k INFO [2.3641486167907715, 2.28696608543396, 9.959922790527344, 17.36892318725586, 0.7711434364318848, 23400, 9.62349682889948e-05]
|
595 |
+
2023-02-10 07:21:01,068 32k INFO ====> Epoch: 308
|
596 |
+
2023-02-10 07:22:25,526 32k INFO ====> Epoch: 309
|
597 |
+
2023-02-10 07:23:50,156 32k INFO ====> Epoch: 310
|
598 |
+
2023-02-10 07:24:45,168 32k INFO Train Epoch: 311 [53%]
|
599 |
+
2023-02-10 07:24:45,169 32k INFO [2.5586600303649902, 2.3868942260742188, 8.270402908325195, 11.41346549987793, 0.6984982490539551, 23600, 9.619888468671259e-05]
|
600 |
+
2023-02-10 07:25:14,874 32k INFO ====> Epoch: 311
|
601 |
+
2023-02-10 07:26:39,243 32k INFO ====> Epoch: 312
|
602 |
+
2023-02-10 07:28:03,612 32k INFO ====> Epoch: 313
|
603 |
+
2023-02-10 07:28:35,612 32k INFO Train Epoch: 314 [16%]
|
604 |
+
2023-02-10 07:28:35,613 32k INFO [1.8997191190719604, 2.9339983463287354, 9.256068229675293, 12.056029319763184, 0.7804506421089172, 23800, 9.61628146140899e-05]
|
605 |
+
2023-02-10 07:29:28,369 32k INFO ====> Epoch: 314
|
606 |
+
2023-02-10 07:30:52,694 32k INFO ====> Epoch: 315
|
607 |
+
2023-02-10 07:32:04,287 32k INFO Train Epoch: 316 [79%]
|
608 |
+
2023-02-10 07:32:04,287 32k INFO [2.4035465717315674, 2.7094507217407227, 9.096111297607422, 14.042303085327148, 0.8556358814239502, 24000, 9.613877541298036e-05]
|
609 |
+
2023-02-10 07:32:08,663 32k INFO Saving model and optimizer state at iteration 316 to ./logs\32k\G_24000.pth
|
610 |
+
2023-02-10 07:32:24,726 32k INFO Saving model and optimizer state at iteration 316 to ./logs\32k\D_24000.pth
|
611 |
+
2023-02-10 07:32:41,400 32k INFO ====> Epoch: 316
|
612 |
+
2023-02-10 07:34:06,691 32k INFO ====> Epoch: 317
|