David Portes
commited on
Commit
•
e40b4f0
1
Parent(s):
7b5769f
model_nospace_binary
Browse files- CKPT.yaml +7 -7
- brain.ckpt +1 -1
- counter.ckpt +1 -1
- hyperparams.yaml +12 -21
- model.ckpt +1 -1
- optimizer.ckpt +2 -2
- scheduler.ckpt +0 -0
CKPT.yaml
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
# yamllint disable
|
2 |
-
attn_loss: 0.
|
3 |
-
attn_weight:
|
4 |
end-of-epoch: true
|
5 |
-
epoch:
|
6 |
-
gate_loss: 0.
|
7 |
-
loss: 0.
|
8 |
-
mel_loss: 0.
|
9 |
-
unixtime:
|
|
|
1 |
# yamllint disable
|
2 |
+
attn_loss: 0.0
|
3 |
+
attn_weight: 0.0
|
4 |
end-of-epoch: true
|
5 |
+
epoch: 128
|
6 |
+
gate_loss: 0.0010953631717711687
|
7 |
+
loss: 0.44426456093788147
|
8 |
+
mel_loss: 0.4431692063808441
|
9 |
+
unixtime: 1665297109.2974067
|
brain.ckpt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
avg_train_loss: 0.0
|
2 |
-
optimizer_step:
|
3 |
step: 0
|
|
|
1 |
avg_train_loss: 0.0
|
2 |
+
optimizer_step: 23552
|
3 |
step: 0
|
counter.ckpt
CHANGED
@@ -1 +1 @@
|
|
1 |
-
|
|
|
1 |
+
128
|
hyperparams.yaml
CHANGED
@@ -13,11 +13,11 @@
|
|
13 |
###################################
|
14 |
# Experiment Parameters and setup #
|
15 |
###################################
|
16 |
-
seed:
|
17 |
-
__set_seed: !apply:torch.manual_seed [
|
18 |
-
output_folder: ./results/tacotron2/
|
19 |
-
save_folder: ./results/tacotron2/
|
20 |
-
train_log: ./results/tacotron2/
|
21 |
epochs: 750
|
22 |
keep_checkpoint_interval: 50
|
23 |
|
@@ -32,7 +32,7 @@ keep_checkpoint_interval: 50
|
|
32 |
progress_samples: true
|
33 |
|
34 |
# The path where the samples will be stored
|
35 |
-
progress_sample_path: ./results/tacotron2/
|
36 |
# The interval, in epochs. For instance, if it is set to 5,
|
37 |
# progress samples will be output every 5 epochs
|
38 |
progress_samples_interval: 1
|
@@ -46,9 +46,9 @@ progress_batch_sample_size: 3
|
|
46 |
data_folder: ../../../../LJSpeech-1.1
|
47 |
# e.g, /localscratch/ljspeech
|
48 |
|
49 |
-
train_json: ./results/tacotron2/
|
50 |
-
valid_json: ./results/tacotron2/
|
51 |
-
test_json: ./results/tacotron2/
|
52 |
|
53 |
splits: [train, valid]
|
54 |
split_ratio: [90, 10]
|
@@ -208,7 +208,7 @@ epoch_counter: &id003 !new:speechbrain.utils.epoch_loop.EpochCounter
|
|
208 |
limit: 750
|
209 |
|
210 |
train_logger: !new:speechbrain.utils.train_logger.FileTrainLogger
|
211 |
-
save_file: ./results/tacotron2/
|
212 |
|
213 |
#annealing_function
|
214 |
lr_annealing: &id004 !new:speechbrain.nnet.schedulers.IntervalScheduler
|
@@ -225,22 +225,13 @@ lr_annealing: &id004 !new:speechbrain.nnet.schedulers.IntervalScheduler
|
|
225 |
|
226 |
#checkpointer
|
227 |
checkpointer: !new:speechbrain.utils.checkpoints.Checkpointer
|
228 |
-
checkpoints_dir: ./results/tacotron2/
|
229 |
recoverables:
|
230 |
model: *id002
|
231 |
counter: *id003
|
232 |
scheduler: *id004
|
233 |
progress_sample_logger: !new:speechbrain.utils.train_logger.ProgressSampleLogger
|
234 |
-
output_path: ./results/tacotron2/
|
235 |
batch_sample_size: 3
|
236 |
formats:
|
237 |
raw_batch: raw
|
238 |
-
|
239 |
-
|
240 |
-
text_to_sequence: !name:speechbrain.utils.text_to_sequence.text_to_sequence
|
241 |
-
|
242 |
-
pretrainer: !new:speechbrain.utils.parameter_transfer.Pretrainer
|
243 |
-
loadables:
|
244 |
-
model: !ref <model>
|
245 |
-
paths:
|
246 |
-
model: Daporte/speechbrain_tacotron2_exp/model.ckpt
|
|
|
13 |
###################################
|
14 |
# Experiment Parameters and setup #
|
15 |
###################################
|
16 |
+
seed: 9234
|
17 |
+
__set_seed: !apply:torch.manual_seed [9234]
|
18 |
+
output_folder: ./results/tacotron2/9234
|
19 |
+
save_folder: ./results/tacotron2/9234/save
|
20 |
+
train_log: ./results/tacotron2/9234/train_log.txt
|
21 |
epochs: 750
|
22 |
keep_checkpoint_interval: 50
|
23 |
|
|
|
32 |
progress_samples: true
|
33 |
|
34 |
# The path where the samples will be stored
|
35 |
+
progress_sample_path: ./results/tacotron2/9234/samples
|
36 |
# The interval, in epochs. For instance, if it is set to 5,
|
37 |
# progress samples will be output every 5 epochs
|
38 |
progress_samples_interval: 1
|
|
|
46 |
data_folder: ../../../../LJSpeech-1.1
|
47 |
# e.g, /localscratch/ljspeech
|
48 |
|
49 |
+
train_json: ./results/tacotron2/9234/save/train.json
|
50 |
+
valid_json: ./results/tacotron2/9234/save/valid.json
|
51 |
+
test_json: ./results/tacotron2/9234/save/test.json
|
52 |
|
53 |
splits: [train, valid]
|
54 |
split_ratio: [90, 10]
|
|
|
208 |
limit: 750
|
209 |
|
210 |
train_logger: !new:speechbrain.utils.train_logger.FileTrainLogger
|
211 |
+
save_file: ./results/tacotron2/9234/train_log.txt
|
212 |
|
213 |
#annealing_function
|
214 |
lr_annealing: &id004 !new:speechbrain.nnet.schedulers.IntervalScheduler
|
|
|
225 |
|
226 |
#checkpointer
|
227 |
checkpointer: !new:speechbrain.utils.checkpoints.Checkpointer
|
228 |
+
checkpoints_dir: ./results/tacotron2/9234/save
|
229 |
recoverables:
|
230 |
model: *id002
|
231 |
counter: *id003
|
232 |
scheduler: *id004
|
233 |
progress_sample_logger: !new:speechbrain.utils.train_logger.ProgressSampleLogger
|
234 |
+
output_path: ./results/tacotron2/9234/samples
|
235 |
batch_sample_size: 3
|
236 |
formats:
|
237 |
raw_batch: raw
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
model.ckpt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 112834302
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7b18310779c8d32e1f34eff19600a83843a67be7d599341ef9fef2ef985604ac
|
3 |
size 112834302
|
optimizer.ckpt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9804a404d098e71c62b71cd8b2a31cdae83bd26ff3ab7fe7a1c7ec969b7fac45
|
3 |
+
size 225603417
|
scheduler.ckpt
CHANGED
Binary files a/scheduler.ckpt and b/scheduler.ckpt differ
|
|