David Portes commited on
Commit
e40b4f0
1 Parent(s): 7b5769f

model_nospace_binary

Browse files
Files changed (7) hide show
  1. CKPT.yaml +7 -7
  2. brain.ckpt +1 -1
  3. counter.ckpt +1 -1
  4. hyperparams.yaml +12 -21
  5. model.ckpt +1 -1
  6. optimizer.ckpt +2 -2
  7. scheduler.ckpt +0 -0
CKPT.yaml CHANGED
@@ -1,9 +1,9 @@
1
  # yamllint disable
2
- attn_loss: 0.004663344006985426
3
- attn_weight: 35.35533905029297
4
  end-of-epoch: true
5
- epoch: 5
6
- gate_loss: 0.0033020121045410633
7
- loss: 0.8078759908676147
8
- mel_loss: 0.7999106049537659
9
- unixtime: 1665210947.8861418
 
1
  # yamllint disable
2
+ attn_loss: 0.0
3
+ attn_weight: 0.0
4
  end-of-epoch: true
5
+ epoch: 128
6
+ gate_loss: 0.0010953631717711687
7
+ loss: 0.44426456093788147
8
+ mel_loss: 0.4431692063808441
9
+ unixtime: 1665297109.2974067
brain.ckpt CHANGED
@@ -1,3 +1,3 @@
1
  avg_train_loss: 0.0
2
- optimizer_step: 920
3
  step: 0
 
1
  avg_train_loss: 0.0
2
+ optimizer_step: 23552
3
  step: 0
counter.ckpt CHANGED
@@ -1 +1 @@
1
- 5
 
1
+ 128
hyperparams.yaml CHANGED
@@ -13,11 +13,11 @@
13
  ###################################
14
  # Experiment Parameters and setup #
15
  ###################################
16
- seed: 7234
17
- __set_seed: !apply:torch.manual_seed [7234]
18
- output_folder: ./results/tacotron2/7234
19
- save_folder: ./results/tacotron2/7234/save
20
- train_log: ./results/tacotron2/7234/train_log.txt
21
  epochs: 750
22
  keep_checkpoint_interval: 50
23
 
@@ -32,7 +32,7 @@ keep_checkpoint_interval: 50
32
  progress_samples: true
33
 
34
  # The path where the samples will be stored
35
- progress_sample_path: ./results/tacotron2/7234/samples
36
  # The interval, in epochs. For instance, if it is set to 5,
37
  # progress samples will be output every 5 epochs
38
  progress_samples_interval: 1
@@ -46,9 +46,9 @@ progress_batch_sample_size: 3
46
  data_folder: ../../../../LJSpeech-1.1
47
  # e.g, /localscratch/ljspeech
48
 
49
- train_json: ./results/tacotron2/7234/save/train.json
50
- valid_json: ./results/tacotron2/7234/save/valid.json
51
- test_json: ./results/tacotron2/7234/save/test.json
52
 
53
  splits: [train, valid]
54
  split_ratio: [90, 10]
@@ -208,7 +208,7 @@ epoch_counter: &id003 !new:speechbrain.utils.epoch_loop.EpochCounter
208
  limit: 750
209
 
210
  train_logger: !new:speechbrain.utils.train_logger.FileTrainLogger
211
- save_file: ./results/tacotron2/7234/train_log.txt
212
 
213
  #annealing_function
214
  lr_annealing: &id004 !new:speechbrain.nnet.schedulers.IntervalScheduler
@@ -225,22 +225,13 @@ lr_annealing: &id004 !new:speechbrain.nnet.schedulers.IntervalScheduler
225
 
226
  #checkpointer
227
  checkpointer: !new:speechbrain.utils.checkpoints.Checkpointer
228
- checkpoints_dir: ./results/tacotron2/7234/save
229
  recoverables:
230
  model: *id002
231
  counter: *id003
232
  scheduler: *id004
233
  progress_sample_logger: !new:speechbrain.utils.train_logger.ProgressSampleLogger
234
- output_path: ./results/tacotron2/7234/samples
235
  batch_sample_size: 3
236
  formats:
237
  raw_batch: raw
238
-
239
-
240
- text_to_sequence: !name:speechbrain.utils.text_to_sequence.text_to_sequence
241
-
242
- pretrainer: !new:speechbrain.utils.parameter_transfer.Pretrainer
243
- loadables:
244
- model: !ref <model>
245
- paths:
246
- model: Daporte/speechbrain_tacotron2_exp/model.ckpt
 
13
  ###################################
14
  # Experiment Parameters and setup #
15
  ###################################
16
+ seed: 9234
17
+ __set_seed: !apply:torch.manual_seed [9234]
18
+ output_folder: ./results/tacotron2/9234
19
+ save_folder: ./results/tacotron2/9234/save
20
+ train_log: ./results/tacotron2/9234/train_log.txt
21
  epochs: 750
22
  keep_checkpoint_interval: 50
23
 
 
32
  progress_samples: true
33
 
34
  # The path where the samples will be stored
35
+ progress_sample_path: ./results/tacotron2/9234/samples
36
  # The interval, in epochs. For instance, if it is set to 5,
37
  # progress samples will be output every 5 epochs
38
  progress_samples_interval: 1
 
46
  data_folder: ../../../../LJSpeech-1.1
47
  # e.g, /localscratch/ljspeech
48
 
49
+ train_json: ./results/tacotron2/9234/save/train.json
50
+ valid_json: ./results/tacotron2/9234/save/valid.json
51
+ test_json: ./results/tacotron2/9234/save/test.json
52
 
53
  splits: [train, valid]
54
  split_ratio: [90, 10]
 
208
  limit: 750
209
 
210
  train_logger: !new:speechbrain.utils.train_logger.FileTrainLogger
211
+ save_file: ./results/tacotron2/9234/train_log.txt
212
 
213
  #annealing_function
214
  lr_annealing: &id004 !new:speechbrain.nnet.schedulers.IntervalScheduler
 
225
 
226
  #checkpointer
227
  checkpointer: !new:speechbrain.utils.checkpoints.Checkpointer
228
+ checkpoints_dir: ./results/tacotron2/9234/save
229
  recoverables:
230
  model: *id002
231
  counter: *id003
232
  scheduler: *id004
233
  progress_sample_logger: !new:speechbrain.utils.train_logger.ProgressSampleLogger
234
+ output_path: ./results/tacotron2/9234/samples
235
  batch_sample_size: 3
236
  formats:
237
  raw_batch: raw
 
 
 
 
 
 
 
 
 
model.ckpt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:35bc1da9ae78fa7d0ffa8a7771e3bc1fde52bde5b51de7596c8b3aa2c262618f
3
  size 112834302
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b18310779c8d32e1f34eff19600a83843a67be7d599341ef9fef2ef985604ac
3
  size 112834302
optimizer.ckpt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:459bc00d79101be7dc2d458f3435baf1abd332033d68e6f624ec36247dad7124
3
- size 225602585
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9804a404d098e71c62b71cd8b2a31cdae83bd26ff3ab7fe7a1c7ec969b7fac45
3
+ size 225603417
scheduler.ckpt CHANGED
Binary files a/scheduler.ckpt and b/scheduler.ckpt differ