shenyunhang commited on
Commit
0763970
·
1 Parent(s): 077821d
vita_tts_ckpt/codec/final.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c04d618827b0e2778280d7c4701d6c8450c4104de4c1a1dca2b5c3120017c7a
3
+ size 253718273
vita_tts_ckpt/codec/model.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "resblock": "1",
3
+ "num_gpus": 8,
4
+ "batch_size": 160,
5
+ "learning_rate": 0.0002,
6
+ "adam_b1": 0.5,
7
+ "adam_b2": 0.9,
8
+ "lr_decay": 0.98,
9
+ "seed": 1234,
10
+
11
+ "upsample_rates": [8,5,5,3],
12
+ "upsample_kernel_sizes": [16,11,11,5],
13
+ "upsample_initial_channel": 512,
14
+ "resblock_kernel_sizes": [3,7,11],
15
+ "resblock_dilation_sizes": [[1,3,5], [1,3,5], [1,3,5]],
16
+
17
+ "segment_size": 24000,
18
+ "num_mels": 80,
19
+ "num_freq": 1025,
20
+ "n_fft": 1024,
21
+ "hop_size": 240,
22
+ "win_size": 1024,
23
+
24
+ "sampling_rate": 24000,
25
+
26
+ "n_code_groups": 1,
27
+ "residul_layer": 1,
28
+ "n_codes": 1024,
29
+ "codebook_loss_lambda": 1.0,
30
+ "commitment_loss_lambda": 0.25,
31
+ "global_code_num": 8,
32
+ "global_feature_conv":[128, 64, 128, 3, 1],
33
+ "global_tokens": [473,975,419,219,565,121,550,616],
34
+
35
+ "fmin": 0,
36
+ "fmax": 8000,
37
+ "fmax_for_loss": null,
38
+
39
+ "num_workers": 12
40
+ }
vita_tts_ckpt/decoder/final.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d3274da94685758ce37209759ac7542ea8f2d6b47c00d1cc18c051d0d33cc3e
3
+ size 1194900847
vita_tts_ckpt/decoder/model.json ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ 896,
3
+ 1024,
4
+ {
5
+ "accum_grad": 3,
6
+ "char_list": [],
7
+ "debugmode": 0,
8
+ "encoder_criterion": "ce",
9
+ "encoder_drop_rate": 0.1,
10
+ "encoder_input_dim": 896,
11
+ "encoder_layer_config": "transformer",
12
+ "encoder_output_dim": 896,
13
+ "encoder_pre_norm_type": "ln",
14
+ "encoder_upsample_rate": 9,
15
+ "kv_cache_prefix_finetune": 0,
16
+ "epochs": 100,
17
+ "eps": 1e-08,
18
+ "eps_decay": 0.8,
19
+ "gpu_id": null,
20
+ "gpu_num": 1,
21
+ "grad_clip": 5,
22
+ "grad_noise": false,
23
+ "idim": 896,
24
+ "init_lr": 0.0005,
25
+ "lsm_weight": 0.0,
26
+ "max_batch_size": 25,
27
+ "max_duration": 256,
28
+ "max_mem": 20000,
29
+ "mtlalpha": 0.5,
30
+ "n_iter_processes": 8,
31
+ "noam_warmup_steps": 4000,
32
+ "odim": 1024,
33
+ "opt": "noamw",
34
+ "rank": 0,
35
+ "report_interval_iters": 100,
36
+ "resume_trainer": false,
37
+ "save_interval_iters": 2000,
38
+ "seed": 19832,
39
+ "sort_duration": true,
40
+ "start_decay_epoch": 5,
41
+ "stop_learning_rate": 1e-05,
42
+ "sycn_batchnorm": false,
43
+ "tensorboard_dir": null,
44
+ "train_dtype": "bfloat16",
45
+ "transformer_attention_dim": 896,
46
+ "transformer_attention_dropout_rate": 0.1,
47
+ "transformer_attention_heads": 14,
48
+ "transformer_chunk_size": [
49
+ 1
50
+ ],
51
+ "transformer_concat_after": false,
52
+ "transformer_dropout_rate": 0.1,
53
+ "transformer_dynamic_chunks": false,
54
+ "transformer_input_dim": 896,
55
+ "transformer_input_layer": "linear",
56
+ "transformer_left_chunks": [
57
+ -1
58
+ ],
59
+ "transformer_linear_units": 4864,
60
+ "transformer_normalize_before": true,
61
+ "transformer_num_blocks": 4,
62
+ "transformer_output_dim": 896,
63
+ "transformer_pos_enc_class": "rel-enc",
64
+ "transformer_positional_dropout_rate": 0.1,
65
+ "transformer_positionwise_conv_kernel_size": 1,
66
+ "transformer_positionwise_layer_type": "linear",
67
+ "use_zero_redun_opt": false,
68
+ "verbose": 0,
69
+ "weight_decay": 0.05,
70
+ "world_size": 1
71
+ }
72
+ ]