ayymen commited on
Commit
797fb2a
·
verified ·
1 Parent(s): 3d59b6c

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ GPT_XTTS_LUO_FT-February-23-2025_07+52PM-8e59ec3/trainer_0_log.txt filter=lfs diff=lfs merge=lfs -text
GPT_XTTS_LUO_FT-February-23-2025_07+52PM-8e59ec3/best_model.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa490cc0fb7016929e73bc33df1501aefc27930154b6c31c471439ffd72f9d8c
3
+ size 5647440173
GPT_XTTS_LUO_FT-February-23-2025_07+52PM-8e59ec3/best_model_1736086.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa490cc0fb7016929e73bc33df1501aefc27930154b6c31c471439ffd72f9d8c
3
+ size 5647440173
GPT_XTTS_LUO_FT-February-23-2025_07+52PM-8e59ec3/checkpoint_1730000.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5829100335bb78e23f96352a1ade4a00038a42365413bd60c6db169fc15d1ba9
3
+ size 5647440173
GPT_XTTS_LUO_FT-February-23-2025_07+52PM-8e59ec3/checkpoint_1740000.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e2d93a076f01e0eb98ec898d65332e37618318729de3535c8d00c07f50cf1b4
3
+ size 5647440173
GPT_XTTS_LUO_FT-February-23-2025_07+52PM-8e59ec3/config.json ADDED
@@ -0,0 +1,215 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "output_path": "xtts_luo",
3
+ "logger_uri": null,
4
+ "run_name": "GPT_XTTS_LUO_FT",
5
+ "project_name": "XTTS_trainer",
6
+ "run_description": [
7
+ "\n GPT XTTS training\n "
8
+ ],
9
+ "print_step": 50,
10
+ "plot_step": 100,
11
+ "model_param_stats": false,
12
+ "wandb_entity": null,
13
+ "dashboard_logger": "tensorboard",
14
+ "save_on_interrupt": true,
15
+ "log_model_step": 100,
16
+ "save_step": 10000,
17
+ "save_n_checkpoints": 2,
18
+ "save_checkpoints": true,
19
+ "save_all_best": false,
20
+ "save_best_after": 0,
21
+ "target_loss": null,
22
+ "print_eval": false,
23
+ "test_delay_epochs": 0,
24
+ "run_eval": true,
25
+ "run_eval_steps": null,
26
+ "distributed_backend": "nccl",
27
+ "distributed_url": "tcp://localhost:54321",
28
+ "mixed_precision": true,
29
+ "precision": "bf16",
30
+ "epochs": 1000,
31
+ "batch_size": 1,
32
+ "eval_batch_size": 1,
33
+ "grad_clip": 0.0,
34
+ "scheduler_after_epoch": true,
35
+ "lr": 5e-06,
36
+ "optimizer": "AdamW",
37
+ "optimizer_params": {
38
+ "betas": [
39
+ 0.9,
40
+ 0.96
41
+ ],
42
+ "eps": 1e-08,
43
+ "weight_decay": 0.01
44
+ },
45
+ "lr_scheduler": "MultiStepLR",
46
+ "lr_scheduler_params": {
47
+ "milestones": [
48
+ 5000,
49
+ 150000,
50
+ 300000
51
+ ],
52
+ "gamma": 0.5,
53
+ "last_epoch": -1
54
+ },
55
+ "use_grad_scaler": false,
56
+ "allow_tf32": false,
57
+ "cudnn_enable": true,
58
+ "cudnn_deterministic": false,
59
+ "cudnn_benchmark": false,
60
+ "training_seed": 54321,
61
+ "model": "xtts",
62
+ "num_loader_workers": 8,
63
+ "num_eval_loader_workers": 0,
64
+ "use_noise_augment": false,
65
+ "audio": {
66
+ "sample_rate": 22050,
67
+ "output_sample_rate": 24000,
68
+ "dvae_sample_rate": 22050
69
+ },
70
+ "use_phonemes": false,
71
+ "phonemizer": null,
72
+ "phoneme_language": null,
73
+ "compute_input_seq_cache": false,
74
+ "text_cleaner": null,
75
+ "enable_eos_bos_chars": false,
76
+ "test_sentences_file": "",
77
+ "phoneme_cache_path": null,
78
+ "characters": null,
79
+ "add_blank": false,
80
+ "batch_group_size": 0,
81
+ "loss_masking": null,
82
+ "min_audio_len": 1,
83
+ "max_audio_len": Infinity,
84
+ "min_text_len": 1,
85
+ "max_text_len": Infinity,
86
+ "compute_f0": false,
87
+ "compute_energy": false,
88
+ "compute_linear_spec": false,
89
+ "precompute_num_workers": 0,
90
+ "start_by_longest": false,
91
+ "shuffle": false,
92
+ "drop_last": false,
93
+ "datasets": [
94
+ {
95
+ "formatter": "",
96
+ "dataset_name": "",
97
+ "path": "",
98
+ "meta_file_train": "",
99
+ "ignored_speakers": null,
100
+ "language": "",
101
+ "phonemizer": "",
102
+ "meta_file_val": "",
103
+ "meta_file_attn_mask": ""
104
+ }
105
+ ],
106
+ "test_sentences": [
107
+ {
108
+ "text": "jo kolosai achiel.",
109
+ "speaker_wav": [
110
+ "/app/data/clips/1CH_009_031.wav"
111
+ ],
112
+ "language": "luo"
113
+ },
114
+ {
115
+ "text": "magoyo erokamano ni wuoro ka un gi mor.",
116
+ "speaker_wav": [
117
+ "/app/data/clips/1CH_009_031.wav"
118
+ ],
119
+ "language": "luo"
120
+ },
121
+ {
122
+ "text": "epafra bende nonyisowa kuom hera ma roho maler osemiyou.",
123
+ "speaker_wav": [
124
+ "/app/data/clips/1CH_009_031.wav"
125
+ ],
126
+ "language": "luo"
127
+ }
128
+ ],
129
+ "eval_split_max_size": null,
130
+ "eval_split_size": 0.01,
131
+ "use_speaker_weighted_sampler": false,
132
+ "speaker_weighted_sampler_alpha": 1.0,
133
+ "use_language_weighted_sampler": false,
134
+ "language_weighted_sampler_alpha": 1.0,
135
+ "use_length_weighted_sampler": false,
136
+ "length_weighted_sampler_alpha": 1.0,
137
+ "model_args": {
138
+ "gpt_batch_size": 1,
139
+ "enable_redaction": false,
140
+ "kv_cache": true,
141
+ "gpt_checkpoint": "",
142
+ "clvp_checkpoint": null,
143
+ "decoder_checkpoint": null,
144
+ "num_chars": 255,
145
+ "tokenizer_file": "xtts_luo/XTTS_v2.0_original_model_files/vocab.json",
146
+ "gpt_max_audio_tokens": 605,
147
+ "gpt_max_text_tokens": 402,
148
+ "gpt_max_prompt_tokens": 70,
149
+ "gpt_layers": 30,
150
+ "gpt_n_model_channels": 1024,
151
+ "gpt_n_heads": 16,
152
+ "gpt_number_text_tokens": 8288,
153
+ "gpt_start_text_token": 261,
154
+ "gpt_stop_text_token": 0,
155
+ "gpt_num_audio_tokens": 1026,
156
+ "gpt_start_audio_token": 1024,
157
+ "gpt_stop_audio_token": 1025,
158
+ "gpt_code_stride_len": 1024,
159
+ "gpt_use_masking_gt_prompt_approach": true,
160
+ "gpt_use_perceiver_resampler": true,
161
+ "input_sample_rate": 22050,
162
+ "output_sample_rate": 24000,
163
+ "output_hop_length": 256,
164
+ "decoder_input_dim": 1024,
165
+ "d_vector_dim": 512,
166
+ "cond_d_vector_in_each_upsampling_layer": true,
167
+ "duration_const": 102400,
168
+ "min_conditioning_length": 11025,
169
+ "max_conditioning_length": 132300,
170
+ "gpt_loss_text_ce_weight": 0.01,
171
+ "gpt_loss_mel_ce_weight": 1.0,
172
+ "debug_loading_failures": true,
173
+ "max_wav_length": 264600,
174
+ "max_text_length": 300,
175
+ "mel_norm_file": "xtts_luo/XTTS_v2.0_original_model_files/mel_stats.pth",
176
+ "dvae_checkpoint": "xtts_luo/XTTS_v2.0_original_model_files/dvae.pth",
177
+ "xtts_checkpoint": "xtts_luo/XTTS_v2.0_original_model_files/model.pth",
178
+ "vocoder": ""
179
+ },
180
+ "model_dir": null,
181
+ "languages": [
182
+ "en",
183
+ "es",
184
+ "fr",
185
+ "de",
186
+ "it",
187
+ "pt",
188
+ "pl",
189
+ "tr",
190
+ "ru",
191
+ "nl",
192
+ "cs",
193
+ "ar",
194
+ "zh-cn",
195
+ "hu",
196
+ "ko",
197
+ "ja",
198
+ "hi",
199
+ "luo"
200
+ ],
201
+ "temperature": 0.75,
202
+ "length_penalty": 1.0,
203
+ "repetition_penalty": 5.0,
204
+ "top_k": 50,
205
+ "top_p": 0.85,
206
+ "num_gpt_outputs": 1,
207
+ "gpt_cond_len": 30,
208
+ "gpt_cond_chunk_len": 4,
209
+ "max_ref_len": 30,
210
+ "sound_norm_refs": false,
211
+ "optimizer_wd_only_on_weights": true,
212
+ "weighted_loss_attrs": {},
213
+ "weighted_loss_multipliers": {},
214
+ "github_branch": "* main"
215
+ }
GPT_XTTS_LUO_FT-February-23-2025_07+52PM-8e59ec3/events.out.tfevents.1740340368.815d4e6ec33e.1.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f433ef41e47ed43f98c09cdabb6f61d869d34e4a6fd1c1d55045f6ce9a847b2
3
+ size 30766257
GPT_XTTS_LUO_FT-February-23-2025_07+52PM-8e59ec3/train_gpt_xtts.py ADDED
@@ -0,0 +1,204 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os
2
+
3
+ from trainer import Trainer, TrainerArgs
4
+
5
+ from TTS.config.shared_configs import BaseDatasetConfig
6
+ from TTS.tts.datasets import load_tts_samples
7
+ from TTS.tts.layers.xtts.trainer.gpt_trainer import GPTArgs, GPTTrainer, GPTTrainerConfig, XttsAudioConfig
8
+ from TTS.utils.manage import ModelManager
9
+ from math import ceil
10
+
11
+ LANG_TO_ISO = {
12
+ "hausa": "ha",
13
+ "luo": "luo"
14
+ }
15
+
16
+ subdirs = [d for d in os.listdir() if os.path.isdir(d) and d.startswith('xtts')]
17
+ OUT_PATH = subdirs[0]
18
+ LANG_NAME = OUT_PATH.split('_')[1]
19
+
20
+ # Logging parameters
21
+ RUN_NAME = f"GPT_XTTS_{LANG_NAME.upper()}_FT"
22
+ PROJECT_NAME = "XTTS_trainer"
23
+ DASHBOARD_LOGGER = "tensorboard"
24
+ LOGGER_URI = None
25
+
26
+ # Training Parameters
27
+ OPTIMIZER_WD_ONLY_ON_WEIGHTS = True # for multi-gpu training please make it False
28
+ START_WITH_EVAL = True # if True it will start with evaluation
29
+ BATCH_SIZE = 1 # set here the batch size
30
+ GRAD_ACUMM_STEPS = ceil(252 / BATCH_SIZE) # set here the grad accumulation steps
31
+ # Note: we recommend that BATCH_SIZE * GRAD_ACUMM_STEPS need to be at least 252 for more efficient training. You can increase/decrease BATCH_SIZE but then set GRAD_ACUMM_STEPS accordingly.
32
+
33
+ # Define here the dataset that you want to use for the fine-tuning on.
34
+ config_dataset = BaseDatasetConfig(
35
+ formatter="coqui",
36
+ dataset_name="ft_dataset",
37
+ path="data/",
38
+ meta_file_train="manifest_train.csv",
39
+ meta_file_val="manifest_dev.csv",
40
+ language=LANG_TO_ISO[LANG_NAME],
41
+ )
42
+
43
+ # Add here the configs of the datasets
44
+ DATASETS_CONFIG_LIST = [config_dataset]
45
+
46
+ # Define the path where XTTS v2.0.1 files will be downloaded
47
+ CHECKPOINTS_OUT_PATH = os.path.join(OUT_PATH, "XTTS_v2.0_original_model_files/")
48
+ os.makedirs(CHECKPOINTS_OUT_PATH, exist_ok=True)
49
+
50
+
51
+ # DVAE files
52
+ DVAE_CHECKPOINT_LINK = "https://coqui.gateway.scarf.sh/hf-coqui/XTTS-v2/main/dvae.pth"
53
+ MEL_NORM_LINK = "https://coqui.gateway.scarf.sh/hf-coqui/XTTS-v2/main/mel_stats.pth"
54
+
55
+ # Set the path to the downloaded files
56
+ DVAE_CHECKPOINT = os.path.join(CHECKPOINTS_OUT_PATH, os.path.basename(DVAE_CHECKPOINT_LINK))
57
+ MEL_NORM_FILE = os.path.join(CHECKPOINTS_OUT_PATH, os.path.basename(MEL_NORM_LINK))
58
+
59
+ # download DVAE files if needed
60
+ if not os.path.isfile(DVAE_CHECKPOINT) or not os.path.isfile(MEL_NORM_FILE):
61
+ print(" > Downloading DVAE files!")
62
+ ModelManager._download_model_files([MEL_NORM_LINK, DVAE_CHECKPOINT_LINK], CHECKPOINTS_OUT_PATH, progress_bar=True)
63
+
64
+
65
+ # Download XTTS v2.0 checkpoint if needed
66
+ TOKENIZER_FILE_LINK = "https://coqui.gateway.scarf.sh/hf-coqui/XTTS-v2/main/vocab.json"
67
+ XTTS_CHECKPOINT_LINK = "https://coqui.gateway.scarf.sh/hf-coqui/XTTS-v2/main/model.pth"
68
+ XTTS_CONFIG_LINK = "https://coqui.gateway.scarf.sh/hf-coqui/XTTS-v2/main/config.json"
69
+
70
+ # XTTS transfer learning parameters: You we need to provide the paths of XTTS model checkpoint that you want to do the fine tuning.
71
+ TOKENIZER_FILE = os.path.join(CHECKPOINTS_OUT_PATH, os.path.basename(TOKENIZER_FILE_LINK)) # vocab.json file
72
+ XTTS_CHECKPOINT = os.path.join(CHECKPOINTS_OUT_PATH, os.path.basename(XTTS_CHECKPOINT_LINK)) # model.pth file
73
+ XTTS_CONFIG_FILE = os.path.join(CHECKPOINTS_OUT_PATH, os.path.basename(XTTS_CONFIG_LINK)) # config.json file
74
+
75
+ # download XTTS v2.0 files if needed
76
+ if not os.path.isfile(TOKENIZER_FILE):
77
+ print(" > Downloading XTTS v2.0 tokenizer!")
78
+ ModelManager._download_model_files(
79
+ [TOKENIZER_FILE_LINK], CHECKPOINTS_OUT_PATH, progress_bar=True
80
+ )
81
+ if not os.path.isfile(XTTS_CHECKPOINT):
82
+ print(" > Downloading XTTS v2.0 checkpoint!")
83
+ ModelManager._download_model_files(
84
+ [XTTS_CHECKPOINT_LINK], CHECKPOINTS_OUT_PATH, progress_bar=True
85
+ )
86
+ if not os.path.isfile(XTTS_CONFIG_FILE):
87
+ print(" > Downloading XTTS v2.0 config!")
88
+ ModelManager._download_model_files(
89
+ [XTTS_CONFIG_LINK], CHECKPOINTS_OUT_PATH, progress_bar=True
90
+ )
91
+
92
+ # load training samples
93
+ train_samples, eval_samples = load_tts_samples(
94
+ DATASETS_CONFIG_LIST,
95
+ eval_split=True,
96
+ )
97
+ print(f"Train samples: {len(train_samples)}")
98
+ print(f"Eval samples: {len(eval_samples)}")
99
+
100
+ # get the longest text audio file to use as speaker reference
101
+ samples_len = [len(item["text"].split(" ")) for item in train_samples]
102
+ longest_text_idx = samples_len.index(max(samples_len))
103
+ SPEAKER_REFERENCE = [train_samples[longest_text_idx]["audio_file"]] # speaker reference to be used in training test sentences
104
+ print(f"Using speaker reference: {SPEAKER_REFERENCE}")
105
+ LANGUAGE = config_dataset.language
106
+
107
+ HAUSA_TEST_SENTENCES = [
108
+ "Umarnai don zaman tsarki.",
109
+ "wanda kuma ya faɗa mana ƙaunar da kuke yi cikin Ruhu.",
110
+ "Gama mun ji labarin bangaskiyarku a cikin Yesu Kiristi da kuma ƙaunar da kuke yi saboda dukan tsarkaka."
111
+ ]
112
+
113
+ LUO_TEST_SENTENCES = [
114
+ "jo kolosai achiel.",
115
+ "magoyo erokamano ni wuoro ka un gi mor.",
116
+ "epafra bende nonyisowa kuom hera ma roho maler osemiyou."
117
+ ]
118
+
119
+ TEST_SENTENCES = {
120
+ "hausa": [{"text": text, "speaker_wav": SPEAKER_REFERENCE, "language": LANGUAGE} for text in HAUSA_TEST_SENTENCES],
121
+ "luo": [{"text": text, "speaker_wav": SPEAKER_REFERENCE, "language": LANGUAGE} for text in LUO_TEST_SENTENCES]
122
+ }
123
+
124
+
125
+ def main():
126
+ # init args and config
127
+ model_args = GPTArgs(
128
+ max_conditioning_length=132300, # 6 secs
129
+ min_conditioning_length=11025, # 0.5 secs
130
+ debug_loading_failures=True,
131
+ max_wav_length=12*22050, # 12 secs
132
+ max_text_length=300,
133
+ mel_norm_file=MEL_NORM_FILE,
134
+ dvae_checkpoint=DVAE_CHECKPOINT,
135
+ xtts_checkpoint=XTTS_CHECKPOINT, # checkpoint path of the model that you want to fine-tune
136
+ tokenizer_file=TOKENIZER_FILE,
137
+ gpt_num_audio_tokens=1026,
138
+ gpt_start_audio_token=1024,
139
+ gpt_stop_audio_token=1025,
140
+ gpt_use_masking_gt_prompt_approach=True,
141
+ gpt_use_perceiver_resampler=True,
142
+ )
143
+ # define audio config
144
+ audio_config = XttsAudioConfig(sample_rate=22050, dvae_sample_rate=22050, output_sample_rate=24000)
145
+ # training parameters config
146
+ config = GPTTrainerConfig()
147
+
148
+ config.load_json(XTTS_CONFIG_FILE)
149
+
150
+ config.mixed_precision = True
151
+ config.precision = "bf16"
152
+ config.epochs = 1000
153
+ config.output_path = OUT_PATH
154
+ config.model_args = model_args
155
+ config.run_name = RUN_NAME
156
+ config.project_name = PROJECT_NAME
157
+ config.run_description = """
158
+ GPT XTTS training
159
+ """,
160
+ config.dashboard_logger = DASHBOARD_LOGGER
161
+ config.logger_uri = LOGGER_URI
162
+ config.audio = audio_config
163
+ config.batch_size = BATCH_SIZE
164
+ config.eval_batch_size = BATCH_SIZE
165
+ config.num_loader_workers = 8
166
+ config.print_step = 50
167
+ config.plot_step = 100
168
+ config.log_model_step = 100
169
+ config.save_step = 10000
170
+ config.save_n_checkpoints = 2
171
+ config.save_checkpoints = True
172
+ config.save_best_after = 0
173
+ config.print_eval = False
174
+ # Optimizer values like tortoise, pytorch implementation with modifications to not apply WD to non-weight parameters.
175
+ config.optimizer = "AdamW"
176
+ config.optimizer_wd_only_on_weights = OPTIMIZER_WD_ONLY_ON_WEIGHTS
177
+ config.optimizer_params = {"betas": [0.9, 0.96], "eps": 1e-8, "weight_decay": 1e-2}
178
+ config.lr = 5e-06 # learning rate
179
+ config.lr_scheduler = "MultiStepLR"
180
+ config.lr_scheduler_params = {"milestones": [5000, 150000, 300000], "gamma": 0.5, "last_epoch": -1}
181
+ config.test_sentences=TEST_SENTENCES[LANG_NAME]
182
+
183
+ # init the model from config
184
+ model = GPTTrainer.init_from_config(config)
185
+
186
+ # init the trainer and 🚀
187
+ trainer = Trainer(
188
+ TrainerArgs(
189
+ restore_path=None, # xtts checkpoint is restored via xtts_checkpoint key so no need of restore it using Trainer restore_path parameter
190
+ skip_train_epoch=False,
191
+ start_with_eval=START_WITH_EVAL,
192
+ grad_accum_steps=GRAD_ACUMM_STEPS,
193
+ ),
194
+ config,
195
+ output_path=OUT_PATH,
196
+ model=model,
197
+ train_samples=train_samples,
198
+ eval_samples=eval_samples,
199
+ )
200
+ trainer.fit()
201
+
202
+
203
+ if __name__ == "__main__":
204
+ main()
GPT_XTTS_LUO_FT-February-23-2025_07+52PM-8e59ec3/trainer_0_log.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b24d54ee53003b32d035cf4de87a46e284b25178640fcd42c6efb121126b4711
3
+ size 13144267
XTTS_v2.0_original_model_files/config.json ADDED
@@ -0,0 +1,160 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "output_path": "output",
3
+ "logger_uri": null,
4
+ "run_name": "run",
5
+ "project_name": null,
6
+ "run_description": "\ud83d\udc38Coqui trainer run.",
7
+ "print_step": 25,
8
+ "plot_step": 100,
9
+ "model_param_stats": false,
10
+ "wandb_entity": null,
11
+ "dashboard_logger": "tensorboard",
12
+ "save_on_interrupt": true,
13
+ "log_model_step": null,
14
+ "save_step": 10000,
15
+ "save_n_checkpoints": 5,
16
+ "save_checkpoints": true,
17
+ "save_all_best": false,
18
+ "save_best_after": 10000,
19
+ "target_loss": null,
20
+ "print_eval": false,
21
+ "test_delay_epochs": 0,
22
+ "run_eval": true,
23
+ "run_eval_steps": null,
24
+ "distributed_backend": "nccl",
25
+ "distributed_url": "tcp://localhost:54321",
26
+ "mixed_precision": false,
27
+ "precision": "fp16",
28
+ "epochs": 1000,
29
+ "batch_size": 32,
30
+ "eval_batch_size": 16,
31
+ "grad_clip": 0.0,
32
+ "scheduler_after_epoch": true,
33
+ "lr": 0.001,
34
+ "optimizer": "radam",
35
+ "optimizer_params": null,
36
+ "lr_scheduler": null,
37
+ "lr_scheduler_params": {},
38
+ "use_grad_scaler": false,
39
+ "allow_tf32": false,
40
+ "cudnn_enable": true,
41
+ "cudnn_deterministic": false,
42
+ "cudnn_benchmark": false,
43
+ "training_seed": 54321,
44
+ "model": "xtts",
45
+ "num_loader_workers": 0,
46
+ "num_eval_loader_workers": 0,
47
+ "use_noise_augment": false,
48
+ "audio": {
49
+ "sample_rate": 22050,
50
+ "output_sample_rate": 24000
51
+ },
52
+ "use_phonemes": false,
53
+ "phonemizer": null,
54
+ "phoneme_language": null,
55
+ "compute_input_seq_cache": false,
56
+ "text_cleaner": null,
57
+ "enable_eos_bos_chars": false,
58
+ "test_sentences_file": "",
59
+ "phoneme_cache_path": null,
60
+ "characters": null,
61
+ "add_blank": false,
62
+ "batch_group_size": 0,
63
+ "loss_masking": null,
64
+ "min_audio_len": 1,
65
+ "max_audio_len": Infinity,
66
+ "min_text_len": 1,
67
+ "max_text_len": Infinity,
68
+ "compute_f0": false,
69
+ "compute_energy": false,
70
+ "compute_linear_spec": false,
71
+ "precompute_num_workers": 0,
72
+ "start_by_longest": false,
73
+ "shuffle": false,
74
+ "drop_last": false,
75
+ "datasets": [
76
+ {
77
+ "formatter": "",
78
+ "dataset_name": "",
79
+ "path": "",
80
+ "meta_file_train": "",
81
+ "ignored_speakers": null,
82
+ "language": "",
83
+ "phonemizer": "",
84
+ "meta_file_val": "",
85
+ "meta_file_attn_mask": ""
86
+ }
87
+ ],
88
+ "test_sentences": [],
89
+ "eval_split_max_size": null,
90
+ "eval_split_size": 0.01,
91
+ "use_speaker_weighted_sampler": false,
92
+ "speaker_weighted_sampler_alpha": 1.0,
93
+ "use_language_weighted_sampler": false,
94
+ "language_weighted_sampler_alpha": 1.0,
95
+ "use_length_weighted_sampler": false,
96
+ "length_weighted_sampler_alpha": 1.0,
97
+ "model_args": {
98
+ "gpt_batch_size": 1,
99
+ "enable_redaction": false,
100
+ "kv_cache": true,
101
+ "gpt_checkpoint": null,
102
+ "clvp_checkpoint": null,
103
+ "decoder_checkpoint": null,
104
+ "num_chars": 255,
105
+ "tokenizer_file": "",
106
+ "gpt_max_audio_tokens": 605,
107
+ "gpt_max_text_tokens": 402,
108
+ "gpt_max_prompt_tokens": 70,
109
+ "gpt_layers": 30,
110
+ "gpt_n_model_channels": 1024,
111
+ "gpt_n_heads": 16,
112
+ "gpt_number_text_tokens": 6681,
113
+ "gpt_start_text_token": null,
114
+ "gpt_stop_text_token": null,
115
+ "gpt_num_audio_tokens": 1026,
116
+ "gpt_start_audio_token": 1024,
117
+ "gpt_stop_audio_token": 1025,
118
+ "gpt_code_stride_len": 1024,
119
+ "gpt_use_masking_gt_prompt_approach": true,
120
+ "gpt_use_perceiver_resampler": true,
121
+ "input_sample_rate": 22050,
122
+ "output_sample_rate": 24000,
123
+ "output_hop_length": 256,
124
+ "decoder_input_dim": 1024,
125
+ "d_vector_dim": 512,
126
+ "cond_d_vector_in_each_upsampling_layer": true,
127
+ "duration_const": 102400
128
+ },
129
+ "model_dir": null,
130
+ "languages": [
131
+ "en",
132
+ "es",
133
+ "fr",
134
+ "de",
135
+ "it",
136
+ "pt",
137
+ "pl",
138
+ "tr",
139
+ "ru",
140
+ "nl",
141
+ "cs",
142
+ "ar",
143
+ "zh-cn",
144
+ "hu",
145
+ "ko",
146
+ "ja",
147
+ "hi",
148
+ "luo"
149
+ ],
150
+ "temperature": 0.75,
151
+ "length_penalty": 1.0,
152
+ "repetition_penalty": 5.0,
153
+ "top_k": 50,
154
+ "top_p": 0.85,
155
+ "num_gpt_outputs": 1,
156
+ "gpt_cond_len": 30,
157
+ "gpt_cond_chunk_len": 4,
158
+ "max_ref_len": 30,
159
+ "sound_norm_refs": false
160
+ }
XTTS_v2.0_original_model_files/dvae.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1e91bf10e24ba7172ac30c59790df2ee2df027394f4f05c6f3cb4649e01acba
3
+ size 210893114
XTTS_v2.0_original_model_files/mel_stats.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f69422a8a8f344c4fca2f0c6b8d41d2151d6615b7321e48e6bb15ae949b119c
3
+ size 1067
XTTS_v2.0_original_model_files/model.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7ea20001c6a0a841c77e252d8409f6a74fb423e79b3206a0771ba5989776187
3
+ size 1867929118
XTTS_v2.0_original_model_files/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
dvae_logs.txt ADDED
The diff for this file is too large to render. See raw diff