-
Notifications
You must be signed in to change notification settings - Fork 28
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG🐛] Size match in converted xttsv2 models #43
Comments
Ah ha, figured it out. Coqui xtts2 v2.0.2 differs from v2.0.3 in the # of tokens https://huggingface.co/coqui/XTTS-v2/commit/6b8036b35d787cf43d18d640587956b9db8fd1b8 the above models were training on v2.0.2 The convertor script needs to be aware of this, since any difference will cause it to not work once converted, since the config/etc don't match the actual trained gpt section of the model Correct me if I'm wrong, but basically, either this means the gpt config must be adjusted in this case, since it no longer matches the stock config/etc. OR you should just fail the convertor, and complain that only v2.0.3 models can be converted. |
same issue here with 2.0.0 model version used for training, this would also maybe explain the difference in quality/output #27 when I am converting coqui 2.0.0 model using provided script... |
Error[rank0]: raise RuntimeError( Explanationsame issue here i trained for a new languague when i run checkpoint_converter.py script it download json file for xtts config and gpt tokenizer config then i update json file with new languague vocabulary. But i also get following error. |
Cool I didn't knew about this, I'll be looking into it |
@mlinmg I have tried the new conversion script, but after conversion I needed manually replace the tokenizer from version 2.0.3 to 2.0.0 which is the model version I used for finetuning. And adjust the setting accordingly to the 2.0.0. repo. The quality is much better but it is still worse than in original... I couldn't figure it why maybe you are calculating latents from reference differently or something. Or I am missing some settings that are different between implementations... @elvinzade you also need to change the tokenizer size in all of the config.json and some python files where it is referenced, just search for the number in all files and change to the version corresponding to the coqui model repo version on huggingface |
I don't believe it's quite a drop-in 'downgrade' to go backwards. |
@C00reNUT thanks for your proposal. I appreciate your proposal regarding the Coqui model. However, I've trained a new language that is not part of the languages introduced in the official Coqui model repository. During this process, I extended the tokenizer vocabulary size to accommodate the new language. The model should work seamlessly with the extended tokenizer and support the newly added language. If I attempt to adjust the tokenizer size, I lose the ability to use the newly added language feature. |
Bug Description
The text was updated successfully, but these errors were encountered: