Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Is pretrained model of Tacotron2 + LPCNet is avalilible? #1

Open
mrgloom opened this issue May 16, 2019 · 24 comments
Open

Is pretrained model of Tacotron2 + LPCNet is avalilible? #1

mrgloom opened this issue May 16, 2019 · 24 comments

Comments

@mrgloom
Copy link

mrgloom commented May 16, 2019

Is pretrained model of Tacotron2 + LPCNet is avalilible?

@MlWoo
Copy link
Owner

MlWoo commented May 16, 2019

We have the pretrained model of the two on mandarin dataset. But I think it is not illegal without the permission of my company to release them personally. It is not hard to train the model with your materials following the steps in readme. I will follow the steps to rerun the training and synthesising procedures to make sure it is right.

@sheepHavingPurpleLeaf
Copy link

Hi @MlWoo, i have trained Tacotron2 about 21000 steps on a female mandarin dataset and connected to LPCNet. At 21000, the error is about 0.33 and decoder is already aligned with encoder. The output wav has really bad quality, ie. large portion of the sentence is silence and cannot tell even gender from voicing part. How many steps have you trained tacotron to achieve good sound?

@MlWoo
Copy link
Owner

MlWoo commented May 25, 2019

I have pointed that the quality of vocoder is sensitive to the estimation of pitch parameters. maybe you could achieve it with 210000 steps. we have different params to train the tacotron2. and I think it has no much meaningful.info.for you. our loss is less than 0.1.

@sheepHavingPurpleLeaf
Copy link

@MlWoo I waited for another 10k steps, the loss stays above 0.3. Any advice on taco2 params? Thanks in advance.

@superhg2012
Copy link

Hi, @MlWoo Did you train T2 with 16k mandarin dataset?

@MlWoo
Copy link
Owner

MlWoo commented May 31, 2019

@superhg2012 yes.

@superhg2012
Copy link

@sheepHavingPurpleLeaf which Tacotron repo did you use? any better results?

@superhg2012
Copy link

@MlWoo audio processing parameters is not used when training T2 with .f32 feature files. I tried different hparams, but can only achieve 0.2 loss, did you adjust the T2 network params? thanks in advance!

@MlWoo
Copy link
Owner

MlWoo commented Jun 10, 2019

@superhg2012 The loss is gained with the teacher forcing mode.

@superhg2012
Copy link

@MlWoo thanks!! constant and scheduled, which mode is preferenced?

@MlWoo
Copy link
Owner

MlWoo commented Jun 10, 2019

@superhg2012 constant mode.

@superhg2012
Copy link

many thanks !!

@superhg2012
Copy link

superhg2012 commented Jun 13, 2019

I trained T2 for 130k steps and the lowest loss value is 0.13, and the synthsized audio is still not good as expected. some post processing needed?

demo.zip

@MlWoo I think LPCNet is ok, the cause is pitch parameters predicted from Tacotron2, could you give some suggestions?

@sheepHavingPurpleLeaf
Copy link

sheepHavingPurpleLeaf commented Jun 13, 2019

@superhg2012 Can you share your hparams? you are using pinyin to train or phoneme?

@MlWoo
Copy link
Owner

MlWoo commented Jun 13, 2019

@estherxue Could you post your samples?

@superhg2012
Copy link

@sheepHavingPurpleLeaf I used pinyin to train Tacotron2 and parameters is common.

@sheepHavingPurpleLeaf
Copy link

@superhg2012 I have got similar result with yours. Did you train LPCNet with the English dataset provided in Mozilla's repo or you used your mandarin dataset?

@superhg2012
Copy link

@sheepHavingPurpleLeaf I used same mandarin dataset for LPCNet and Tacotron2. The sound quality is almost same while loss is around 0.13 ~ 0.17.

@estherxue
Copy link

Hi, here are my samples trained with Tacotron 2 + LPCNet.
tacotron2+lpcnet.zip

@superhg2012
Copy link

superhg2012 commented Jun 13, 2019

@estherxue hi, the examples sounds good, I have several questions.
1 . you are using pinyin to train or phoneme?
2. you are using same dataset to train both t2 and lpcnet?
3. how many steps takes to train t2 part? and last loss?
4. you are training in GTA mode?

thanks in advance!!

@MlWoo
Copy link
Owner

MlWoo commented Jun 13, 2019

@superhg2012 Our team (Xue is my collegue) does not use any other trick to train the tacotron2.

  1. pinyin
  2. the same dataset
  3. 280k if I remember correctly. and loss is about 0.1. Maybe the lr scheduling is not same as the t2 repo because T2 repo is updated recently.
  4. No GTA. if you want to use gta mode, there is a lot tricky work (like round audio to the frames)to be done.

@superhg2012
Copy link

@MlWoo get it, thanks for kind reply!!

@ajaysg-zz
Copy link

how long does it take to synthesize on GPU as well as CPU

@estherxue
Copy link

estherxue commented Oct 12, 2019 via email

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

6 participants