Hello, Yi-Chiao WU!
I appreciate that if you can read the issue and give me some feedback.
I respectively used the WORLD and the model of QPPWGaf_20(checkpoint400000) from you to be a vocoder to synthesis speech based on my own speech file(from the corpus named LJSpeech-1.1).
The process is Speech→Extract feature→synthesis→Speech followed by readme.
But the output of QPPWGaf_20 is neither better nor worse than the output of WORLD.
Just because I didn't take the VCC-corpus to be the input?
Or there are other reasons?