-
Notifications
You must be signed in to change notification settings - Fork 21
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Use with different languages #4
Comments
Me too, I'm interested in Italian |
Models for French, German, and Dutch have just been added. It will take more time to additional languages, but fortunately the data is available: http://openslr.org/94/ |
Spanish would be a great addition :D |
Hi @synesthesiam do you have any news about other languages? |
Not yet, but Spanish, Portuguese, Polish, and Italian should be possible with the MLS dataset. |
Great, but i don't know how to train the model :( maybe do you have any instructions or something? |
@synesthesiam do you have any guidelines how to export LibriTTS-R generator from a checkpoint? |
Hi @synesthesiam I've tried several times over the past few days to create a new model (.pt) to use with this notebook: https://colab.research.google.com/drive/1q1oe2zOyZp7UsB3jJiQ1IFn8z5YfjwEb#scrollTo=1cbqBebHXjFD to create my custom Italian wake word. I've attempted to follow various guides, including this one: https://github.com/rhasspy/piper/blob/master/TRAINING.md. I also tried starting with the 15 GB dataset you mentioned here: https://openslr.org/94/, but I haven't been successful. Could you please tell me how to do it? It would be a great help. |
I was able to export a .pt file, but not get it working with piper-sample-generator. (I'm using English US, but cannot use a model trained with piper in Piper Sample Generator) Here is what I've done so far:
Then ran
(To be clear, this doesn't train the model, just export it from a checkpoint (.ckpt file).)
I installed PyTorch 2.0.0 and it's dependencies, piper-phonemize, and webrtcvad. By running import os
import sys
if "piper-sample-generator/" not in sys.path:
sys.path.append("piper-sample-generator/")
from generate_samples import generate_samples
target_word = 'edward'
def text_to_speech(text):
generate_samples(text = text, max_samples=1, length_scales=[1.1], noise_scales=[0.7], noise_scale_ws = [0.7], output_dir = './', batch_size=1, auto_reduce_batch_size=True, file_names=["test_generation.wav"], model='James5.pt')
text_to_speech(target_word) This was the output:
So, I'm not really sure what to do from here, I don't really understand how AI in Python works, but I've gotten this far, help would be greatly appreciated. Edit: If it's any help, here's what is outputted when I run Second edit: I just realised that I'm likely getting this error because I'm using a single speaker model. Would this be likely @synesthesiam? |
I was also trying to use the same notebook with referencing the German model (that's included in the release) instead of the English one, using the config from this repo, but can't get any proper human text with the sample generator, it sounds just some random phonemes... |
After some trial, my outputs are following:
parser.add_argument(
"--ckpt-to-pt", help="Convert .ckpt file to .pt file and exit", metavar="CHECKPOINT_PATH"
)
...
if args.ckpt_to_pt:
convert_ckpt_to_pt(args.ckpt_to_pt)
return
...
def convert_ckpt_to_pt(ckpt_path):
_LOGGER.debug(f"Converting .ckpt file: {ckpt_path} to .pt file")
model = VitsModel.load_from_checkpoint(ckpt_path)
pt_path = ckpt_path.replace(".ckpt", ".pt")
torch.save(model.model_g, pt_path)
_LOGGER.debug(f"Model saved as .pt file: {pt_path}")
python3 generate_samples.py '<your-word>' --model '<created-pt-model-path>' --max-samples 10 --output-dir <output-dir> |
Hi there,
I want to use the steps in https://www.home-assistant.io/voice_control/create_wake_word/ to create my custom wake word.
Is there a chance to use/add different languages?
I'm interested in Polish, but I think that support for any additional language would be awesome.
The text was updated successfully, but these errors were encountered: