diff --git a/notebooks/piper_model_exporter.ipynb b/notebooks/piper_model_exporter.ipynb index 276f7fe..3b98bec 100644 --- a/notebooks/piper_model_exporter.ipynb +++ b/notebooks/piper_model_exporter.ipynb @@ -5,7 +5,7 @@ "colab": { "provenance": [], "gpuType": "T4", - "authorship_tag": "ABX9TyPBXwWHq64s9HTowl1s6H6d", + "authorship_tag": "ABX9TyNHZIC4OUHDX+ElHmpW/AAV", "include_colab_link": true }, "kernelspec": { @@ -99,7 +99,7 @@ " config = json.load(file)\n", " sample_rate = config[\"audio\"][\"sample_rate\"]\n", " num_speakers = config[\"num_speakers\"]\n", - " output.eval_js('new Audio(f\"{guideurl}/waiting.wav?raw=true\").play()')\n", + " output.eval_js(f'new Audio(\"{guideurl}/waiting.wav?raw=true\").play()')\n", " text_area = widgets.Textarea(\n", " description = \"fill in this following template and press start to generate the voice package\",\n", " value=f'# Model card for {voice_name} ({quality})\\n\\n* Language: {language} (normaliced)\\n* Speakers: {num_speakers}\\n* Quality: {quality}\\n* Samplerate: {sample_rate}Hz\\n\\n## Dataset\\n\\n* URL: \\n* License: \\n\\n## Training\\n\\nTrained from scratch.\\nOr finetuned from: ',\n", @@ -114,7 +114,7 @@ " text_area.close()\n", " button.close()\n", " output.clear()\n", - " output.eval_js('new Audio(f\"{guideurl}/starting.wav?raw=true\").play()')\n", + " output.eval_js(f'new Audio(\"{guideurl}/starting.wav?raw=true\").play()')\n", " start_process()\n", "\n", " button.on_click(create_model_card)\n", @@ -125,7 +125,7 @@ " !python -m piper_train.export_onnx \"/content/project/model.ckpt\" \"{export_voice_path}/{export_voice_name}.onnx\"\n", " print(\"compressing...\")\n", " !tar -czvf \"{packages_path}/voice-{export_voice_name}.tar.gz\" -C \"{export_voice_path}\" .\n", - " output.eval_js('new Audio(\"{guideurl}/success.wav?raw=true\").play()')" + " output.eval_js(f'new Audio(\"{guideurl}/success.wav?raw=true\").play()')" ], "metadata": { "cellView": "form",