Merge branch 'v.2' of github.com:w-okada/voice-changer into v.2

This commit is contained in:
wok 2024-08-07 19:48:43 +09:00
commit 9ce408ccf5

View File

@ -52,38 +52,12 @@
},
{
"cell_type": "code",
"execution_count": 1,
"execution_count": null,
"metadata": {
"id": "W2GYWTHWmRIY",
"outputId": "e4ce3296-81f9-4004-daf0-219a2977077b",
"colab": {
"base_uri": "https://localhost:8080/"
},
"cellView": "form"
},
"outputs": [
{
"output_type": "stream",
"name": "stdout",
"text": [
"GPU is available\n",
"GPU Name: Tesla T4\n",
"Welcome to ColabMod\n",
"Mounted at /content/drive\n",
"Checking latest version...\n",
"current_version_hash: eae1b7592ec7c44b88730d2044b7bd9a\n",
"latest_version_hash : de736169a1c6213e63479eda109c1baf\n",
"hash not match -> download latest version\n",
" % Total % Received % Xferd Average Speed Time Time Time Current\n",
" Dload Upload Total Spent Left Speed\n",
"100 1161 100 1161 0 0 6715 0 --:--:-- --:--:-- --:--:-- 6750\n",
"100 3028M 100 3028M 0 0 23.8M 0 0:02:07 0:02:07 --:--:-- 53.9M\n",
"Download is done.\n",
"/content/drive/MyDrive/vcclient\n",
"Installing modules... Install is done.\n"
]
}
],
"outputs": [],
"source": [
"#=================Updated=================\n",
"# @title **[1]** Clone repository and install dependencies\n",
@ -110,7 +84,7 @@
"\n",
"# Configs\n",
"Run_Cell=0\n",
"Use_Drive=True #@param {type:\"boolean\"}\n",
"Use_Drive=False #@param {type:\"boolean\"}\n",
"\n",
"current_version_hash=None\n",
"latest_version_hash=None\n",
@ -166,6 +140,8 @@
"\n",
" if mode == \"elf\":\n",
" !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab -o {work_dir}/vcclient_latest_for_colab\n",
" # !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_2.0.44-alpha_for_colab -o {work_dir}/vcclient_latest_for_colab\n",
"\n",
" elif mode == \"zip\":\n",
" !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab.zip -o {work_dir}/vcclient_latest_for_colab.zip\n",
"\n",
@ -182,6 +158,7 @@
"\n",
" if mode == \"elf\":\n",
" !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab -o {work_dir}/vcclient_latest_for_colab\n",
" # !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_2.0.44-alpha_for_colab -o {work_dir}/vcclient_latest_for_colab\n",
" elif mode == \"zip\":\n",
" !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab.zip -o {work_dir}/vcclient_latest_for_colab.zip\n",
"\n",
@ -209,44 +186,12 @@
},
{
"cell_type": "code",
"execution_count": 2,
"execution_count": null,
"metadata": {
"id": "s7mYqKtW6VOI",
"outputId": "e86ad79b-cb62-4ea5-a8d9-dc7b9ca9e6ed",
"colab": {
"base_uri": "https://localhost:8080/",
"height": 148
}
"cellView": "form"
},
"outputs": [
{
"output_type": "stream",
"name": "stdout",
"text": [
"--------- SERVER READY! ---------\n",
"Your server is available. elapsed: 173sec\n",
"https://2bj201er26o-496ff2e9c6d22116-8003-colab.googleusercontent.com/\n",
"---------------------------------\n"
]
},
{
"output_type": "display_data",
"data": {
"text/plain": [
"<IPython.lib.display.Audio object>"
],
"text/html": [
"\n",
" <audio controls=\"controls\" autoplay=\"autoplay\">\n",
" <source src=\"https://huggingface.co/wok000/voices/resolve/main/vcclient001_vctk229_gpt-sovits_vcclient-ready.wav\" type=\"audio/x-wav\" />\n",
" Your browser does not support the audio element.\n",
" </audio>\n",
" "
]
},
"metadata": {}
}
],
"outputs": [],
"source": [
"# @title **[2]** Start server\n",
"# @markdown This cell will start the server, the first time that you run it will download the models, so it can take a while (2~4 minutes)\n",
@ -334,7 +279,7 @@
"colab": {
"provenance": [],
"gpuType": "T4",
"authorship_tag": "ABX9TyPEPBkGLPAGJGofiLPcnQhG",
"authorship_tag": "ABX9TyPpbFBssircyl/qT8uDX2zy",
"include_colab_link": true
},
"kernelspec": {