From 52411aa1d8338a0431bab4c434b8ff886021db44 Mon Sep 17 00:00:00 2001 From: w-okada <48346627+w-okada@users.noreply.github.com> Date: Wed, 7 Aug 2024 00:18:08 +0900 Subject: [PATCH 1/2] =?UTF-8?q?Colab=20=E3=82=92=E4=BD=BF=E7=94=A8?= =?UTF-8?q?=E3=81=97=E3=81=A6=E4=BD=9C=E6=88=90=E3=81=95=E3=82=8C=E3=81=BE?= =?UTF-8?q?=E3=81=97=E3=81=9F?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- w_okada's_Voice_Changer_version_2_x.ipynb | 37 +++++++++++------------ 1 file changed, 18 insertions(+), 19 deletions(-) diff --git a/w_okada's_Voice_Changer_version_2_x.ipynb b/w_okada's_Voice_Changer_version_2_x.ipynb index b9f6abed..98db70cc 100644 --- a/w_okada's_Voice_Changer_version_2_x.ipynb +++ b/w_okada's_Voice_Changer_version_2_x.ipynb @@ -55,11 +55,10 @@ "execution_count": 1, "metadata": { "id": "W2GYWTHWmRIY", - "outputId": "e4ce3296-81f9-4004-daf0-219a2977077b", + "outputId": "88177d0e-d7a2-4a0b-fb09-578a5c27c430", "colab": { "base_uri": "https://localhost:8080/" - }, - "cellView": "form" + } }, "outputs": [ { @@ -69,17 +68,13 @@ "GPU is available\n", "GPU Name: Tesla T4\n", "Welcome to ColabMod\n", - "Mounted at /content/drive\n", - "Checking latest version...\n", - "current_version_hash: eae1b7592ec7c44b88730d2044b7bd9a\n", - "latest_version_hash : de736169a1c6213e63479eda109c1baf\n", - "hash not match -> download latest version\n", + "Downloading the latest vcclient... \n", " % Total % Received % Xferd Average Speed Time Time Time Current\n", " Dload Upload Total Spent Left Speed\n", - "100 1161 100 1161 0 0 6715 0 --:--:-- --:--:-- --:--:-- 6750\n", - "100 3028M 100 3028M 0 0 23.8M 0 0:02:07 0:02:07 --:--:-- 53.9M\n", + "100 1149 100 1149 0 0 3004 0 --:--:-- --:--:-- --:--:-- 3007\n", + "100 2967M 100 2967M 0 0 154M 0 0:00:19 0:00:19 --:--:-- 219M\n", "Download is done.\n", - "/content/drive/MyDrive/vcclient\n", + "/content\n", "Installing modules... Install is done.\n" ] } @@ -110,7 +105,7 @@ "\n", "# Configs\n", "Run_Cell=0\n", - "Use_Drive=True #@param {type:\"boolean\"}\n", + "Use_Drive=False #@param {type:\"boolean\"}\n", "\n", "current_version_hash=None\n", "latest_version_hash=None\n", @@ -165,7 +160,9 @@ " latest_hash_path=f'{work_dir}/latest_hash.txt'\n", "\n", " if mode == \"elf\":\n", - " !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab -o {work_dir}/vcclient_latest_for_colab\n", + " # !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab -o {work_dir}/vcclient_latest_for_colab\n", + " !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_2.0.44-alpha_for_colab -o {work_dir}/vcclient_latest_for_colab\n", + "\n", " elif mode == \"zip\":\n", " !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab.zip -o {work_dir}/vcclient_latest_for_colab.zip\n", "\n", @@ -181,7 +178,8 @@ " latest_version_hash = open('latest_hash.txt').read().strip()\n", "\n", " if mode == \"elf\":\n", - " !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab -o {work_dir}/vcclient_latest_for_colab\n", + " # !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab -o {work_dir}/vcclient_latest_for_colab\n", + " !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_2.0.44-alpha_for_colab -o {work_dir}/vcclient_latest_for_colab\n", " elif mode == \"zip\":\n", " !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab.zip -o {work_dir}/vcclient_latest_for_colab.zip\n", "\n", @@ -212,11 +210,12 @@ "execution_count": 2, "metadata": { "id": "s7mYqKtW6VOI", - "outputId": "e86ad79b-cb62-4ea5-a8d9-dc7b9ca9e6ed", + "outputId": "3c88d3c3-5888-479f-9256-c6719d15b1f2", "colab": { "base_uri": "https://localhost:8080/", "height": 148 - } + }, + "cellView": "form" }, "outputs": [ { @@ -224,8 +223,8 @@ "name": "stdout", "text": [ "--------- SERVER READY! ---------\n", - "Your server is available. elapsed: 173sec\n", - "https://2bj201er26o-496ff2e9c6d22116-8003-colab.googleusercontent.com/\n", + "Your server is available. elapsed: 98sec\n", + "https://wdc8fs91tuo-496ff2e9c6d22116-8003-colab.googleusercontent.com/\n", "---------------------------------\n" ] }, @@ -334,7 +333,7 @@ "colab": { "provenance": [], "gpuType": "T4", - "authorship_tag": "ABX9TyPEPBkGLPAGJGofiLPcnQhG", + "authorship_tag": "ABX9TyPL8HlIEcd7fZuat/q9Ir1k", "include_colab_link": true }, "kernelspec": { From bb6cdc8e72a3948b685be845a0b0ea9640492e32 Mon Sep 17 00:00:00 2001 From: w-okada <48346627+w-okada@users.noreply.github.com> Date: Wed, 7 Aug 2024 02:18:44 +0900 Subject: [PATCH 2/2] =?UTF-8?q?Colab=20=E3=82=92=E4=BD=BF=E7=94=A8?= =?UTF-8?q?=E3=81=97=E3=81=A6=E4=BD=9C=E6=88=90=E3=81=95=E3=82=8C=E3=81=BE?= =?UTF-8?q?=E3=81=97=E3=81=9F?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- w_okada's_Voice_Changer_version_2_x.ipynb | 74 +++-------------------- 1 file changed, 10 insertions(+), 64 deletions(-) diff --git a/w_okada's_Voice_Changer_version_2_x.ipynb b/w_okada's_Voice_Changer_version_2_x.ipynb index 98db70cc..e6bfe8d3 100644 --- a/w_okada's_Voice_Changer_version_2_x.ipynb +++ b/w_okada's_Voice_Changer_version_2_x.ipynb @@ -52,33 +52,12 @@ }, { "cell_type": "code", - "execution_count": 1, + "execution_count": null, "metadata": { "id": "W2GYWTHWmRIY", - "outputId": "88177d0e-d7a2-4a0b-fb09-578a5c27c430", - "colab": { - "base_uri": "https://localhost:8080/" - } + "cellView": "form" }, - "outputs": [ - { - "output_type": "stream", - "name": "stdout", - "text": [ - "GPU is available\n", - "GPU Name: Tesla T4\n", - "Welcome to ColabMod\n", - "Downloading the latest vcclient... \n", - " % Total % Received % Xferd Average Speed Time Time Time Current\n", - " Dload Upload Total Spent Left Speed\n", - "100 1149 100 1149 0 0 3004 0 --:--:-- --:--:-- --:--:-- 3007\n", - "100 2967M 100 2967M 0 0 154M 0 0:00:19 0:00:19 --:--:-- 219M\n", - "Download is done.\n", - "/content\n", - "Installing modules... Install is done.\n" - ] - } - ], + "outputs": [], "source": [ "#=================Updated=================\n", "# @title **[1]** Clone repository and install dependencies\n", @@ -160,8 +139,8 @@ " latest_hash_path=f'{work_dir}/latest_hash.txt'\n", "\n", " if mode == \"elf\":\n", - " # !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab -o {work_dir}/vcclient_latest_for_colab\n", - " !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_2.0.44-alpha_for_colab -o {work_dir}/vcclient_latest_for_colab\n", + " !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab -o {work_dir}/vcclient_latest_for_colab\n", + " # !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_2.0.44-alpha_for_colab -o {work_dir}/vcclient_latest_for_colab\n", "\n", " elif mode == \"zip\":\n", " !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab.zip -o {work_dir}/vcclient_latest_for_colab.zip\n", @@ -178,8 +157,8 @@ " latest_version_hash = open('latest_hash.txt').read().strip()\n", "\n", " if mode == \"elf\":\n", - " # !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab -o {work_dir}/vcclient_latest_for_colab\n", - " !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_2.0.44-alpha_for_colab -o {work_dir}/vcclient_latest_for_colab\n", + " !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab -o {work_dir}/vcclient_latest_for_colab\n", + " # !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_2.0.44-alpha_for_colab -o {work_dir}/vcclient_latest_for_colab\n", " elif mode == \"zip\":\n", " !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab.zip -o {work_dir}/vcclient_latest_for_colab.zip\n", "\n", @@ -207,45 +186,12 @@ }, { "cell_type": "code", - "execution_count": 2, + "execution_count": null, "metadata": { "id": "s7mYqKtW6VOI", - "outputId": "3c88d3c3-5888-479f-9256-c6719d15b1f2", - "colab": { - "base_uri": "https://localhost:8080/", - "height": 148 - }, "cellView": "form" }, - "outputs": [ - { - "output_type": "stream", - "name": "stdout", - "text": [ - "--------- SERVER READY! ---------\n", - "Your server is available. elapsed: 98sec\n", - "https://wdc8fs91tuo-496ff2e9c6d22116-8003-colab.googleusercontent.com/\n", - "---------------------------------\n" - ] - }, - { - "output_type": "display_data", - "data": { - "text/plain": [ - "" - ], - "text/html": [ - "\n", - " \n", - " " - ] - }, - "metadata": {} - } - ], + "outputs": [], "source": [ "# @title **[2]** Start server\n", "# @markdown This cell will start the server, the first time that you run it will download the models, so it can take a while (2~4 minutes)\n", @@ -333,7 +279,7 @@ "colab": { "provenance": [], "gpuType": "T4", - "authorship_tag": "ABX9TyPL8HlIEcd7fZuat/q9Ir1k", + "authorship_tag": "ABX9TyPpbFBssircyl/qT8uDX2zy", "include_colab_link": true }, "kernelspec": {