2024-06-22 10:00:37 +03:00
{
"cells": [
{
"cell_type": "markdown",
"metadata": {
"id": "view-in-github",
"colab_type": "text"
},
"source": [
"<a href=\"https://colab.research.google.com/github/w-okada/voice-changer/blob/v.2/w_okada's_Voice_Changer_version_2_x.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>"
]
},
{
"cell_type": "markdown",
"metadata": {
"id": "wNCGmSXbfZRr"
},
"source": [
"### w-okada's Voice Changer version 2.x | **Google Colab**\n",
"\n",
"## READ ME - VERY IMPORTANT\n",
"This is an attempt to run [Realtime Voice Changer](https://github.com/w-okada/voice-changer) on Google Colab, still not perfect but is totally usable, you can use the following settings for better results:\n",
"\n",
2024-06-30 09:42:31 +03:00
"If you're using a index: `f0: RMVPE_ONNX | Chunk: 24000 or higher | Extra: 7680`\\\n",
"If you're not using a index: `f0: RMVPE_ONNX | Chunk: 24000 or higher | Extra: 7680`\\\n",
2024-06-22 10:00:37 +03:00
"**Don't forget to select your Colab GPU in the GPU field (<b>Tesla T4</b>, for free users)*\n",
"> Seems that PTH models performance better than ONNX for now, you can still try ONNX models and see if it satisfies you\n",
"\n",
"\n",
"*You can always [click here](https://rentry.co/VoiceChangerGuide#gpu-chart-for-known-working-chunkextra\n",
") to check if these settings are up-to-date*\n",
"<br><br>\n",
"\n",
"---\n",
"\n",
"###Always use Colab GPU (**VERY VERY VERY IMPORTANT!**)\n",
"You need to use a Colab GPU so the Voice Changer can work faster and better\\\n",
"Use the menu above and click on **Runtime** » **Change runtime** » **Hardware acceleration** to select a GPU (**T4 is the free one**)\n",
"\n",
"---\n",
"\n",
"\n",
"# **Credits and Support**\n",
"Realtime Voice Changer by [w-okada](https://github.com/w-okada)\\\n",
"Colab files updated by [rafacasari](https://github.com/Rafacasari)\\\n",
"Recommended settings by [Raven](https://github.com/ravencutie21)\\\n",
"Modified again by [Hina](https://github.com/HinaBl)\\\n",
"Enable FCPE by [TheTrustedComputer](https://github.com/TheTrustedComputer)\n",
"\n",
"Need help? [AI Hub Discord](https://discord.gg/aihub) » ***#help-realtime-vc***\n"
]
},
{
"cell_type": "code",
2024-06-30 09:42:31 +03:00
"execution_count": 1,
2024-06-22 10:00:37 +03:00
"metadata": {
"id": "W2GYWTHWmRIY",
2024-06-30 09:42:31 +03:00
"outputId": "e4ce3296-81f9-4004-daf0-219a2977077b",
2024-06-23 21:36:15 +03:00
"colab": {
"base_uri": "https://localhost:8080/"
2024-06-30 09:42:31 +03:00
},
"cellView": "form"
2024-06-22 10:00:37 +03:00
},
2024-06-23 21:36:15 +03:00
"outputs": [
{
"output_type": "stream",
"name": "stdout",
"text": [
"GPU is available\n",
"GPU Name: Tesla T4\n",
"Welcome to ColabMod\n",
"Mounted at /content/drive\n",
"Checking latest version...\n",
2024-06-30 09:42:31 +03:00
"current_version_hash: eae1b7592ec7c44b88730d2044b7bd9a\n",
"latest_version_hash : de736169a1c6213e63479eda109c1baf\n",
2024-06-23 21:36:15 +03:00
"hash not match -> download latest version\n",
" % Total % Received % Xferd Average Speed Time Time Time Current\n",
" Dload Upload Total Spent Left Speed\n",
2024-06-30 09:42:31 +03:00
"100 1161 100 1161 0 0 6715 0 --:--:-- --:--:-- --:--:-- 6750\n",
"100 3028M 100 3028M 0 0 23.8M 0 0:02:07 0:02:07 --:--:-- 53.9M\n",
2024-06-23 21:36:15 +03:00
"Download is done.\n",
"/content/drive/MyDrive/vcclient\n",
"Installing modules... Install is done.\n"
]
}
],
2024-06-22 10:00:37 +03:00
"source": [
"#=================Updated=================\n",
"# @title **[1]** Clone repository and install dependencies\n",
2024-06-30 09:42:31 +03:00
"# @markdown This first step will download the latest version of Voice Changer and install the dependencies. **It can take some time to complete.(~5min)**\n",
2024-06-22 10:00:37 +03:00
"\n",
"#@markdown ---\n",
2024-06-23 13:59:09 +03:00
"# @markdown By using Google Drive, you can avoid re-downloading already downloaded versions.\n",
2024-06-22 10:00:37 +03:00
"\n",
"\n",
"import os\n",
"import time\n",
"import subprocess\n",
"import threading\n",
"import shutil\n",
"import base64\n",
"import codecs\n",
2024-06-23 21:36:15 +03:00
"import torch\n",
"import sys\n",
"\n",
"from typing import Literal, TypeAlias\n",
"\n",
"Mode: TypeAlias = Literal[\"elf\", \"zip\"]\n",
"mode:Mode=\"elf\"\n",
2024-06-22 10:00:37 +03:00
"\n",
"# Configs\n",
"Run_Cell=0\n",
"Use_Drive=True #@param {type:\"boolean\"}\n",
"\n",
2024-06-23 13:59:09 +03:00
"current_version_hash=None\n",
"latest_version_hash=None\n",
"\n",
2024-06-23 21:36:15 +03:00
"# Check GPU\n",
"if torch.cuda.is_available():\n",
" print(\"GPU is available\")\n",
" print(\"GPU Name:\", torch.cuda.get_device_name(0))\n",
"else:\n",
" print(\"GPU is not available\")\n",
" # sys.exit(\"No GPU available. Change runtime.\")\n",
"\n",
2024-06-23 13:59:09 +03:00
"\n",
2024-06-22 10:00:37 +03:00
"notebook_env=0\n",
"if os.path.exists('/content'):\n",
" notebook_env=1\n",
" print(\"Welcome to ColabMod\")\n",
" from google.colab import drive\n",
"\n",
"elif os.path.exists('/kaggle/working'):\n",
" notebook_env=2\n",
" print(\"Welcome to Kaggle Mod\")\n",
"else:\n",
" notebook_env=3\n",
" print(\"Welcome!\")\n",
"\n",
"from IPython.display import clear_output, Javascript\n",
"\n",
"if notebook_env==1 and Use_Drive==True:\n",
" work_dir = \"/content/drive/MyDrive/vcclient\"\n",
" if not os.path.exists('/content/drive'):\n",
" drive.mount('/content/drive')\n",
"\n",
" if not os.path.exists(work_dir):\n",
" !mkdir -p {work_dir}\n",
"\n",
" print(\"Checking latest version...\")\n",
" if os.path.exists(f'{work_dir}/latest_hash.txt'):\n",
" current_version_hash = open(f'{work_dir}/latest_hash.txt').read().strip()\n",
" else:\n",
" current_version_hash = None\n",
"\n",
" !curl -s -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/latest_hash.txt -o latest_hash.txt\n",
" latest_version_hash = open('latest_hash.txt').read().strip()\n",
"\n",
" print(f\"current_version_hash: {current_version_hash}\")\n",
" print(f\"latest_version_hash : {latest_version_hash}\")\n",
"\n",
" if current_version_hash != latest_version_hash:\n",
" print(f\"hash not match -> download latest version\")\n",
"\n",
" latest_hash_path=f'{work_dir}/latest_hash.txt'\n",
2024-06-23 21:36:15 +03:00
"\n",
" if mode == \"elf\":\n",
" !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab -o {work_dir}/vcclient_latest_for_colab\n",
" elif mode == \"zip\":\n",
" !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab.zip -o {work_dir}/vcclient_latest_for_colab.zip\n",
"\n",
2024-06-22 10:00:37 +03:00
" !cp latest_hash.txt {latest_hash_path}\n",
2024-06-23 13:59:09 +03:00
" print(\"Download is done.\")\n",
2024-06-22 10:00:37 +03:00
" else:\n",
" print(\"hash matched. skip download\")\n",
"\n",
"else:\n",
" work_dir = \"/content\"\n",
" print(\"Downloading the latest vcclient... \")\n",
2024-06-23 13:59:09 +03:00
" !curl -s -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/latest_hash.txt -o latest_hash.txt\n",
" latest_version_hash = open('latest_hash.txt').read().strip()\n",
2024-06-23 21:36:15 +03:00
"\n",
" if mode == \"elf\":\n",
" !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab -o {work_dir}/vcclient_latest_for_colab\n",
" elif mode == \"zip\":\n",
" !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab.zip -o {work_dir}/vcclient_latest_for_colab.zip\n",
"\n",
2024-06-23 13:59:09 +03:00
" print(\"Download is done.\")\n",
2024-06-22 10:00:37 +03:00
"\n",
2024-06-23 21:36:15 +03:00
"if current_version_hash != latest_version_hash and mode == \"zip\":\n",
2024-06-23 13:59:09 +03:00
" print(f\"Unzip vcclient to {latest_version_hash} ... \")\n",
" !cd {work_dir} && unzip -q vcclient_latest_for_colab.zip -d {latest_version_hash}\n",
" print(f\"Unzip is done.\")\n",
"\n",
2024-06-23 21:36:15 +03:00
"if mode == \"elf\":\n",
2024-06-30 09:42:31 +03:00
" %cd {work_dir}\n",
2024-06-23 21:36:15 +03:00
" !chmod 0700 vcclient_latest_for_colab\n",
"elif mode == \"zip\":\n",
" %cd {work_dir}/{latest_version_hash}/main\n",
" !chmod 0700 main\n",
2024-06-23 13:59:09 +03:00
"\n",
"print(\"Installing modules... \",end=\"\")\n",
"!sudo apt-get install -y libportaudio2 > /dev/null 2>&1\n",
"!pip install pyngrok > /dev/null 2>&1\n",
"print(\"Install is done.\")\n",
2024-06-22 10:00:37 +03:00
"\n",
"Run_Cell=1\n"
]
},
{
"cell_type": "code",
2024-06-30 09:42:31 +03:00
"execution_count": 2,
2024-06-22 10:00:37 +03:00
"metadata": {
"id": "s7mYqKtW6VOI",
2024-06-30 09:42:31 +03:00
"outputId": "e86ad79b-cb62-4ea5-a8d9-dc7b9ca9e6ed",
2024-06-23 21:36:15 +03:00
"colab": {
2024-06-30 09:42:31 +03:00
"base_uri": "https://localhost:8080/",
"height": 148
},
"cellView": "form"
2024-06-22 10:00:37 +03:00
},
2024-06-23 21:36:15 +03:00
"outputs": [
{
"output_type": "stream",
"name": "stdout",
"text": [
"--------- SERVER READY! ---------\n",
2024-06-30 09:42:31 +03:00
"Your server is available. elapsed: 173sec\n",
"https://2bj201er26o-496ff2e9c6d22116-8003-colab.googleusercontent.com/\n",
2024-06-23 21:36:15 +03:00
"---------------------------------\n"
]
2024-06-30 09:42:31 +03:00
},
{
"output_type": "display_data",
"data": {
"text/plain": [
"<IPython.lib.display.Audio object>"
],
"text/html": [
"\n",
" <audio controls=\"controls\" autoplay=\"autoplay\">\n",
" <source src=\"https://huggingface.co/wok000/voices/resolve/main/vcclient001_vctk229_gpt-sovits_vcclient-ready.wav\" type=\"audio/x-wav\" />\n",
" Your browser does not support the audio element.\n",
" </audio>\n",
" "
]
},
"metadata": {}
2024-06-23 21:36:15 +03:00
}
],
2024-06-22 10:00:37 +03:00
"source": [
2024-06-30 09:42:31 +03:00
"# @title **[2]** Start server\n",
"# @markdown This cell will start the server, the first time that you run it will download the models, so it can take a while (2~4 minutes)\n",
"\n",
"#@markdown - Options:\n",
"ClearConsole = True # @param {type:\"boolean\"}\n",
"Play_Notification = True # @param {type:\"boolean\"}\n",
"\n",
"PORT=8003\n",
2024-06-22 10:00:37 +03:00
"\n",
2024-06-23 21:36:15 +03:00
"LOG_FILE = f\"/content/LOG_FILE_{PORT}\"\n",
2024-06-30 09:42:31 +03:00
"\n",
"from IPython.display import Audio, display\n",
"def play_notification_sound(url):\n",
" display(Audio(url=url, autoplay=True))\n",
"\n",
"from google.colab.output import eval_js\n",
"\n",
"\n",
2024-06-23 21:36:15 +03:00
"if mode == \"elf\":\n",
" # !LD_LIBRARY_PATH=/usr/lib64-nvidia:/usr/lib/x86_64-linux-gnu ./vcclient_latest_for_colab cui --port {PORT} --no_cui true &\n",
"\n",
" get_ipython().system_raw(f'LD_LIBRARY_PATH=/usr/lib64-nvidia:/usr/lib/x86_64-linux-gnu ./vcclient_latest_for_colab cui --port {PORT} --no_cui true >{LOG_FILE} 2>&1 &')\n",
"elif mode == \"zip\":\n",
" !LD_LIBRARY_PATH=/usr/lib64-nvidia:/usr/lib/x86_64-linux-gnu ./main cui --port {PORT} --no_cui true &\n",
"\n",
"\n",
"import socket\n",
"def wait_for_server():\n",
" elapsed_time = 0\n",
" start_time = time.time()\n",
"\n",
"\n",
" while True:\n",
" time.sleep(1)\n",
" sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n",
" result = sock.connect_ex(('127.0.0.1', PORT))\n",
" if result == 0:\n",
" break\n",
" sock.close()\n",
" # 時刻を出力\n",
" current_time = time.time()\n",
" elapsed_time = int(current_time - start_time)\n",
" clear_output(wait=True)\n",
" print(f\"Waiting for server... elapsed: {elapsed_time}sec\")\n",
" try:\n",
" with open(LOG_FILE, 'r') as f:\n",
" lines = f.readlines()[-5:]\n",
" for line in lines:\n",
" print(line.strip())\n",
" except:\n",
" pass\n",
"\n",
" if ClearConsole:\n",
" clear_output()\n",
" print(\"--------- SERVER READY! ---------\")\n",
" print(f\"Your server is available. elapsed: {elapsed_time}sec\")\n",
2024-06-30 09:42:31 +03:00
" proxy = eval_js( \"google.colab.kernel.proxyPort(\" + str(PORT) + \")\" )\n",
" print(f\"{proxy}\")\n",
2024-06-23 21:36:15 +03:00
" print(\"---------------------------------\")\n",
2024-06-30 09:42:31 +03:00
" if Play_Notification==True:\n",
" play_notification_sound('https://huggingface.co/wok000/voices/resolve/main/vcclient001_vctk229_gpt-sovits_vcclient-ready.wav')\n",
2024-06-23 21:36:15 +03:00
"wait_for_server()\n"
2024-06-22 10:00:37 +03:00
]
}
],
"metadata": {
"colab": {
"provenance": [],
2024-06-23 21:36:15 +03:00
"gpuType": "T4",
2024-06-30 09:42:31 +03:00
"authorship_tag": "ABX9TyO4xBCjHQ5g0B28Cfbnr1eo",
2024-06-22 10:00:37 +03:00
"include_colab_link": true
},
"kernelspec": {
"display_name": "Python 3",
"name": "python3"
},
"language_info": {
"name": "python"
2024-06-23 21:36:15 +03:00
},
"accelerator": "GPU"
2024-06-22 10:00:37 +03:00
},
"nbformat": 4,
"nbformat_minor": 0
}