Colaboratory を使用して作成しました

This commit is contained in:
w-okada 2023-01-11 23:21:23 +09:00
parent 91aa1bb931
commit 7750b75ba0

View File

@ -4,7 +4,7 @@
"metadata": {
"colab": {
"provenance": [],
"authorship_tag": "ABX9TyMVsTPyfpbt98oy7xm7MwHc",
"authorship_tag": "ABX9TyNML9VVT2rBjHi7qE5G+dci",
"include_colab_link": true
},
"kernelspec": {
@ -73,15 +73,15 @@
"base_uri": "https://localhost:8080/"
},
"id": "vV1t7PBRm-o6",
"outputId": "aa02d901-b65d-40bf-ff88-c06b88155031"
"outputId": "4604e7b1-cdeb-4cfc-cf1b-3922270c4d16"
},
"execution_count": null,
"execution_count": 1,
"outputs": [
{
"output_type": "stream",
"name": "stdout",
"text": [
"Wed Jan 11 08:52:42 2023 \n",
"Wed Jan 11 14:14:21 2023 \n",
"+-----------------------------------------------------------------------------+\n",
"| NVIDIA-SMI 460.32.03 Driver Version: 460.32.03 CUDA Version: 11.2 |\n",
"|-------------------------------+----------------------+----------------------+\n",
@ -90,7 +90,7 @@
"| | | MIG M. |\n",
"|===============================+======================+======================|\n",
"| 0 Tesla T4 Off | 00000000:00:04.0 Off | 0 |\n",
"| N/A 64C P8 12W / 70W | 0MiB / 15109MiB | 0% Default |\n",
"| N/A 39C P8 9W / 70W | 0MiB / 15109MiB | 0% Default |\n",
"| | | N/A |\n",
"+-------------------------------+----------------------+----------------------+\n",
" \n",
@ -119,7 +119,7 @@
"cell_type": "code",
"source": [
"# (2) リポジトリのクローン\n",
"!git clone --depth 1 https://github.com/w-okada/voice-changer.git -b v.1.3.0\n",
"!git clone --depth 1 https://github.com/w-okada/voice-changer.git -b v.1.3.1\n",
"%cd voice-changer/server\n",
"!git clone --depth 1 https://github.com/isletennos/MMVC_Trainer.git -b v1.3.2.2\n",
"!cd MMVC_Trainer/monotonic_align/ && python setup.py build_ext --inplace && cd -\n"
@ -129,9 +129,9 @@
"base_uri": "https://localhost:8080/"
},
"id": "86wTFmqsNMnD",
"outputId": "6bf049ac-02e3-43e9-95d3-77cd42c05eb1"
"outputId": "9be63d14-731d-41ce-9e8e-5c7e8abf0d93"
},
"execution_count": null,
"execution_count": 2,
"outputs": [
{
"output_type": "stream",
@ -140,9 +140,9 @@
"Cloning into 'voice-changer'...\n",
"remote: Enumerating objects: 138, done.\u001b[K\n",
"remote: Counting objects: 100% (138/138), done.\u001b[K\n",
"remote: Compressing objects: 100% (119/119), done.\u001b[K\n",
"remote: Total 138 (delta 18), reused 64 (delta 8), pack-reused 0\u001b[K\n",
"Receiving objects: 100% (138/138), 1.51 MiB | 4.72 MiB/s, done.\n",
"remote: Compressing objects: 100% (120/120), done.\u001b[K\n",
"remote: Total 138 (delta 18), reused 63 (delta 7), pack-reused 0\u001b[K\n",
"Receiving objects: 100% (138/138), 1.51 MiB | 4.06 MiB/s, done.\n",
"Resolving deltas: 100% (18/18), done.\n",
"/content/voice-changer/server\n",
"Cloning into 'MMVC_Trainer'...\n",
@ -150,7 +150,7 @@
"remote: Counting objects: 100% (920/920), done.\u001b[K\n",
"remote: Compressing objects: 100% (830/830), done.\u001b[K\n",
"remote: Total 920 (delta 4), reused 893 (delta 1), pack-reused 0\u001b[K\n",
"Receiving objects: 100% (920/920), 53.04 MiB | 33.98 MiB/s, done.\n",
"Receiving objects: 100% (920/920), 53.04 MiB | 13.66 MiB/s, done.\n",
"Resolving deltas: 100% (4/4), done.\n",
"Note: checking out 'f17c8c57d1ab7512633e6c57521f1eef6851bd0e'.\n",
"\n",
@ -206,40 +206,10 @@
"!pip install onnxruntime-gpu &> /dev/null"
],
"metadata": {
"id": "LwZAAuqxX7yY",
"outputId": "42f08570-62bf-4573-bd73-f0a3bb7f4f03",
"colab": {
"base_uri": "https://localhost:8080/"
}
"id": "LwZAAuqxX7yY"
},
"execution_count": null,
"outputs": [
{
"output_type": "stream",
"name": "stdout",
"text": [
"Looking in indexes: https://pypi.org/simple, https://us-python.pkg.dev/colab-wheels/public/simple/\n",
"Collecting onnxruntime-gpu\n",
" Downloading onnxruntime_gpu-1.13.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (115.3 MB)\n",
"\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m115.3/115.3 MB\u001b[0m \u001b[31m9.3 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n",
"\u001b[?25hRequirement already satisfied: numpy>=1.21.6 in /usr/local/lib/python3.8/dist-packages (from onnxruntime-gpu) (1.21.6)\n",
"Requirement already satisfied: packaging in /usr/local/lib/python3.8/dist-packages (from onnxruntime-gpu) (21.3)\n",
"Requirement already satisfied: protobuf in /usr/local/lib/python3.8/dist-packages (from onnxruntime-gpu) (3.19.6)\n",
"Requirement already satisfied: flatbuffers in /usr/local/lib/python3.8/dist-packages (from onnxruntime-gpu) (1.12)\n",
"Requirement already satisfied: sympy in /usr/local/lib/python3.8/dist-packages (from onnxruntime-gpu) (1.7.1)\n",
"Collecting coloredlogs\n",
" Downloading coloredlogs-15.0.1-py2.py3-none-any.whl (46 kB)\n",
"\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m46.0/46.0 KB\u001b[0m \u001b[31m7.6 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n",
"\u001b[?25hCollecting humanfriendly>=9.1\n",
" Downloading humanfriendly-10.0-py2.py3-none-any.whl (86 kB)\n",
"\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m86.8/86.8 KB\u001b[0m \u001b[31m13.7 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n",
"\u001b[?25hRequirement already satisfied: pyparsing!=3.0.5,>=2.0.2 in /usr/local/lib/python3.8/dist-packages (from packaging->onnxruntime-gpu) (3.0.9)\n",
"Requirement already satisfied: mpmath>=0.19 in /usr/local/lib/python3.8/dist-packages (from sympy->onnxruntime-gpu) (1.2.1)\n",
"Installing collected packages: humanfriendly, coloredlogs, onnxruntime-gpu\n",
"Successfully installed coloredlogs-15.0.1 humanfriendly-10.0 onnxruntime-gpu-1.13.1\n"
]
}
]
"execution_count": 3,
"outputs": []
},
{
"cell_type": "markdown",
@ -284,7 +254,7 @@
"metadata": {
"id": "G-nMdPxEW1rc"
},
"execution_count": null,
"execution_count": 4,
"outputs": []
},
{
@ -299,9 +269,9 @@
"base_uri": "https://localhost:8080/"
},
"id": "chu06KpAjEK6",
"outputId": "a714391f-616b-44ec-9290-c29246bc773f"
"outputId": "a5947a31-4e50-45d4-e7ae-e4bc2911b81c"
},
"execution_count": null,
"execution_count": 5,
"outputs": [
{
"output_type": "stream",
@ -336,7 +306,7 @@
"# (5) プロキシを起動\n",
"from google.colab.output import eval_js\n",
"proxy = eval_js( \"google.colab.kernel.proxyPort(\" + str(PORT) + \")\" )\n",
"print(f\"{proxy}front/\")"
"print(f\"{proxy}front/?colab=true\")"
],
"metadata": {
"id": "nkRjZm95l87C",
@ -344,15 +314,15 @@
"base_uri": "https://localhost:8080/",
"height": 34
},
"outputId": "8ca21533-b852-45d5-ca3c-19f0b5d8bf1f"
"outputId": "8c324bf3-e678-44cc-d67f-2d0f48c0b3fb"
},
"execution_count": null,
"execution_count": 6,
"outputs": [
{
"output_type": "stream",
"name": "stdout",
"text": [
"https://3n0321xyvvj-496ff2e9c6d22116-11464-colab.googleusercontent.com/front/\n"
"https://ucm1weit8m-496ff2e9c6d22116-14519-colab.googleusercontent.com/front/?colab=true\n"
]
}
]