mirror of
https://github.com/w-okada/voice-changer.git
synced 2025-02-09 03:37:51 +03:00
sending sample rate
This commit is contained in:
parent
2706810b4e
commit
e073eed95c
17
client/demo/dist/index.js
vendored
17
client/demo/dist/index.js
vendored
File diff suppressed because one or more lines are too long
@ -37,7 +37,6 @@ import { IndexRatioRow, IndexRatioRowProps } from "./components/609_IndexRatioRo
|
|||||||
import { RVCQualityRow, RVCQualityRowProps } from "./components/810_RVCQualityRow"
|
import { RVCQualityRow, RVCQualityRowProps } from "./components/810_RVCQualityRow"
|
||||||
import { ModelSamplingRateRow, ModelSamplingRateRowProps } from "./components/303_ModelSamplingRateRow"
|
import { ModelSamplingRateRow, ModelSamplingRateRowProps } from "./components/303_ModelSamplingRateRow"
|
||||||
// import { OnnxExportRow, OnnxExportRowProps } from "./components/304_OnnxExportRow"
|
// import { OnnxExportRow, OnnxExportRowProps } from "./components/304_OnnxExportRow"
|
||||||
import { SolaEnableRow, SolaEnableRowProps } from "./components/811_SolaEnableRow"
|
|
||||||
import { EnableEnhancerRow, EnableEnhancerRowProps } from "./components/610_EnableEnhancerRow"
|
import { EnableEnhancerRow, EnableEnhancerRowProps } from "./components/610_EnableEnhancerRow"
|
||||||
import { DstIdRow2, DstIdRow2Props } from "./components/602v2_DstIdRow2"
|
import { DstIdRow2, DstIdRow2Props } from "./components/602v2_DstIdRow2"
|
||||||
import { SilenceFrontRow, SilenceFrontRowProps } from "./components/812_SilenceFrontRow"
|
import { SilenceFrontRow, SilenceFrontRowProps } from "./components/812_SilenceFrontRow"
|
||||||
@ -107,7 +106,6 @@ const initialize = () => {
|
|||||||
addToCatalog("downSamplingMode", (props: DownSamplingModeRowProps) => { return <DownSamplingModeRow {...props} /> })
|
addToCatalog("downSamplingMode", (props: DownSamplingModeRowProps) => { return <DownSamplingModeRow {...props} /> })
|
||||||
addToCatalog("trancateNumThreshold", (props: TrancateNumTresholdRowProps) => { return <TrancateNumTresholdRow {...props} /> })
|
addToCatalog("trancateNumThreshold", (props: TrancateNumTresholdRowProps) => { return <TrancateNumTresholdRow {...props} /> })
|
||||||
addToCatalog("rvcQuality", (props: RVCQualityRowProps) => { return <RVCQualityRow {...props} /> })
|
addToCatalog("rvcQuality", (props: RVCQualityRowProps) => { return <RVCQualityRow {...props} /> })
|
||||||
addToCatalog("solaEnable", (props: SolaEnableRowProps) => { return <SolaEnableRow {...props} /> })
|
|
||||||
addToCatalog("silenceFront", (props: SilenceFrontRowProps) => { return <SilenceFrontRow {...props} /> })
|
addToCatalog("silenceFront", (props: SilenceFrontRowProps) => { return <SilenceFrontRow {...props} /> })
|
||||||
|
|
||||||
|
|
||||||
|
@ -12,10 +12,10 @@ export const NoiseScaleRow = (_props: NoiseScaleRowProps) => {
|
|||||||
<div className="body-row split-3-3-4 left-padding-1 guided">
|
<div className="body-row split-3-3-4 left-padding-1 guided">
|
||||||
<div className="body-item-title left-padding-1 ">Noise Scale</div>
|
<div className="body-item-title left-padding-1 ">Noise Scale</div>
|
||||||
<div>
|
<div>
|
||||||
<input type="range" className="body-item-input-slider" min="0" max="1" step="0.1" value={appState.serverSetting.serverSetting.noiceScale || 0} onChange={(e) => {
|
<input type="range" className="body-item-input-slider" min="0" max="1" step="0.1" value={appState.serverSetting.serverSetting.noiseScale || 0} onChange={(e) => {
|
||||||
appState.serverSetting.updateServerSettings({ ...appState.serverSetting.serverSetting, noiceScale: Number(e.target.value) })
|
appState.serverSetting.updateServerSettings({ ...appState.serverSetting.serverSetting, noiseScale: Number(e.target.value) })
|
||||||
}}></input>
|
}}></input>
|
||||||
<span className="body-item-input-slider-val">{appState.serverSetting.serverSetting.noiceScale}</span>
|
<span className="body-item-input-slider-val">{appState.serverSetting.serverSetting.noiseScale}</span>
|
||||||
</div>
|
</div>
|
||||||
<div className="body-button-container">
|
<div className="body-button-container">
|
||||||
</div>
|
</div>
|
||||||
|
@ -1,31 +0,0 @@
|
|||||||
import React, { useMemo } from "react"
|
|
||||||
import { useAppState } from "../../../001_provider/001_AppStateProvider"
|
|
||||||
|
|
||||||
export type SolaEnableRowProps = {
|
|
||||||
}
|
|
||||||
|
|
||||||
export const SolaEnableRow = (_props: SolaEnableRowProps) => {
|
|
||||||
const appState = useAppState()
|
|
||||||
|
|
||||||
const trancateNumTresholdRow = useMemo(() => {
|
|
||||||
const onSolaEnableChanged = (val: number) => {
|
|
||||||
appState.serverSetting.updateServerSettings({
|
|
||||||
...appState.serverSetting.serverSetting,
|
|
||||||
solaEnabled: val
|
|
||||||
})
|
|
||||||
}
|
|
||||||
return (
|
|
||||||
<div className="body-row split-3-7 left-padding-1 guided">
|
|
||||||
<div className="body-item-title left-padding-1">Sola enable</div>
|
|
||||||
<div className="body-input-container">
|
|
||||||
<select value={appState.serverSetting.serverSetting.solaEnabled} onChange={(e) => { onSolaEnableChanged(Number(e.target.value)) }}>
|
|
||||||
<option value="0" >disable</option>
|
|
||||||
<option value="1" >enable</option>
|
|
||||||
</select>
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
)
|
|
||||||
}, [appState.serverSetting.serverSetting, appState.serverSetting.updateServerSettings])
|
|
||||||
|
|
||||||
return trancateNumTresholdRow
|
|
||||||
}
|
|
@ -74,7 +74,6 @@ export const ServerSettingKey = {
|
|||||||
"crossFadeOffsetRate": "crossFadeOffsetRate",
|
"crossFadeOffsetRate": "crossFadeOffsetRate",
|
||||||
"crossFadeEndRate": "crossFadeEndRate",
|
"crossFadeEndRate": "crossFadeEndRate",
|
||||||
"crossFadeOverlapSize": "crossFadeOverlapSize",
|
"crossFadeOverlapSize": "crossFadeOverlapSize",
|
||||||
"solaEnabled": "solaEnabled",
|
|
||||||
|
|
||||||
"framework": "framework",
|
"framework": "framework",
|
||||||
"onnxExecutionProvider": "onnxExecutionProvider",
|
"onnxExecutionProvider": "onnxExecutionProvider",
|
||||||
@ -84,7 +83,7 @@ export const ServerSettingKey = {
|
|||||||
"recordIO": "recordIO",
|
"recordIO": "recordIO",
|
||||||
|
|
||||||
"tran": "tran",
|
"tran": "tran",
|
||||||
"noiceScale": "noiceScale",
|
"noiseScale": "noiseScale",
|
||||||
"predictF0": "predictF0",
|
"predictF0": "predictF0",
|
||||||
"silentThreshold": "silentThreshold",
|
"silentThreshold": "silentThreshold",
|
||||||
"extraConvertSize": "extraConvertSize",
|
"extraConvertSize": "extraConvertSize",
|
||||||
@ -111,7 +110,6 @@ export type VoiceChangerServerSetting = {
|
|||||||
crossFadeOffsetRate: number,
|
crossFadeOffsetRate: number,
|
||||||
crossFadeEndRate: number,
|
crossFadeEndRate: number,
|
||||||
crossFadeOverlapSize: CrossFadeOverlapSize,
|
crossFadeOverlapSize: CrossFadeOverlapSize,
|
||||||
solaEnabled: number,
|
|
||||||
|
|
||||||
framework: Framework
|
framework: Framework
|
||||||
onnxExecutionProvider: OnnxExecutionProvider,
|
onnxExecutionProvider: OnnxExecutionProvider,
|
||||||
@ -121,7 +119,7 @@ export type VoiceChangerServerSetting = {
|
|||||||
recordIO: number // 0:off, 1:on
|
recordIO: number // 0:off, 1:on
|
||||||
|
|
||||||
tran: number // so-vits-svc
|
tran: number // so-vits-svc
|
||||||
noiceScale: number // so-vits-svc
|
noiseScale: number // so-vits-svc
|
||||||
predictF0: number // so-vits-svc
|
predictF0: number // so-vits-svc
|
||||||
silentThreshold: number // so-vits-svc
|
silentThreshold: number // so-vits-svc
|
||||||
extraConvertSize: number// so-vits-svc
|
extraConvertSize: number// so-vits-svc
|
||||||
@ -149,248 +147,35 @@ export type ServerInfo = VoiceChangerServerSetting & {
|
|||||||
export type ServerInfoSoVitsSVC = ServerInfo & {
|
export type ServerInfoSoVitsSVC = ServerInfo & {
|
||||||
speakers: { [key: string]: number }
|
speakers: { [key: string]: number }
|
||||||
}
|
}
|
||||||
export const DefaultServerSetting_MMVCv15: ServerInfo = {
|
|
||||||
srcId: 0,
|
export const DefaultServerSetting: ServerInfo = {
|
||||||
dstId: 101,
|
// VC Common
|
||||||
gpu: 0,
|
inputSampleRate: 48000,
|
||||||
|
|
||||||
crossFadeOffsetRate: 0.0,
|
crossFadeOffsetRate: 0.0,
|
||||||
crossFadeEndRate: 1.0,
|
crossFadeEndRate: 1.0,
|
||||||
crossFadeOverlapSize: CrossFadeOverlapSize[1024],
|
crossFadeOverlapSize: CrossFadeOverlapSize[1024],
|
||||||
solaEnabled: 0,
|
|
||||||
|
|
||||||
framework: Framework.PyTorch,
|
|
||||||
f0Factor: 1.0,
|
|
||||||
onnxExecutionProvider: OnnxExecutionProvider.CPUExecutionProvider,
|
|
||||||
f0Detector: F0Detector.dio,
|
|
||||||
recordIO: 0,
|
recordIO: 0,
|
||||||
|
|
||||||
tran: 0,
|
|
||||||
noiceScale: 0,
|
|
||||||
predictF0: 0,
|
|
||||||
silentThreshold: 0,
|
|
||||||
extraConvertSize: 0,
|
|
||||||
clusterInferRatio: 0,
|
|
||||||
|
|
||||||
indexRatio: 0,
|
// VC Specific
|
||||||
rvcQuality: 0,
|
|
||||||
modelSamplingRate: 48000,
|
|
||||||
silenceFront: 0,
|
|
||||||
|
|
||||||
enableEnhancer: 0,
|
|
||||||
enhancerTune: 0,
|
|
||||||
|
|
||||||
inputSampleRate: 48000,
|
|
||||||
|
|
||||||
//
|
|
||||||
status: "ok",
|
|
||||||
configFile: "",
|
|
||||||
pyTorchModelFile: "",
|
|
||||||
onnxModelFile: "",
|
|
||||||
onnxExecutionProviders: []
|
|
||||||
}
|
|
||||||
|
|
||||||
export const DefaultServerSetting_MMVCv13: ServerInfo = {
|
|
||||||
srcId: 107,
|
|
||||||
dstId: 100,
|
|
||||||
gpu: 0,
|
|
||||||
|
|
||||||
crossFadeOffsetRate: 0.0,
|
|
||||||
crossFadeEndRate: 1.0,
|
|
||||||
crossFadeOverlapSize: CrossFadeOverlapSize[1024],
|
|
||||||
solaEnabled: 0,
|
|
||||||
|
|
||||||
framework: Framework.ONNX,
|
|
||||||
f0Factor: 1.0,
|
|
||||||
onnxExecutionProvider: OnnxExecutionProvider.CPUExecutionProvider,
|
|
||||||
f0Detector: F0Detector.dio,
|
|
||||||
recordIO: 0,
|
|
||||||
|
|
||||||
tran: 0,
|
|
||||||
noiceScale: 0,
|
|
||||||
predictF0: 0,
|
|
||||||
silentThreshold: 0,
|
|
||||||
extraConvertSize: 0,
|
|
||||||
clusterInferRatio: 0,
|
|
||||||
|
|
||||||
indexRatio: 0,
|
|
||||||
rvcQuality: 0,
|
|
||||||
modelSamplingRate: 48000,
|
|
||||||
silenceFront: 0,
|
|
||||||
|
|
||||||
enableEnhancer: 0,
|
|
||||||
enhancerTune: 0,
|
|
||||||
|
|
||||||
inputSampleRate: 48000,
|
|
||||||
|
|
||||||
//
|
|
||||||
status: "ok",
|
|
||||||
configFile: "",
|
|
||||||
pyTorchModelFile: "",
|
|
||||||
onnxModelFile: "",
|
|
||||||
onnxExecutionProviders: []
|
|
||||||
}
|
|
||||||
|
|
||||||
export const DefaultServerSetting_so_vits_svc_40: ServerInfo = {
|
|
||||||
srcId: 0,
|
|
||||||
dstId: 0,
|
|
||||||
gpu: 0,
|
|
||||||
|
|
||||||
crossFadeOffsetRate: 0.0,
|
|
||||||
crossFadeEndRate: 1.0,
|
|
||||||
crossFadeOverlapSize: CrossFadeOverlapSize[1024],
|
|
||||||
solaEnabled: 0,
|
|
||||||
|
|
||||||
framework: Framework.PyTorch,
|
|
||||||
f0Factor: 1.0,
|
|
||||||
onnxExecutionProvider: OnnxExecutionProvider.CPUExecutionProvider,
|
|
||||||
f0Detector: F0Detector.dio,
|
|
||||||
recordIO: 0,
|
|
||||||
|
|
||||||
// tran: 0,
|
|
||||||
// noiceScale: 0,
|
|
||||||
// predictF0: 0,
|
|
||||||
// silentThreshold: 0,
|
|
||||||
tran: 10,
|
|
||||||
noiceScale: 0.3,
|
|
||||||
predictF0: 0,
|
|
||||||
silentThreshold: 0.00001,
|
|
||||||
extraConvertSize: 1024 * 32,
|
|
||||||
clusterInferRatio: 0.1,
|
|
||||||
|
|
||||||
indexRatio: 0,
|
|
||||||
rvcQuality: 0,
|
|
||||||
modelSamplingRate: 48000,
|
|
||||||
silenceFront: 0,
|
|
||||||
|
|
||||||
enableEnhancer: 0,
|
|
||||||
enhancerTune: 0,
|
|
||||||
|
|
||||||
inputSampleRate: 48000,
|
|
||||||
|
|
||||||
//
|
|
||||||
status: "ok",
|
|
||||||
configFile: "",
|
|
||||||
pyTorchModelFile: "",
|
|
||||||
onnxModelFile: "",
|
|
||||||
onnxExecutionProviders: []
|
|
||||||
}
|
|
||||||
|
|
||||||
export const DefaultServerSetting_so_vits_svc_40_c: ServerInfo = {
|
|
||||||
srcId: 0,
|
|
||||||
dstId: 0,
|
|
||||||
gpu: 0,
|
|
||||||
|
|
||||||
crossFadeOffsetRate: 0.0,
|
|
||||||
crossFadeEndRate: 1.0,
|
|
||||||
crossFadeOverlapSize: CrossFadeOverlapSize[1024],
|
|
||||||
solaEnabled: 0,
|
|
||||||
|
|
||||||
framework: Framework.ONNX,
|
|
||||||
f0Factor: 1.0,
|
|
||||||
onnxExecutionProvider: OnnxExecutionProvider.CPUExecutionProvider,
|
|
||||||
f0Detector: F0Detector.dio,
|
|
||||||
recordIO: 0,
|
|
||||||
|
|
||||||
// tran: 0,
|
|
||||||
// noiceScale: 0,
|
|
||||||
// predictF0: 0,
|
|
||||||
// silentThreshold: 0,
|
|
||||||
tran: 10,
|
|
||||||
noiceScale: 0.3,
|
|
||||||
predictF0: 0,
|
|
||||||
silentThreshold: 0.00001,
|
|
||||||
extraConvertSize: 1024 * 32,
|
|
||||||
clusterInferRatio: 0.1,
|
|
||||||
|
|
||||||
indexRatio: 0,
|
|
||||||
rvcQuality: 0,
|
|
||||||
modelSamplingRate: 48000,
|
|
||||||
silenceFront: 0,
|
|
||||||
|
|
||||||
enableEnhancer: 0,
|
|
||||||
enhancerTune: 0,
|
|
||||||
|
|
||||||
inputSampleRate: 48000,
|
|
||||||
|
|
||||||
//
|
|
||||||
status: "ok",
|
|
||||||
configFile: "",
|
|
||||||
pyTorchModelFile: "",
|
|
||||||
onnxModelFile: "",
|
|
||||||
onnxExecutionProviders: []
|
|
||||||
}
|
|
||||||
export const DefaultServerSetting_so_vits_svc_40v2: ServerInfo = {
|
|
||||||
srcId: 0,
|
|
||||||
dstId: 0,
|
|
||||||
gpu: 0,
|
|
||||||
|
|
||||||
crossFadeOffsetRate: 0.0,
|
|
||||||
crossFadeEndRate: 1.0,
|
|
||||||
crossFadeOverlapSize: CrossFadeOverlapSize[1024],
|
|
||||||
solaEnabled: 0,
|
|
||||||
|
|
||||||
framework: Framework.PyTorch,
|
|
||||||
f0Factor: 1.0,
|
|
||||||
onnxExecutionProvider: OnnxExecutionProvider.CPUExecutionProvider,
|
|
||||||
f0Detector: F0Detector.dio,
|
|
||||||
recordIO: 0,
|
|
||||||
|
|
||||||
// tran: 0,
|
|
||||||
// noiceScale: 0,
|
|
||||||
// predictF0: 0,
|
|
||||||
// silentThreshold: 0,
|
|
||||||
tran: 10,
|
|
||||||
noiceScale: 0.3,
|
|
||||||
predictF0: 0,
|
|
||||||
silentThreshold: 0.00001,
|
|
||||||
extraConvertSize: 1024 * 32,
|
|
||||||
clusterInferRatio: 0.1,
|
|
||||||
|
|
||||||
indexRatio: 0,
|
|
||||||
rvcQuality: 0,
|
|
||||||
modelSamplingRate: 48000,
|
|
||||||
silenceFront: 0,
|
|
||||||
|
|
||||||
enableEnhancer: 0,
|
|
||||||
enhancerTune: 0,
|
|
||||||
|
|
||||||
inputSampleRate: 48000,
|
|
||||||
|
|
||||||
//
|
|
||||||
status: "ok",
|
|
||||||
configFile: "",
|
|
||||||
pyTorchModelFile: "",
|
|
||||||
onnxModelFile: "",
|
|
||||||
onnxExecutionProviders: []
|
|
||||||
}
|
|
||||||
|
|
||||||
export const DefaultServerSetting_DDSP_SVC: ServerInfo = {
|
|
||||||
srcId: 0,
|
srcId: 0,
|
||||||
dstId: 1,
|
dstId: 1,
|
||||||
gpu: 0,
|
gpu: 0,
|
||||||
|
|
||||||
crossFadeOffsetRate: 0.0,
|
|
||||||
crossFadeEndRate: 1.0,
|
|
||||||
crossFadeOverlapSize: CrossFadeOverlapSize[1024],
|
|
||||||
solaEnabled: 0,
|
|
||||||
|
|
||||||
framework: Framework.PyTorch,
|
framework: Framework.PyTorch,
|
||||||
f0Factor: 1.0,
|
f0Factor: 1.0,
|
||||||
onnxExecutionProvider: OnnxExecutionProvider.CPUExecutionProvider,
|
onnxExecutionProvider: OnnxExecutionProvider.CPUExecutionProvider,
|
||||||
f0Detector: F0Detector.dio,
|
f0Detector: F0Detector.dio,
|
||||||
recordIO: 0,
|
|
||||||
|
|
||||||
// tran: 0,
|
tran: 0,
|
||||||
// noiceScale: 0,
|
noiseScale: 0,
|
||||||
// predictF0: 0,
|
|
||||||
// silentThreshold: 0,
|
|
||||||
tran: 10,
|
|
||||||
noiceScale: 0.3,
|
|
||||||
predictF0: 0,
|
predictF0: 0,
|
||||||
silentThreshold: 0.00001,
|
silentThreshold: 0,
|
||||||
extraConvertSize: 1024 * 32,
|
extraConvertSize: 0,
|
||||||
clusterInferRatio: 0.1,
|
clusterInferRatio: 0,
|
||||||
|
|
||||||
indexRatio: 0,
|
indexRatio: 0,
|
||||||
rvcQuality: 0,
|
rvcQuality: 0,
|
||||||
@ -400,7 +185,6 @@ export const DefaultServerSetting_DDSP_SVC: ServerInfo = {
|
|||||||
enableEnhancer: 0,
|
enableEnhancer: 0,
|
||||||
enhancerTune: 0,
|
enhancerTune: 0,
|
||||||
|
|
||||||
inputSampleRate: 48000,
|
|
||||||
|
|
||||||
//
|
//
|
||||||
status: "ok",
|
status: "ok",
|
||||||
@ -409,51 +193,31 @@ export const DefaultServerSetting_DDSP_SVC: ServerInfo = {
|
|||||||
onnxModelFile: "",
|
onnxModelFile: "",
|
||||||
onnxExecutionProviders: []
|
onnxExecutionProviders: []
|
||||||
}
|
}
|
||||||
|
export const DefaultServerSetting_MMVCv15: ServerInfo = {
|
||||||
|
...DefaultServerSetting, dstId: 101,
|
||||||
|
}
|
||||||
|
export const DefaultServerSetting_MMVCv13: ServerInfo = {
|
||||||
|
...DefaultServerSetting, srcId: 107, dstId: 100,
|
||||||
|
}
|
||||||
|
|
||||||
|
export const DefaultServerSetting_so_vits_svc_40: ServerInfo = {
|
||||||
|
...DefaultServerSetting, tran: 10, noiseScale: 0.3, extraConvertSize: 1024 * 32, clusterInferRatio: 0.1,
|
||||||
|
}
|
||||||
|
|
||||||
|
export const DefaultServerSetting_so_vits_svc_40_c: ServerInfo = {
|
||||||
|
...DefaultServerSetting, tran: 10, noiseScale: 0.3, extraConvertSize: 1024 * 32, clusterInferRatio: 0.1,
|
||||||
|
}
|
||||||
|
export const DefaultServerSetting_so_vits_svc_40v2: ServerInfo = {
|
||||||
|
...DefaultServerSetting, tran: 10, noiseScale: 0.3, extraConvertSize: 1024 * 32, clusterInferRatio: 0.1,
|
||||||
|
}
|
||||||
|
|
||||||
|
export const DefaultServerSetting_DDSP_SVC: ServerInfo = {
|
||||||
|
...DefaultServerSetting, dstId: 1, tran: 10, extraConvertSize: 1024 * 32, clusterInferRatio: 0.1, enableEnhancer: 1
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
export const DefaultServerSetting_RVC: ServerInfo = {
|
export const DefaultServerSetting_RVC: ServerInfo = {
|
||||||
srcId: 0,
|
...DefaultServerSetting, tran: 10, extraConvertSize: 1024 * 32,
|
||||||
dstId: 0,
|
|
||||||
gpu: 0,
|
|
||||||
|
|
||||||
crossFadeOffsetRate: 0.1,
|
|
||||||
crossFadeEndRate: 0.8,
|
|
||||||
crossFadeOverlapSize: CrossFadeOverlapSize[1024],
|
|
||||||
solaEnabled: 1,
|
|
||||||
|
|
||||||
framework: Framework.PyTorch,
|
|
||||||
f0Factor: 1.0,
|
|
||||||
onnxExecutionProvider: OnnxExecutionProvider.CPUExecutionProvider,
|
|
||||||
f0Detector: F0Detector.harvest,
|
|
||||||
recordIO: 0,
|
|
||||||
|
|
||||||
// tran: 0,
|
|
||||||
// noiceScale: 0,
|
|
||||||
// predictF0: 0,
|
|
||||||
// silentThreshold: 0,
|
|
||||||
tran: 10,
|
|
||||||
noiceScale: 0.3,
|
|
||||||
predictF0: 0,
|
|
||||||
silentThreshold: 0.00001,
|
|
||||||
extraConvertSize: 1024 * 32,
|
|
||||||
clusterInferRatio: 0.1,
|
|
||||||
|
|
||||||
indexRatio: 0,
|
|
||||||
rvcQuality: 0,
|
|
||||||
modelSamplingRate: 48000,
|
|
||||||
silenceFront: 1,
|
|
||||||
|
|
||||||
enableEnhancer: 0,
|
|
||||||
enhancerTune: 0,
|
|
||||||
|
|
||||||
inputSampleRate: 48000,
|
|
||||||
|
|
||||||
//
|
|
||||||
status: "ok",
|
|
||||||
configFile: "",
|
|
||||||
pyTorchModelFile: "",
|
|
||||||
onnxModelFile: "",
|
|
||||||
onnxExecutionProviders: []
|
|
||||||
}
|
}
|
||||||
|
|
||||||
///////////////////////
|
///////////////////////
|
||||||
@ -503,41 +267,25 @@ export type WorkletNodeSetting = {
|
|||||||
export const DefaultWorkletNodeSetting: WorkletNodeSetting = {
|
export const DefaultWorkletNodeSetting: WorkletNodeSetting = {
|
||||||
serverUrl: "",
|
serverUrl: "",
|
||||||
protocol: "sio",
|
protocol: "sio",
|
||||||
sendingSampleRate: 24000,
|
sendingSampleRate: 48000,
|
||||||
inputChunkNum: 48,
|
inputChunkNum: 48,
|
||||||
downSamplingMode: "average"
|
downSamplingMode: "average"
|
||||||
}
|
}
|
||||||
|
|
||||||
export const DefaultWorkletNodeSetting_so_vits_svc_40: WorkletNodeSetting = {
|
export const DefaultWorkletNodeSetting_so_vits_svc_40: WorkletNodeSetting = {
|
||||||
serverUrl: "",
|
...DefaultWorkletNodeSetting, inputChunkNum: 128,
|
||||||
protocol: "sio",
|
|
||||||
sendingSampleRate: 24000,
|
|
||||||
inputChunkNum: 128,
|
|
||||||
downSamplingMode: "average"
|
|
||||||
}
|
}
|
||||||
|
|
||||||
export const DefaultWorkletNodeSetting_so_vits_svc_40v2: WorkletNodeSetting = {
|
export const DefaultWorkletNodeSetting_so_vits_svc_40v2: WorkletNodeSetting = {
|
||||||
serverUrl: "",
|
...DefaultWorkletNodeSetting, inputChunkNum: 128,
|
||||||
protocol: "sio",
|
|
||||||
sendingSampleRate: 24000,
|
|
||||||
inputChunkNum: 128,
|
|
||||||
downSamplingMode: "average"
|
|
||||||
}
|
}
|
||||||
|
|
||||||
export const DefaultWorkletNodeSetting_DDSP_SVC: WorkletNodeSetting = {
|
export const DefaultWorkletNodeSetting_DDSP_SVC: WorkletNodeSetting = {
|
||||||
serverUrl: "",
|
...DefaultWorkletNodeSetting, inputChunkNum: 256,
|
||||||
protocol: "sio",
|
|
||||||
sendingSampleRate: 48000,
|
|
||||||
inputChunkNum: 256,
|
|
||||||
downSamplingMode: "average"
|
|
||||||
}
|
}
|
||||||
|
|
||||||
export const DefaultWorkletNodeSetting_RVC: WorkletNodeSetting = {
|
export const DefaultWorkletNodeSetting_RVC: WorkletNodeSetting = {
|
||||||
serverUrl: "",
|
...DefaultWorkletNodeSetting, inputChunkNum: 256,
|
||||||
protocol: "sio",
|
|
||||||
sendingSampleRate: 48000,
|
|
||||||
inputChunkNum: 256,
|
|
||||||
downSamplingMode: "average"
|
|
||||||
}
|
}
|
||||||
|
|
||||||
///////////////////////
|
///////////////////////
|
||||||
|
@ -98,7 +98,7 @@ export const useServerSetting = (props: UseServerSettingProps): ServerSettingSta
|
|||||||
const cachedServerSetting = await getItem(INDEXEDDB_KEY_SERVER)
|
const cachedServerSetting = await getItem(INDEXEDDB_KEY_SERVER)
|
||||||
let initialSetting: ServerInfo
|
let initialSetting: ServerInfo
|
||||||
if (cachedServerSetting) {
|
if (cachedServerSetting) {
|
||||||
initialSetting = { ...defaultServerSetting, ...cachedServerSetting as ServerInfo }
|
initialSetting = { ...defaultServerSetting, ...cachedServerSetting as ServerInfo, inputSampleRate: 48000 }// sample rateは時限措置
|
||||||
console.log("Initial Setting1:", initialSetting)
|
console.log("Initial Setting1:", initialSetting)
|
||||||
} else {
|
} else {
|
||||||
initialSetting = { ...defaultServerSetting }
|
initialSetting = { ...defaultServerSetting }
|
||||||
|
@ -45,7 +45,7 @@ export const useWorkletNodeSetting = (props: UseWorkletNodeSettingProps): Workle
|
|||||||
const loadCache = async () => {
|
const loadCache = async () => {
|
||||||
const setting = await getItem(INDEXEDDB_KEY_WORKLETNODE) as WorkletNodeSetting
|
const setting = await getItem(INDEXEDDB_KEY_WORKLETNODE) as WorkletNodeSetting
|
||||||
if (setting) {
|
if (setting) {
|
||||||
_setWorkletNodeSetting(setting)
|
_setWorkletNodeSetting({ ...setting, sendingSampleRate: 48000 }) // sample rateは時限措置
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
loadCache()
|
loadCache()
|
||||||
|
@ -51,7 +51,7 @@ class DDSP_SVCSettings():
|
|||||||
|
|
||||||
# ↓mutableな物だけ列挙
|
# ↓mutableな物だけ列挙
|
||||||
intData = ["gpu", "dstId", "tran", "predictF0", "extraConvertSize", "enableEnhancer", "enhancerTune"]
|
intData = ["gpu", "dstId", "tran", "predictF0", "extraConvertSize", "enableEnhancer", "enhancerTune"]
|
||||||
floatData = ["noiceScale", "silentThreshold", "clusterInferRatio"]
|
floatData = ["silentThreshold", "clusterInferRatio"]
|
||||||
strData = ["framework", "f0Detector"]
|
strData = ["framework", "f0Detector"]
|
||||||
|
|
||||||
|
|
||||||
@ -229,33 +229,7 @@ class DDSP_SVC:
|
|||||||
print("[Voice Changer] No onnx session.")
|
print("[Voice Changer] No onnx session.")
|
||||||
raise NoModeLoadedException("ONNX")
|
raise NoModeLoadedException("ONNX")
|
||||||
|
|
||||||
seg_units = data[0]
|
raise NoModeLoadedException("ONNX")
|
||||||
# f0 = data[1]
|
|
||||||
# convertSize = data[2]
|
|
||||||
# vol = data[3]
|
|
||||||
|
|
||||||
if vol < self.settings.silentThreshold:
|
|
||||||
return np.zeros(convertSize).astype(np.int16)
|
|
||||||
|
|
||||||
c, f0, uv = [x.numpy() for x in data]
|
|
||||||
audio1 = self.onnx_session.run(
|
|
||||||
["audio"],
|
|
||||||
{
|
|
||||||
"c": c,
|
|
||||||
"f0": f0,
|
|
||||||
"g": np.array([self.settings.dstId]).astype(np.int64),
|
|
||||||
"uv": np.array([self.settings.dstId]).astype(np.int64),
|
|
||||||
"predict_f0": np.array([self.settings.dstId]).astype(np.int64),
|
|
||||||
"noice_scale": np.array([self.settings.dstId]).astype(np.int64),
|
|
||||||
|
|
||||||
|
|
||||||
})[0][0, 0] * self.hps.data.max_wav_value
|
|
||||||
|
|
||||||
audio1 = audio1 * vol
|
|
||||||
|
|
||||||
result = audio1
|
|
||||||
|
|
||||||
return result
|
|
||||||
|
|
||||||
def _pyTorch_inference(self, data):
|
def _pyTorch_inference(self, data):
|
||||||
|
|
||||||
|
@ -41,7 +41,6 @@ class RVCSettings():
|
|||||||
|
|
||||||
f0Detector: str = "pm" # pm or harvest
|
f0Detector: str = "pm" # pm or harvest
|
||||||
tran: int = 20
|
tran: int = 20
|
||||||
noiceScale: float = 0.3
|
|
||||||
predictF0: int = 0 # 0:False, 1:True
|
predictF0: int = 0 # 0:False, 1:True
|
||||||
silentThreshold: float = 0.00001
|
silentThreshold: float = 0.00001
|
||||||
extraConvertSize: int = 1024 * 32
|
extraConvertSize: int = 1024 * 32
|
||||||
@ -63,7 +62,7 @@ class RVCSettings():
|
|||||||
|
|
||||||
# ↓mutableな物だけ列挙
|
# ↓mutableな物だけ列挙
|
||||||
intData = ["gpu", "dstId", "tran", "predictF0", "extraConvertSize", "rvcQuality", "modelSamplingRate", "silenceFront"]
|
intData = ["gpu", "dstId", "tran", "predictF0", "extraConvertSize", "rvcQuality", "modelSamplingRate", "silenceFront"]
|
||||||
floatData = ["noiceScale", "silentThreshold", "indexRatio"]
|
floatData = ["silentThreshold", "indexRatio"]
|
||||||
strData = ["framework", "f0Detector"]
|
strData = ["framework", "f0Detector"]
|
||||||
|
|
||||||
|
|
||||||
|
@ -40,7 +40,7 @@ class SoVitsSvc40Settings():
|
|||||||
|
|
||||||
f0Detector: str = "dio" # dio or harvest
|
f0Detector: str = "dio" # dio or harvest
|
||||||
tran: int = 20
|
tran: int = 20
|
||||||
noiceScale: float = 0.3
|
noiseScale: float = 0.3
|
||||||
predictF0: int = 0 # 0:False, 1:True
|
predictF0: int = 0 # 0:False, 1:True
|
||||||
silentThreshold: float = 0.00001
|
silentThreshold: float = 0.00001
|
||||||
extraConvertSize: int = 1024 * 32
|
extraConvertSize: int = 1024 * 32
|
||||||
@ -57,7 +57,7 @@ class SoVitsSvc40Settings():
|
|||||||
|
|
||||||
# ↓mutableな物だけ列挙
|
# ↓mutableな物だけ列挙
|
||||||
intData = ["gpu", "dstId", "tran", "predictF0", "extraConvertSize"]
|
intData = ["gpu", "dstId", "tran", "predictF0", "extraConvertSize"]
|
||||||
floatData = ["noiceScale", "silentThreshold", "clusterInferRatio"]
|
floatData = ["noiseScale", "silentThreshold", "clusterInferRatio"]
|
||||||
strData = ["framework", "f0Detector"]
|
strData = ["framework", "f0Detector"]
|
||||||
|
|
||||||
|
|
||||||
@ -308,7 +308,7 @@ class SoVitsSvc40:
|
|||||||
"f0": f0.astype(np.float32),
|
"f0": f0.astype(np.float32),
|
||||||
"uv": uv.astype(np.float32),
|
"uv": uv.astype(np.float32),
|
||||||
"g": sid_target.astype(np.int64),
|
"g": sid_target.astype(np.int64),
|
||||||
"noice_scale": np.array([self.settings.noiceScale]).astype(np.float32),
|
"noise_scale": np.array([self.settings.noiseScale]).astype(np.float32),
|
||||||
# "predict_f0": np.array([self.settings.dstId]).astype(np.int64),
|
# "predict_f0": np.array([self.settings.dstId]).astype(np.int64),
|
||||||
|
|
||||||
|
|
||||||
@ -342,7 +342,7 @@ class SoVitsSvc40:
|
|||||||
# audio1 = self.net_g.infer(c, f0=f0, g=sid_target, uv=uv, predict_f0=True, noice_scale=0.1)[0][0, 0].data.float()
|
# audio1 = self.net_g.infer(c, f0=f0, g=sid_target, uv=uv, predict_f0=True, noice_scale=0.1)[0][0, 0].data.float()
|
||||||
predict_f0_flag = True if self.settings.predictF0 == 1 else False
|
predict_f0_flag = True if self.settings.predictF0 == 1 else False
|
||||||
audio1 = self.net_g.infer(c, f0=f0, g=sid_target, uv=uv, predict_f0=predict_f0_flag,
|
audio1 = self.net_g.infer(c, f0=f0, g=sid_target, uv=uv, predict_f0=predict_f0_flag,
|
||||||
noice_scale=self.settings.noiceScale)
|
noice_scale=self.settings.noiseScale)
|
||||||
audio1 = audio1[0][0].data.float()
|
audio1 = audio1[0][0].data.float()
|
||||||
# audio1 = self.net_g.infer(c, f0=f0, g=sid_target, uv=uv, predict_f0=predict_f0_flag,
|
# audio1 = self.net_g.infer(c, f0=f0, g=sid_target, uv=uv, predict_f0=predict_f0_flag,
|
||||||
# noice_scale=self.settings.noiceScale)[0][0, 0].data.float()
|
# noice_scale=self.settings.noiceScale)[0][0, 0].data.float()
|
||||||
|
@ -36,7 +36,7 @@ class SoVitsSvc40v2Settings():
|
|||||||
|
|
||||||
f0Detector: str = "dio" # dio or harvest
|
f0Detector: str = "dio" # dio or harvest
|
||||||
tran: int = 20
|
tran: int = 20
|
||||||
noiceScale: float = 0.3
|
noiseScale: float = 0.3
|
||||||
predictF0: int = 0 # 0:False, 1:True
|
predictF0: int = 0 # 0:False, 1:True
|
||||||
silentThreshold: float = 0.00001
|
silentThreshold: float = 0.00001
|
||||||
extraConvertSize: int = 1024 * 32
|
extraConvertSize: int = 1024 * 32
|
||||||
@ -53,7 +53,7 @@ class SoVitsSvc40v2Settings():
|
|||||||
|
|
||||||
# ↓mutableな物だけ列挙
|
# ↓mutableな物だけ列挙
|
||||||
intData = ["gpu", "dstId", "tran", "predictF0", "extraConvertSize"]
|
intData = ["gpu", "dstId", "tran", "predictF0", "extraConvertSize"]
|
||||||
floatData = ["noiceScale", "silentThreshold", "clusterInferRatio"]
|
floatData = ["noiseScale", "silentThreshold", "clusterInferRatio"]
|
||||||
strData = ["framework", "f0Detector"]
|
strData = ["framework", "f0Detector"]
|
||||||
|
|
||||||
|
|
||||||
@ -300,7 +300,7 @@ class SoVitsSvc40v2:
|
|||||||
# audio1 = self.net_g.infer(c, f0=f0, g=sid_target, uv=uv, predict_f0=True, noice_scale=0.1)[0][0, 0].data.float()
|
# audio1 = self.net_g.infer(c, f0=f0, g=sid_target, uv=uv, predict_f0=True, noice_scale=0.1)[0][0, 0].data.float()
|
||||||
predict_f0_flag = True if self.settings.predictF0 == 1 else False
|
predict_f0_flag = True if self.settings.predictF0 == 1 else False
|
||||||
audio1 = self.net_g.infer(c, f0=f0, g=sid_target, uv=uv, predict_f0=predict_f0_flag,
|
audio1 = self.net_g.infer(c, f0=f0, g=sid_target, uv=uv, predict_f0=predict_f0_flag,
|
||||||
noice_scale=self.settings.noiceScale)[0][0, 0].data.float()
|
noice_scale=self.settings.noiseScale)[0][0, 0].data.float()
|
||||||
audio1 = audio1 * self.hps.data.max_wav_value
|
audio1 = audio1 * self.hps.data.max_wav_value
|
||||||
|
|
||||||
audio1 = audio1 * vol
|
audio1 = audio1 * vol
|
||||||
|
@ -31,13 +31,12 @@ class VoiceChangerSettings():
|
|||||||
crossFadeOffsetRate: float = 0.1
|
crossFadeOffsetRate: float = 0.1
|
||||||
crossFadeEndRate: float = 0.9
|
crossFadeEndRate: float = 0.9
|
||||||
crossFadeOverlapSize: int = 4096
|
crossFadeOverlapSize: int = 4096
|
||||||
solaEnabled: int = 1 # 0:off, 1:on
|
|
||||||
|
|
||||||
recordIO: int = 0 # 0:off, 1:on
|
recordIO: int = 0 # 0:off, 1:on
|
||||||
|
|
||||||
# ↓mutableな物だけ列挙
|
# ↓mutableな物だけ列挙
|
||||||
intData: list[str] = field(
|
intData: list[str] = field(
|
||||||
default_factory=lambda: ["inputSampleRate", "crossFadeOverlapSize", "recordIO", "solaEnabled"]
|
default_factory=lambda: ["inputSampleRate", "crossFadeOverlapSize", "recordIO"]
|
||||||
)
|
)
|
||||||
floatData: list[str] = field(
|
floatData: list[str] = field(
|
||||||
default_factory=lambda: ["crossFadeOffsetRate", "crossFadeEndRate"]
|
default_factory=lambda: ["crossFadeOffsetRate", "crossFadeEndRate"]
|
||||||
|
Loading…
Reference in New Issue
Block a user