Spaces:
Running
on
Zero
Running
on
Zero
Upload 10 files
Browse files- README.md +12 -12
- app.py +4 -7
- pre-requirements.txt +2 -1
- requirements.txt +7 -5
- src/download_models.py +12 -5
- src/infer_pack/models.py +4 -4
- src/main.py +112 -27
- src/mdx.py +7 -3
- src/rvc.py +15 -4
- src/webui.py +82 -27
README.md
CHANGED
|
@@ -1,13 +1,13 @@
|
|
| 1 |
-
---
|
| 2 |
-
title: AICoverGen
|
| 3 |
-
emoji: 🚀
|
| 4 |
-
colorFrom: red
|
| 5 |
-
colorTo: pink
|
| 6 |
-
sdk: gradio
|
| 7 |
-
sdk_version: 5.
|
| 8 |
-
app_file: app.py
|
| 9 |
-
pinned: false
|
| 10 |
-
license: mit
|
| 11 |
-
---
|
| 12 |
-
|
| 13 |
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
|
|
|
|
| 1 |
+
---
|
| 2 |
+
title: AICoverGen
|
| 3 |
+
emoji: 🚀
|
| 4 |
+
colorFrom: red
|
| 5 |
+
colorTo: pink
|
| 6 |
+
sdk: gradio
|
| 7 |
+
sdk_version: 5.44.0
|
| 8 |
+
app_file: app.py
|
| 9 |
+
pinned: false
|
| 10 |
+
license: mit
|
| 11 |
+
---
|
| 12 |
+
|
| 13 |
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
|
app.py
CHANGED
|
@@ -1,12 +1,9 @@
|
|
| 1 |
import os
|
|
|
|
| 2 |
|
| 3 |
-
|
| 4 |
|
| 5 |
-
|
| 6 |
-
|
| 7 |
-
|
| 8 |
-
python src/download_models.py
|
| 9 |
-
python src/webui.py
|
| 10 |
-
"""
|
| 11 |
|
| 12 |
os.system(cmd)
|
|
|
|
| 1 |
import os
|
| 2 |
+
import sys
|
| 3 |
|
| 4 |
+
os.system("python src/download_models.py")
|
| 5 |
|
| 6 |
+
args = " ".join(sys.argv[1:])
|
| 7 |
+
cmd = f"python src/webui.py {args}"
|
|
|
|
|
|
|
|
|
|
|
|
|
| 8 |
|
| 9 |
os.system(cmd)
|
pre-requirements.txt
CHANGED
|
@@ -1 +1,2 @@
|
|
| 1 |
-
pip
|
|
|
|
|
|
| 1 |
+
pip==23.0.1
|
| 2 |
+
Setuptools<=80.6.0
|
requirements.txt
CHANGED
|
@@ -1,4 +1,3 @@
|
|
| 1 |
-
--extra-index-url=https://download.pytorch.org/whl/cu121
|
| 2 |
torch==2.5.1
|
| 3 |
torchvision==0.20.1
|
| 4 |
torchaudio==2.5.1
|
|
@@ -6,11 +5,11 @@ deemix
|
|
| 6 |
fairseq==0.12.2
|
| 7 |
faiss-cpu==1.7.3
|
| 8 |
ffmpeg-python>=0.2.0
|
| 9 |
-
|
| 10 |
lib==4.0.0
|
| 11 |
librosa==0.9.1
|
| 12 |
numpy==1.23.5
|
| 13 |
-
# onnxruntime #onnxruntime_gpu
|
| 14 |
praat-parselmouth>=0.4.2
|
| 15 |
pedalboard==0.7.7
|
| 16 |
pydub==0.25.1
|
|
@@ -20,5 +19,8 @@ scipy==1.11.1
|
|
| 20 |
soundfile==0.12.1
|
| 21 |
torchcrepe==0.0.20
|
| 22 |
tqdm==4.65.0
|
| 23 |
-
yt_dlp
|
| 24 |
-
sox==1.4.1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
torch==2.5.1
|
| 2 |
torchvision==0.20.1
|
| 3 |
torchaudio==2.5.1
|
|
|
|
| 5 |
fairseq==0.12.2
|
| 6 |
faiss-cpu==1.7.3
|
| 7 |
ffmpeg-python>=0.2.0
|
| 8 |
+
gradio==5.44.0
|
| 9 |
lib==4.0.0
|
| 10 |
librosa==0.9.1
|
| 11 |
numpy==1.23.5
|
| 12 |
+
onnxruntime-gpu==1.22.0 # onnxruntime #onnxruntime_gpu
|
| 13 |
praat-parselmouth>=0.4.2
|
| 14 |
pedalboard==0.7.7
|
| 15 |
pydub==0.25.1
|
|
|
|
| 19 |
soundfile==0.12.1
|
| 20 |
torchcrepe==0.0.20
|
| 21 |
tqdm==4.65.0
|
| 22 |
+
yt_dlp
|
| 23 |
+
sox==1.4.1
|
| 24 |
+
noisereduce
|
| 25 |
+
spaces
|
| 26 |
+
matplotlib-inline
|
src/download_models.py
CHANGED
|
@@ -8,11 +8,20 @@ BASE_DIR = Path(__file__).resolve().parent.parent
|
|
| 8 |
mdxnet_models_dir = BASE_DIR / 'mdxnet_models'
|
| 9 |
rvc_models_dir = BASE_DIR / 'rvc_models'
|
| 10 |
|
|
|
|
|
|
|
|
|
|
| 11 |
|
| 12 |
def dl_model(link, model_name, dir_name):
|
| 13 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 14 |
r.raise_for_status()
|
| 15 |
-
with open(
|
| 16 |
for chunk in r.iter_content(chunk_size=8192):
|
| 17 |
f.write(chunk)
|
| 18 |
|
|
@@ -20,12 +29,10 @@ def dl_model(link, model_name, dir_name):
|
|
| 20 |
if __name__ == '__main__':
|
| 21 |
mdx_model_names = ['UVR-MDX-NET-Inst_HQ_4.onnx', 'UVR-MDX-NET-Voc_FT.onnx', 'UVR_MDXNET_KARA_2.onnx', 'Reverb_HQ_By_FoxJoy.onnx']
|
| 22 |
for model in mdx_model_names:
|
| 23 |
-
print(f'Downloading {model}...')
|
| 24 |
dl_model(MDX_DOWNLOAD_LINK, model, mdxnet_models_dir)
|
| 25 |
|
| 26 |
rvc_model_names = ['hubert_base.pt', 'rmvpe.pt']
|
| 27 |
for model in rvc_model_names:
|
| 28 |
-
print(f'Downloading {model}...')
|
| 29 |
dl_model(RVC_DOWNLOAD_LINK, model, rvc_models_dir)
|
| 30 |
|
| 31 |
-
print('All models
|
|
|
|
| 8 |
mdxnet_models_dir = BASE_DIR / 'mdxnet_models'
|
| 9 |
rvc_models_dir = BASE_DIR / 'rvc_models'
|
| 10 |
|
| 11 |
+
mdxnet_models_dir.mkdir(parents=True, exist_ok=True)
|
| 12 |
+
rvc_models_dir.mkdir(parents=True, exist_ok=True)
|
| 13 |
+
|
| 14 |
|
| 15 |
def dl_model(link, model_name, dir_name):
|
| 16 |
+
model_path = dir_name / model_name
|
| 17 |
+
if model_path.exists():
|
| 18 |
+
# print(f"{model_name} already exists, skipping download.")
|
| 19 |
+
return
|
| 20 |
+
|
| 21 |
+
print(f"Downloading {model_name}...")
|
| 22 |
+
with requests.get(f'{link}{model_name}', stream=True) as r:
|
| 23 |
r.raise_for_status()
|
| 24 |
+
with open(model_path, 'wb') as f:
|
| 25 |
for chunk in r.iter_content(chunk_size=8192):
|
| 26 |
f.write(chunk)
|
| 27 |
|
|
|
|
| 29 |
if __name__ == '__main__':
|
| 30 |
mdx_model_names = ['UVR-MDX-NET-Inst_HQ_4.onnx', 'UVR-MDX-NET-Voc_FT.onnx', 'UVR_MDXNET_KARA_2.onnx', 'Reverb_HQ_By_FoxJoy.onnx']
|
| 31 |
for model in mdx_model_names:
|
|
|
|
| 32 |
dl_model(MDX_DOWNLOAD_LINK, model, mdxnet_models_dir)
|
| 33 |
|
| 34 |
rvc_model_names = ['hubert_base.pt', 'rmvpe.pt']
|
| 35 |
for model in rvc_model_names:
|
|
|
|
| 36 |
dl_model(RVC_DOWNLOAD_LINK, model, rvc_models_dir)
|
| 37 |
|
| 38 |
+
print('All models ready!')
|
src/infer_pack/models.py
CHANGED
|
@@ -607,7 +607,7 @@ class SynthesizerTrnMs256NSFsid(nn.Module):
|
|
| 607 |
inter_channels, hidden_channels, 5, 1, 3, gin_channels=gin_channels
|
| 608 |
)
|
| 609 |
self.emb_g = nn.Embedding(self.spk_embed_dim, gin_channels)
|
| 610 |
-
print("gin_channels:", gin_channels, "self.spk_embed_dim:", self.spk_embed_dim)
|
| 611 |
|
| 612 |
def remove_weight_norm(self):
|
| 613 |
self.dec.remove_weight_norm()
|
|
@@ -718,7 +718,7 @@ class SynthesizerTrnMs768NSFsid(nn.Module):
|
|
| 718 |
inter_channels, hidden_channels, 5, 1, 3, gin_channels=gin_channels
|
| 719 |
)
|
| 720 |
self.emb_g = nn.Embedding(self.spk_embed_dim, gin_channels)
|
| 721 |
-
print("gin_channels:", gin_channels, "self.spk_embed_dim:", self.spk_embed_dim)
|
| 722 |
|
| 723 |
def remove_weight_norm(self):
|
| 724 |
self.dec.remove_weight_norm()
|
|
@@ -826,7 +826,7 @@ class SynthesizerTrnMs256NSFsid_nono(nn.Module):
|
|
| 826 |
inter_channels, hidden_channels, 5, 1, 3, gin_channels=gin_channels
|
| 827 |
)
|
| 828 |
self.emb_g = nn.Embedding(self.spk_embed_dim, gin_channels)
|
| 829 |
-
print("gin_channels:", gin_channels, "self.spk_embed_dim:", self.spk_embed_dim)
|
| 830 |
|
| 831 |
def remove_weight_norm(self):
|
| 832 |
self.dec.remove_weight_norm()
|
|
@@ -928,7 +928,7 @@ class SynthesizerTrnMs768NSFsid_nono(nn.Module):
|
|
| 928 |
inter_channels, hidden_channels, 5, 1, 3, gin_channels=gin_channels
|
| 929 |
)
|
| 930 |
self.emb_g = nn.Embedding(self.spk_embed_dim, gin_channels)
|
| 931 |
-
print("gin_channels:", gin_channels, "self.spk_embed_dim:", self.spk_embed_dim)
|
| 932 |
|
| 933 |
def remove_weight_norm(self):
|
| 934 |
self.dec.remove_weight_norm()
|
|
|
|
| 607 |
inter_channels, hidden_channels, 5, 1, 3, gin_channels=gin_channels
|
| 608 |
)
|
| 609 |
self.emb_g = nn.Embedding(self.spk_embed_dim, gin_channels)
|
| 610 |
+
# print("gin_channels:", gin_channels, "self.spk_embed_dim:", self.spk_embed_dim)
|
| 611 |
|
| 612 |
def remove_weight_norm(self):
|
| 613 |
self.dec.remove_weight_norm()
|
|
|
|
| 718 |
inter_channels, hidden_channels, 5, 1, 3, gin_channels=gin_channels
|
| 719 |
)
|
| 720 |
self.emb_g = nn.Embedding(self.spk_embed_dim, gin_channels)
|
| 721 |
+
# print("gin_channels:", gin_channels, "self.spk_embed_dim:", self.spk_embed_dim)
|
| 722 |
|
| 723 |
def remove_weight_norm(self):
|
| 724 |
self.dec.remove_weight_norm()
|
|
|
|
| 826 |
inter_channels, hidden_channels, 5, 1, 3, gin_channels=gin_channels
|
| 827 |
)
|
| 828 |
self.emb_g = nn.Embedding(self.spk_embed_dim, gin_channels)
|
| 829 |
+
# print("gin_channels:", gin_channels, "self.spk_embed_dim:", self.spk_embed_dim)
|
| 830 |
|
| 831 |
def remove_weight_norm(self):
|
| 832 |
self.dec.remove_weight_norm()
|
|
|
|
| 928 |
inter_channels, hidden_channels, 5, 1, 3, gin_channels=gin_channels
|
| 929 |
)
|
| 930 |
self.emb_g = nn.Embedding(self.spk_embed_dim, gin_channels)
|
| 931 |
+
# print("gin_channels:", gin_channels, "self.spk_embed_dim:", self.spk_embed_dim)
|
| 932 |
|
| 933 |
def remove_weight_norm(self):
|
| 934 |
self.dec.remove_weight_norm()
|
src/main.py
CHANGED
|
@@ -9,6 +9,8 @@ import shlex
|
|
| 9 |
import subprocess
|
| 10 |
from contextlib import suppress
|
| 11 |
from urllib.parse import urlparse, parse_qs
|
|
|
|
|
|
|
| 12 |
|
| 13 |
import gradio as gr
|
| 14 |
import librosa
|
|
@@ -19,6 +21,7 @@ import yt_dlp
|
|
| 19 |
from pedalboard import Pedalboard, Reverb, Compressor, HighpassFilter
|
| 20 |
from pedalboard.io import AudioFile
|
| 21 |
from pydub import AudioSegment
|
|
|
|
| 22 |
|
| 23 |
from mdx import run_mdx
|
| 24 |
from rvc import Config, load_hubert, get_vc, rvc_infer
|
|
@@ -27,12 +30,29 @@ import logging
|
|
| 27 |
logging.getLogger("httpx").setLevel(logging.WARNING)
|
| 28 |
|
| 29 |
BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
|
|
|
|
| 30 |
|
| 31 |
mdxnet_models_dir = os.path.join(BASE_DIR, 'mdxnet_models')
|
| 32 |
rvc_models_dir = os.path.join(BASE_DIR, 'rvc_models')
|
| 33 |
output_dir = os.path.join(BASE_DIR, 'song_output')
|
| 34 |
|
| 35 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 36 |
def get_youtube_video_id(url, ignore_playlist=True):
|
| 37 |
"""
|
| 38 |
Examples:
|
|
@@ -68,6 +88,9 @@ def get_youtube_video_id(url, ignore_playlist=True):
|
|
| 68 |
|
| 69 |
|
| 70 |
def yt_download(link):
|
|
|
|
|
|
|
|
|
|
| 71 |
ydl_opts = {
|
| 72 |
'format': 'bestaudio',
|
| 73 |
'outtmpl': '%(title)s',
|
|
@@ -95,12 +118,12 @@ def raise_exception(error_msg, is_webui):
|
|
| 95 |
def get_rvc_model(voice_model, is_webui):
|
| 96 |
rvc_model_filename, rvc_index_filename = None, None
|
| 97 |
model_dir = os.path.join(rvc_models_dir, voice_model)
|
| 98 |
-
print(model_dir)
|
| 99 |
for file in os.listdir(model_dir):
|
| 100 |
-
print(file)
|
| 101 |
if os.path.isdir(file):
|
| 102 |
for ff in os.listdir(file):
|
| 103 |
-
print("subfile", ff)
|
| 104 |
ext = os.path.splitext(ff)[1]
|
| 105 |
if ext == '.pth':
|
| 106 |
rvc_model_filename = ff
|
|
@@ -136,9 +159,21 @@ def get_audio_paths(song_dir):
|
|
| 136 |
elif file.endswith('_Vocals_Backup.wav'):
|
| 137 |
backup_vocals_path = os.path.join(song_dir, file)
|
| 138 |
|
|
|
|
| 139 |
return orig_song_path, instrumentals_path, main_vocals_dereverb_path, backup_vocals_path
|
| 140 |
|
| 141 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 142 |
def convert_to_stereo(audio_path):
|
| 143 |
wave, sr = librosa.load(audio_path, mono=False, sr=44100)
|
| 144 |
|
|
@@ -216,7 +251,7 @@ hubert_model = load_hubert("cuda", config.is_half, os.path.join(rvc_models_dir,
|
|
| 216 |
print(device, "half>>", config.is_half)
|
| 217 |
|
| 218 |
# @spaces.GPU(enable_queue=True)
|
| 219 |
-
def voice_change(voice_model, vocals_path, output_path, pitch_change, f0_method, index_rate, filter_radius, rms_mix_rate, protect, crepe_hop_length, is_webui):
|
| 220 |
rvc_model_path, rvc_index_path = get_rvc_model(voice_model, is_webui)
|
| 221 |
|
| 222 |
device = "cuda:0" if torch.cuda.is_available() else "cpu"
|
|
@@ -227,8 +262,8 @@ def voice_change(voice_model, vocals_path, output_path, pitch_change, f0_method,
|
|
| 227 |
|
| 228 |
# convert main vocals
|
| 229 |
global hubert_model
|
| 230 |
-
rvc_infer(rvc_index_path, index_rate, vocals_path, output_path, pitch_change, f0_method, cpt, version, net_g, filter_radius, tgt_sr, rms_mix_rate, protect, crepe_hop_length, vc, hubert_model)
|
| 231 |
-
del
|
| 232 |
gc.collect()
|
| 233 |
|
| 234 |
|
|
@@ -267,9 +302,9 @@ def combine_audio(audio_paths, output_path, main_gain, backup_gain, inst_gain, o
|
|
| 267 |
def process_song(
|
| 268 |
song_dir, song_input, mdx_model_params, song_id, is_webui, input_type, progress,
|
| 269 |
keep_files, pitch_change, pitch_change_all, voice_model, index_rate, filter_radius,
|
| 270 |
-
rms_mix_rate, protect, f0_method, crepe_hop_length, output_format, keep_orig, orig_song_path
|
| 271 |
):
|
| 272 |
-
|
| 273 |
if not os.path.exists(song_dir):
|
| 274 |
os.makedirs(song_dir)
|
| 275 |
orig_song_path, vocals_path, instrumentals_path, main_vocals_path, backup_vocals_path, main_vocals_dereverb_path = preprocess_song(song_input, mdx_model_params, song_id, is_webui, input_type, progress, keep_orig, orig_song_path)
|
|
@@ -278,29 +313,72 @@ def process_song(
|
|
| 278 |
paths = get_audio_paths(song_dir)
|
| 279 |
|
| 280 |
# if any of the audio files aren't available or keep intermediate files, rerun preprocess
|
| 281 |
-
if any(path is None for path in paths)
|
| 282 |
orig_song_path, vocals_path, instrumentals_path, main_vocals_path, backup_vocals_path, main_vocals_dereverb_path = preprocess_song(song_input, mdx_model_params, song_id, is_webui, input_type, progress, keep_orig, orig_song_path)
|
| 283 |
else:
|
| 284 |
orig_song_path, instrumentals_path, main_vocals_dereverb_path, backup_vocals_path = paths
|
| 285 |
|
| 286 |
pitch_change = pitch_change * 12 + pitch_change_all
|
| 287 |
-
ai_vocals_path = os.path.join(song_dir, f'{os.path.splitext(os.path.basename(orig_song_path))[0]}_{voice_model}_p{pitch_change}_i{index_rate}_fr{filter_radius}_rms{rms_mix_rate}_pro{protect}_{f0_method}{"" if f0_method != "mangio-crepe" else f"_{crepe_hop_length}"}.wav')
|
| 288 |
ai_cover_path = os.path.join(song_dir, f'{os.path.splitext(os.path.basename(orig_song_path))[0]} ({voice_model} Ver).{output_format}')
|
| 289 |
|
| 290 |
if not os.path.exists(ai_vocals_path):
|
| 291 |
display_progress('[~] Converting voice using RVC...', 0.5, is_webui, progress)
|
| 292 |
-
voice_change(voice_model, main_vocals_dereverb_path, ai_vocals_path, pitch_change, f0_method, index_rate, filter_radius, rms_mix_rate, protect, crepe_hop_length, is_webui)
|
| 293 |
|
| 294 |
return ai_vocals_path, ai_cover_path, instrumentals_path, backup_vocals_path, vocals_path, main_vocals_path
|
| 295 |
|
| 296 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 297 |
|
| 298 |
# @spaces.GPU(duration=140)
|
| 299 |
def song_cover_pipeline(song_input, voice_model, pitch_change, keep_files,
|
| 300 |
is_webui=0, main_gain=0, backup_gain=0, inst_gain=0, index_rate=0.5, filter_radius=3,
|
| 301 |
rms_mix_rate=0.25, f0_method='rmvpe', crepe_hop_length=128, protect=0.33, pitch_change_all=0,
|
| 302 |
reverb_rm_size=0.15, reverb_wet=0.2, reverb_dry=0.8, reverb_damping=0.7, output_format='mp3',
|
|
|
|
| 303 |
progress=gr.Progress()):
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 304 |
try:
|
| 305 |
if not song_input or not voice_model:
|
| 306 |
raise_exception('Ensure that the song input field and voice model field is filled.', is_webui)
|
|
@@ -334,9 +412,8 @@ def song_cover_pipeline(song_input, voice_model, pitch_change, keep_files,
|
|
| 334 |
keep_orig, orig_song_path = get_audio_file(song_input, is_webui, input_type, progress)
|
| 335 |
orig_song_path = convert_to_stereo(orig_song_path)
|
| 336 |
|
| 337 |
-
import time
|
| 338 |
start = time.time()
|
| 339 |
-
|
| 340 |
(
|
| 341 |
ai_vocals_path,
|
| 342 |
ai_cover_path,
|
|
@@ -365,6 +442,7 @@ def song_cover_pipeline(song_input, voice_model, pitch_change, keep_files,
|
|
| 365 |
output_format,
|
| 366 |
keep_orig,
|
| 367 |
orig_song_path,
|
|
|
|
| 368 |
)
|
| 369 |
|
| 370 |
end = time.time()
|
|
@@ -374,20 +452,27 @@ def song_cover_pipeline(song_input, voice_model, pitch_change, keep_files,
|
|
| 374 |
print(f"Audio duration: {duration__:.2f} seconds")
|
| 375 |
|
| 376 |
display_progress('[~] Applying audio effects to Vocals...', 0.8, is_webui, progress)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 377 |
ai_vocals_mixed_path = add_audio_effects(ai_vocals_path, reverb_rm_size, reverb_wet, reverb_dry, reverb_damping)
|
| 378 |
|
| 379 |
-
|
| 380 |
-
|
| 381 |
-
|
| 382 |
-
|
| 383 |
-
|
| 384 |
-
|
| 385 |
-
|
| 386 |
-
|
| 387 |
-
|
| 388 |
-
|
| 389 |
-
|
| 390 |
-
|
|
|
|
|
|
|
| 391 |
|
| 392 |
if pitch_change_all != 0:
|
| 393 |
display_progress('[~] Applying overall pitch change', 0.85, is_webui, progress)
|
|
@@ -399,7 +484,7 @@ def song_cover_pipeline(song_input, voice_model, pitch_change, keep_files,
|
|
| 399 |
|
| 400 |
if not keep_files:
|
| 401 |
display_progress('[~] Removing intermediate audio files...', 0.95, is_webui, progress)
|
| 402 |
-
intermediate_files = [vocals_path, main_vocals_path, ai_vocals_mixed_path]
|
| 403 |
if pitch_change_all != 0:
|
| 404 |
intermediate_files += [instrumentals_path, backup_vocals_path]
|
| 405 |
for file in intermediate_files:
|
|
|
|
| 9 |
import subprocess
|
| 10 |
from contextlib import suppress
|
| 11 |
from urllib.parse import urlparse, parse_qs
|
| 12 |
+
import time
|
| 13 |
+
import shutil
|
| 14 |
|
| 15 |
import gradio as gr
|
| 16 |
import librosa
|
|
|
|
| 21 |
from pedalboard import Pedalboard, Reverb, Compressor, HighpassFilter
|
| 22 |
from pedalboard.io import AudioFile
|
| 23 |
from pydub import AudioSegment
|
| 24 |
+
import noisereduce as nr
|
| 25 |
|
| 26 |
from mdx import run_mdx
|
| 27 |
from rvc import Config, load_hubert, get_vc, rvc_infer
|
|
|
|
| 30 |
logging.getLogger("httpx").setLevel(logging.WARNING)
|
| 31 |
|
| 32 |
BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
|
| 33 |
+
IS_ZERO_GPU = os.getenv("SPACES_ZERO_GPU")
|
| 34 |
|
| 35 |
mdxnet_models_dir = os.path.join(BASE_DIR, 'mdxnet_models')
|
| 36 |
rvc_models_dir = os.path.join(BASE_DIR, 'rvc_models')
|
| 37 |
output_dir = os.path.join(BASE_DIR, 'song_output')
|
| 38 |
|
| 39 |
|
| 40 |
+
def clean_old_folders(base_path: str, max_age_seconds: int = 10800):
|
| 41 |
+
if not os.path.isdir(base_path):
|
| 42 |
+
print(f"Error: {base_path} is not a valid directory.")
|
| 43 |
+
return
|
| 44 |
+
|
| 45 |
+
now = time.time()
|
| 46 |
+
|
| 47 |
+
for folder_name in os.listdir(base_path):
|
| 48 |
+
folder_path = os.path.join(base_path, folder_name)
|
| 49 |
+
if os.path.isdir(folder_path):
|
| 50 |
+
last_modified = os.path.getmtime(folder_path)
|
| 51 |
+
if now - last_modified > max_age_seconds:
|
| 52 |
+
# print(f"Deleting folder: {folder_path}")
|
| 53 |
+
shutil.rmtree(folder_path)
|
| 54 |
+
|
| 55 |
+
|
| 56 |
def get_youtube_video_id(url, ignore_playlist=True):
|
| 57 |
"""
|
| 58 |
Examples:
|
|
|
|
| 88 |
|
| 89 |
|
| 90 |
def yt_download(link):
|
| 91 |
+
if not link.strip():
|
| 92 |
+
gr.Info("You need to provide a download link.")
|
| 93 |
+
return None
|
| 94 |
ydl_opts = {
|
| 95 |
'format': 'bestaudio',
|
| 96 |
'outtmpl': '%(title)s',
|
|
|
|
| 118 |
def get_rvc_model(voice_model, is_webui):
|
| 119 |
rvc_model_filename, rvc_index_filename = None, None
|
| 120 |
model_dir = os.path.join(rvc_models_dir, voice_model)
|
| 121 |
+
# print(model_dir)
|
| 122 |
for file in os.listdir(model_dir):
|
| 123 |
+
# print(file)
|
| 124 |
if os.path.isdir(file):
|
| 125 |
for ff in os.listdir(file):
|
| 126 |
+
# print("subfile", ff)
|
| 127 |
ext = os.path.splitext(ff)[1]
|
| 128 |
if ext == '.pth':
|
| 129 |
rvc_model_filename = ff
|
|
|
|
| 159 |
elif file.endswith('_Vocals_Backup.wav'):
|
| 160 |
backup_vocals_path = os.path.join(song_dir, file)
|
| 161 |
|
| 162 |
+
# print(orig_song_path, instrumentals_path, main_vocals_dereverb_path, backup_vocals_path)
|
| 163 |
return orig_song_path, instrumentals_path, main_vocals_dereverb_path, backup_vocals_path
|
| 164 |
|
| 165 |
|
| 166 |
+
def get_audio_with_suffix(song_dir, suffix="_mysuffix.wav"):
|
| 167 |
+
target_path = None
|
| 168 |
+
|
| 169 |
+
for file in os.listdir(song_dir):
|
| 170 |
+
if file.endswith(suffix):
|
| 171 |
+
target_path = os.path.join(song_dir, file)
|
| 172 |
+
break
|
| 173 |
+
|
| 174 |
+
return target_path
|
| 175 |
+
|
| 176 |
+
|
| 177 |
def convert_to_stereo(audio_path):
|
| 178 |
wave, sr = librosa.load(audio_path, mono=False, sr=44100)
|
| 179 |
|
|
|
|
| 251 |
print(device, "half>>", config.is_half)
|
| 252 |
|
| 253 |
# @spaces.GPU(enable_queue=True)
|
| 254 |
+
def voice_change(voice_model, vocals_path, output_path, pitch_change, f0_method, index_rate, filter_radius, rms_mix_rate, protect, crepe_hop_length, is_webui, steps):
|
| 255 |
rvc_model_path, rvc_index_path = get_rvc_model(voice_model, is_webui)
|
| 256 |
|
| 257 |
device = "cuda:0" if torch.cuda.is_available() else "cpu"
|
|
|
|
| 262 |
|
| 263 |
# convert main vocals
|
| 264 |
global hubert_model
|
| 265 |
+
rvc_infer(rvc_index_path, index_rate, vocals_path, output_path, pitch_change, f0_method, cpt, version, net_g, filter_radius, tgt_sr, rms_mix_rate, protect, crepe_hop_length, vc, hubert_model, steps)
|
| 266 |
+
del cpt
|
| 267 |
gc.collect()
|
| 268 |
|
| 269 |
|
|
|
|
| 302 |
def process_song(
|
| 303 |
song_dir, song_input, mdx_model_params, song_id, is_webui, input_type, progress,
|
| 304 |
keep_files, pitch_change, pitch_change_all, voice_model, index_rate, filter_radius,
|
| 305 |
+
rms_mix_rate, protect, f0_method, crepe_hop_length, output_format, keep_orig, orig_song_path, steps
|
| 306 |
):
|
| 307 |
+
|
| 308 |
if not os.path.exists(song_dir):
|
| 309 |
os.makedirs(song_dir)
|
| 310 |
orig_song_path, vocals_path, instrumentals_path, main_vocals_path, backup_vocals_path, main_vocals_dereverb_path = preprocess_song(song_input, mdx_model_params, song_id, is_webui, input_type, progress, keep_orig, orig_song_path)
|
|
|
|
| 313 |
paths = get_audio_paths(song_dir)
|
| 314 |
|
| 315 |
# if any of the audio files aren't available or keep intermediate files, rerun preprocess
|
| 316 |
+
if any(path is None for path in paths):
|
| 317 |
orig_song_path, vocals_path, instrumentals_path, main_vocals_path, backup_vocals_path, main_vocals_dereverb_path = preprocess_song(song_input, mdx_model_params, song_id, is_webui, input_type, progress, keep_orig, orig_song_path)
|
| 318 |
else:
|
| 319 |
orig_song_path, instrumentals_path, main_vocals_dereverb_path, backup_vocals_path = paths
|
| 320 |
|
| 321 |
pitch_change = pitch_change * 12 + pitch_change_all
|
| 322 |
+
ai_vocals_path = os.path.join(song_dir, f'{os.path.splitext(os.path.basename(orig_song_path))[0]}_{voice_model}_p{pitch_change}_i{index_rate}_fr{filter_radius}_rms{rms_mix_rate}_pro{protect}_{f0_method}{"" if f0_method != "mangio-crepe" else f"_{crepe_hop_length}"}_s{steps}.wav')
|
| 323 |
ai_cover_path = os.path.join(song_dir, f'{os.path.splitext(os.path.basename(orig_song_path))[0]} ({voice_model} Ver).{output_format}')
|
| 324 |
|
| 325 |
if not os.path.exists(ai_vocals_path):
|
| 326 |
display_progress('[~] Converting voice using RVC...', 0.5, is_webui, progress)
|
| 327 |
+
voice_change(voice_model, main_vocals_dereverb_path, ai_vocals_path, pitch_change, f0_method, index_rate, filter_radius, rms_mix_rate, protect, crepe_hop_length, is_webui, steps)
|
| 328 |
|
| 329 |
return ai_vocals_path, ai_cover_path, instrumentals_path, backup_vocals_path, vocals_path, main_vocals_path
|
| 330 |
|
| 331 |
+
|
| 332 |
+
def apply_noisereduce(audio_list, type_output="wav"):
|
| 333 |
+
# https://github.com/sa-if/Audio-Denoiser
|
| 334 |
+
print("Noice reduce")
|
| 335 |
+
|
| 336 |
+
result = []
|
| 337 |
+
for audio_path in audio_list:
|
| 338 |
+
out_path = f"{os.path.splitext(audio_path)[0]}_nr.{type_output}"
|
| 339 |
+
|
| 340 |
+
try:
|
| 341 |
+
# Load audio file
|
| 342 |
+
audio = AudioSegment.from_file(audio_path)
|
| 343 |
+
|
| 344 |
+
# Convert audio to numpy array
|
| 345 |
+
samples = np.array(audio.get_array_of_samples())
|
| 346 |
+
|
| 347 |
+
# Reduce noise
|
| 348 |
+
reduced_noise = nr.reduce_noise(samples, sr=audio.frame_rate, prop_decrease=0.6)
|
| 349 |
+
|
| 350 |
+
# Convert reduced noise signal back to audio
|
| 351 |
+
reduced_audio = AudioSegment(
|
| 352 |
+
reduced_noise.tobytes(),
|
| 353 |
+
frame_rate=audio.frame_rate,
|
| 354 |
+
sample_width=audio.sample_width,
|
| 355 |
+
channels=audio.channels
|
| 356 |
+
)
|
| 357 |
+
|
| 358 |
+
# Save reduced audio to file
|
| 359 |
+
reduced_audio.export(out_path, format=type_output)
|
| 360 |
+
result.append(out_path)
|
| 361 |
+
|
| 362 |
+
except Exception as e:
|
| 363 |
+
print(f"Error noisereduce: {str(e)}")
|
| 364 |
+
result.append(audio_path)
|
| 365 |
+
|
| 366 |
+
return result
|
| 367 |
+
|
| 368 |
|
| 369 |
# @spaces.GPU(duration=140)
|
| 370 |
def song_cover_pipeline(song_input, voice_model, pitch_change, keep_files,
|
| 371 |
is_webui=0, main_gain=0, backup_gain=0, inst_gain=0, index_rate=0.5, filter_radius=3,
|
| 372 |
rms_mix_rate=0.25, f0_method='rmvpe', crepe_hop_length=128, protect=0.33, pitch_change_all=0,
|
| 373 |
reverb_rm_size=0.15, reverb_wet=0.2, reverb_dry=0.8, reverb_damping=0.7, output_format='mp3',
|
| 374 |
+
extra_denoise=False, steps=1,
|
| 375 |
progress=gr.Progress()):
|
| 376 |
+
if not keep_files or IS_ZERO_GPU:
|
| 377 |
+
clean_old_folders("./song_output", 14400)
|
| 378 |
+
|
| 379 |
+
if IS_ZERO_GPU:
|
| 380 |
+
clean_old_folders("./rvc_models", 10800)
|
| 381 |
+
|
| 382 |
try:
|
| 383 |
if not song_input or not voice_model:
|
| 384 |
raise_exception('Ensure that the song input field and voice model field is filled.', is_webui)
|
|
|
|
| 412 |
keep_orig, orig_song_path = get_audio_file(song_input, is_webui, input_type, progress)
|
| 413 |
orig_song_path = convert_to_stereo(orig_song_path)
|
| 414 |
|
|
|
|
| 415 |
start = time.time()
|
| 416 |
+
|
| 417 |
(
|
| 418 |
ai_vocals_path,
|
| 419 |
ai_cover_path,
|
|
|
|
| 442 |
output_format,
|
| 443 |
keep_orig,
|
| 444 |
orig_song_path,
|
| 445 |
+
steps,
|
| 446 |
)
|
| 447 |
|
| 448 |
end = time.time()
|
|
|
|
| 452 |
print(f"Audio duration: {duration__:.2f} seconds")
|
| 453 |
|
| 454 |
display_progress('[~] Applying audio effects to Vocals...', 0.8, is_webui, progress)
|
| 455 |
+
|
| 456 |
+
nr_path = ai_vocals_path # get_audio_with_suffix(song_dir, "_nr.wav")
|
| 457 |
+
if extra_denoise:
|
| 458 |
+
ai_vocals_path = apply_noisereduce([ai_vocals_path])[0]
|
| 459 |
+
|
| 460 |
ai_vocals_mixed_path = add_audio_effects(ai_vocals_path, reverb_rm_size, reverb_wet, reverb_dry, reverb_damping)
|
| 461 |
|
| 462 |
+
ins_path = get_audio_with_suffix(song_dir, "_Voiceless.wav")
|
| 463 |
+
if not ins_path:
|
| 464 |
+
instrumentals_path, _ = run_mdx(
|
| 465 |
+
mdx_model_params,
|
| 466 |
+
os.path.join(output_dir, song_id),
|
| 467 |
+
os.path.join(mdxnet_models_dir, "UVR-MDX-NET-Inst_HQ_4.onnx"),
|
| 468 |
+
instrumentals_path,
|
| 469 |
+
# exclude_main=False,
|
| 470 |
+
exclude_inversion=True,
|
| 471 |
+
suffix="Voiceless",
|
| 472 |
+
denoise=False,
|
| 473 |
+
keep_orig=True,
|
| 474 |
+
base_device=("" if IS_ZERO_GPU else "cuda")
|
| 475 |
+
)
|
| 476 |
|
| 477 |
if pitch_change_all != 0:
|
| 478 |
display_progress('[~] Applying overall pitch change', 0.85, is_webui, progress)
|
|
|
|
| 484 |
|
| 485 |
if not keep_files:
|
| 486 |
display_progress('[~] Removing intermediate audio files...', 0.95, is_webui, progress)
|
| 487 |
+
intermediate_files = [vocals_path, main_vocals_path, ai_vocals_mixed_path, ins_path, nr_path]
|
| 488 |
if pitch_change_all != 0:
|
| 489 |
intermediate_files += [instrumentals_path, backup_vocals_path]
|
| 490 |
for file in intermediate_files:
|
src/mdx.py
CHANGED
|
@@ -246,20 +246,19 @@ class MDX:
|
|
| 246 |
|
| 247 |
|
| 248 |
def run_mdx(model_params, output_dir, model_path, filename, exclude_main=False, exclude_inversion=False, suffix=None, invert_suffix=None, denoise=False, keep_orig=True, m_threads=2, base_device="cuda"):
|
| 249 |
-
|
|
|
|
| 250 |
if base_device == "cuda" and torch.cuda.is_available():
|
| 251 |
device = torch.device("cuda:0")
|
| 252 |
device_properties = torch.cuda.get_device_properties(device)
|
| 253 |
vram_gb = device_properties.total_memory / 1024**3
|
| 254 |
m_threads = 1 if vram_gb < 8 else (8 if vram_gb > 32 else 2)
|
| 255 |
-
print(f"threads: {m_threads} vram: {vram_gb}")
|
| 256 |
processor_num = 0
|
| 257 |
else:
|
| 258 |
device = torch.device("cpu")
|
| 259 |
m_threads = 2
|
| 260 |
if torch.cuda.is_available():
|
| 261 |
m_threads = 8
|
| 262 |
-
print(f"threads: {m_threads}")
|
| 263 |
processor_num = -1
|
| 264 |
|
| 265 |
model_hash = MDX.get_hash(model_path)
|
|
@@ -275,6 +274,11 @@ def run_mdx(model_params, output_dir, model_path, filename, exclude_main=False,
|
|
| 275 |
|
| 276 |
mdx_sess = MDX(model_path, model, processor=processor_num)
|
| 277 |
wave, sr = librosa.load(filename, mono=False, sr=44100)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 278 |
# normalizing input wave gives better output
|
| 279 |
peak = max(np.max(wave), abs(np.min(wave)))
|
| 280 |
wave /= peak
|
|
|
|
| 246 |
|
| 247 |
|
| 248 |
def run_mdx(model_params, output_dir, model_path, filename, exclude_main=False, exclude_inversion=False, suffix=None, invert_suffix=None, denoise=False, keep_orig=True, m_threads=2, base_device="cuda"):
|
| 249 |
+
vram_gb = 0
|
| 250 |
+
|
| 251 |
if base_device == "cuda" and torch.cuda.is_available():
|
| 252 |
device = torch.device("cuda:0")
|
| 253 |
device_properties = torch.cuda.get_device_properties(device)
|
| 254 |
vram_gb = device_properties.total_memory / 1024**3
|
| 255 |
m_threads = 1 if vram_gb < 8 else (8 if vram_gb > 32 else 2)
|
|
|
|
| 256 |
processor_num = 0
|
| 257 |
else:
|
| 258 |
device = torch.device("cpu")
|
| 259 |
m_threads = 2
|
| 260 |
if torch.cuda.is_available():
|
| 261 |
m_threads = 8
|
|
|
|
| 262 |
processor_num = -1
|
| 263 |
|
| 264 |
model_hash = MDX.get_hash(model_path)
|
|
|
|
| 274 |
|
| 275 |
mdx_sess = MDX(model_path, model, processor=processor_num)
|
| 276 |
wave, sr = librosa.load(filename, mono=False, sr=44100)
|
| 277 |
+
duration = librosa.get_duration(y=wave, sr=sr)
|
| 278 |
+
if duration < 60:
|
| 279 |
+
m_threads = 1
|
| 280 |
+
print(f"threads: {m_threads} vram: {vram_gb}")
|
| 281 |
+
|
| 282 |
# normalizing input wave gives better output
|
| 283 |
peak = max(np.max(wave), abs(np.min(wave)))
|
| 284 |
wave /= peak
|
src/rvc.py
CHANGED
|
@@ -157,9 +157,20 @@ def get_vc(device, is_half, config, model_path):
|
|
| 157 |
return cpt, version, net_g, tgt_sr, vc
|
| 158 |
|
| 159 |
|
| 160 |
-
def rvc_infer(index_path, index_rate, input_path, output_path, pitch_change, f0_method, cpt, version, net_g, filter_radius, tgt_sr, rms_mix_rate, protect, crepe_hop_length, vc, hubert_model):
|
| 161 |
-
audio = load_audio(input_path, 16000)
|
| 162 |
times = [0, 0, 0]
|
| 163 |
if_f0 = cpt.get('f0', 1)
|
| 164 |
-
|
| 165 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 157 |
return cpt, version, net_g, tgt_sr, vc
|
| 158 |
|
| 159 |
|
| 160 |
+
def rvc_infer(index_path, index_rate, input_path, output_path, pitch_change, f0_method, cpt, version, net_g, filter_radius, tgt_sr, rms_mix_rate, protect, crepe_hop_length, vc, hubert_model, steps):
|
|
|
|
| 161 |
times = [0, 0, 0]
|
| 162 |
if_f0 = cpt.get('f0', 1)
|
| 163 |
+
|
| 164 |
+
working_path = input_path
|
| 165 |
+
|
| 166 |
+
for step in range(steps):
|
| 167 |
+
audio = load_audio(working_path, 16000)
|
| 168 |
+
|
| 169 |
+
audio_opt = vc.pipeline(
|
| 170 |
+
hubert_model, net_g, step, audio, working_path, times, pitch_change,
|
| 171 |
+
f0_method, index_path, index_rate, if_f0, filter_radius, tgt_sr,
|
| 172 |
+
0, rms_mix_rate, version, protect, crepe_hop_length
|
| 173 |
+
)
|
| 174 |
+
|
| 175 |
+
wavfile.write(output_path, tgt_sr, audio_opt)
|
| 176 |
+
working_path = output_path
|
src/webui.py
CHANGED
|
@@ -6,10 +6,28 @@ import zipfile
|
|
| 6 |
from argparse import ArgumentParser
|
| 7 |
import spaces
|
| 8 |
import gradio as gr
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
|
| 10 |
-
from main import song_cover_pipeline
|
| 11 |
|
| 12 |
BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
|
|
|
|
| 13 |
|
| 14 |
mdxnet_models_dir = os.path.join(BASE_DIR, 'mdxnet_models')
|
| 15 |
rvc_models_dir = os.path.join(BASE_DIR, 'rvc_models')
|
|
@@ -78,10 +96,21 @@ def download_online_model(url, dir_name, progress=gr.Progress()):
|
|
| 78 |
if 'pixeldrain.com' in url:
|
| 79 |
url = f'https://pixeldrain.com/api/file/{zip_name}'
|
| 80 |
|
| 81 |
-
|
| 82 |
-
|
| 83 |
-
|
| 84 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 85 |
return f'[+] {dir_name} Model successfully downloaded!'
|
| 86 |
|
| 87 |
except Exception as e:
|
|
@@ -157,33 +186,37 @@ def show_hop_slider(pitch_detection_algo):
|
|
| 157 |
if __name__ == '__main__':
|
| 158 |
parser = ArgumentParser(description='Generate a AI cover song in the song_output/id directory.', add_help=True)
|
| 159 |
parser.add_argument("--share", action="store_true", dest="share_enabled", default=False, help="Enable sharing")
|
|
|
|
| 160 |
parser.add_argument("--listen", action="store_true", default=False, help="Make the WebUI reachable from your local network.")
|
| 161 |
parser.add_argument('--listen-host', type=str, help='The hostname that the server will use.')
|
| 162 |
parser.add_argument('--listen-port', type=int, help='The listening port that the server will use.')
|
|
|
|
|
|
|
| 163 |
args = parser.parse_args()
|
| 164 |
|
| 165 |
voice_models = get_current_models(rvc_models_dir)
|
| 166 |
with open(os.path.join(rvc_models_dir, 'public_models.json'), encoding='utf8') as infile:
|
| 167 |
public_models = json.load(infile)
|
| 168 |
|
| 169 |
-
with gr.Blocks(title='AICoverGenWebUI') as app:
|
| 170 |
-
|
| 171 |
-
gr.Label('AICoverGen WebUI ZeroGPU mode created with ❤️', show_label=False)
|
| 172 |
-
|
| 173 |
-
|
| 174 |
-
|
| 175 |
-
<
|
| 176 |
-
|
| 177 |
-
<
|
| 178 |
-
|
| 179 |
-
|
| 180 |
-
|
| 181 |
-
|
| 182 |
-
|
|
|
|
| 183 |
|
| 184 |
# main tab
|
| 185 |
with gr.Tab("Generate"):
|
| 186 |
-
|
| 187 |
with gr.Accordion('Main Options'):
|
| 188 |
with gr.Row():
|
| 189 |
with gr.Column():
|
|
@@ -191,12 +224,19 @@ if __name__ == '__main__':
|
|
| 191 |
ref_btn = gr.Button('Refresh Models 🔁', variant='primary')
|
| 192 |
|
| 193 |
with gr.Column(visible=False) as yt_link_col:
|
| 194 |
-
song_input = gr.Text(label='Song input', info='Link to a song on YouTube or full path to a local file. For file upload, click the button below.
|
| 195 |
show_file_upload_button = gr.Button('Upload file instead')
|
| 196 |
|
| 197 |
with gr.Column(visible=True) as file_upload_col:
|
| 198 |
audio_extensions = ['.mp3', '.m4a', '.flac', '.wav', '.aac', '.ogg', '.wma', '.alac', '.aiff', '.opus', 'amr']
|
| 199 |
-
local_file = gr.File(label='Audio file', interactive=True, type="filepath", file_types=audio_extensions)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 200 |
song_input_file = gr.UploadButton('Upload 📂', file_types=['audio'], variant='primary', visible=False)
|
| 201 |
show_yt_link_button = gr.Button('Paste YouTube link/Path to local file instead', visible=False)
|
| 202 |
song_input_file.upload(process_file_upload, inputs=[song_input_file], outputs=[local_file, song_input])
|
|
@@ -217,7 +257,12 @@ if __name__ == '__main__':
|
|
| 217 |
f0_method = gr.Dropdown(['rmvpe+', 'rmvpe', 'mangio-crepe'], value='rmvpe+', label='Pitch detection algorithm', info='Best option is rmvpe (clarity in vocals), then mangio-crepe (smoother vocals), rmvpe+ use a minimum and maximum allowed pitch values.')
|
| 218 |
crepe_hop_length = gr.Slider(32, 320, value=128, step=1, visible=False, label='Crepe hop length', info='Lower values leads to longer conversions and higher risk of voice cracks, but better pitch accuracy.')
|
| 219 |
f0_method.change(show_hop_slider, inputs=f0_method, outputs=crepe_hop_length)
|
| 220 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 221 |
|
| 222 |
with gr.Accordion('Audio mixing options', open=False):
|
| 223 |
gr.Markdown('### Volume Change (decibels)')
|
|
@@ -239,7 +284,12 @@ if __name__ == '__main__':
|
|
| 239 |
with gr.Row():
|
| 240 |
clear_btn = gr.ClearButton(value='Clear', components=[song_input, rvc_model, keep_files, local_file])
|
| 241 |
generate_btn = gr.Button("Generate", variant='primary')
|
| 242 |
-
ai_cover =
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 243 |
|
| 244 |
ref_btn.click(update_models_list, None, outputs=rvc_model)
|
| 245 |
is_webui = gr.Number(value=1, visible=False)
|
|
@@ -247,12 +297,12 @@ if __name__ == '__main__':
|
|
| 247 |
inputs=[local_file, rvc_model, pitch, keep_files, is_webui, main_gain, backup_gain,
|
| 248 |
inst_gain, index_rate, filter_radius, rms_mix_rate, f0_method, crepe_hop_length,
|
| 249 |
protect, pitch_all, reverb_rm_size, reverb_wet, reverb_dry, reverb_damping,
|
| 250 |
-
output_format],
|
| 251 |
outputs=[ai_cover])
|
| 252 |
-
clear_btn.click(lambda: [0, 0, 0, 0, 0.5, 3, 0.25, 0.33, 'rmvpe+', 128, 0, 0.15, 0.2, 0.8, 0.7, 'mp3', None],
|
| 253 |
outputs=[pitch, main_gain, backup_gain, inst_gain, index_rate, filter_radius, rms_mix_rate,
|
| 254 |
protect, f0_method, crepe_hop_length, pitch_all, reverb_rm_size, reverb_wet,
|
| 255 |
-
reverb_dry, reverb_damping, output_format, ai_cover])
|
| 256 |
|
| 257 |
# Download tab
|
| 258 |
with gr.Tab('Download model'):
|
|
@@ -271,6 +321,8 @@ if __name__ == '__main__':
|
|
| 271 |
gr.Markdown('## Input Examples')
|
| 272 |
gr.Examples(
|
| 273 |
[
|
|
|
|
|
|
|
| 274 |
['https://huggingface.co/phant0m4r/LiSA/resolve/main/LiSA.zip', 'Lisa'],
|
| 275 |
['https://pixeldrain.com/u/3tJmABXA', 'Gura'],
|
| 276 |
['https://huggingface.co/Kit-Lemonfoot/kitlemonfoot_rvc_models/resolve/main/AZKi%20(Hybrid).zip', 'Azki']
|
|
@@ -329,7 +381,10 @@ if __name__ == '__main__':
|
|
| 329 |
|
| 330 |
app.launch(
|
| 331 |
share=args.share_enabled,
|
|
|
|
|
|
|
| 332 |
# enable_queue=True,
|
| 333 |
server_name=None if not args.listen else (args.listen_host or '0.0.0.0'),
|
| 334 |
server_port=args.listen_port,
|
|
|
|
| 335 |
)
|
|
|
|
| 6 |
from argparse import ArgumentParser
|
| 7 |
import spaces
|
| 8 |
import gradio as gr
|
| 9 |
+
import logging
|
| 10 |
+
def configure_logging_libs(debug=False):
|
| 11 |
+
modules = [
|
| 12 |
+
"numba",
|
| 13 |
+
"httpx",
|
| 14 |
+
"markdown_it",
|
| 15 |
+
"fairseq",
|
| 16 |
+
"faiss",
|
| 17 |
+
]
|
| 18 |
+
try:
|
| 19 |
+
for module in modules:
|
| 20 |
+
logging.getLogger(module).setLevel(logging.WARNING)
|
| 21 |
+
os.environ['TF_CPP_MIN_LOG_LEVEL'] = "3" if not debug else "1"
|
| 22 |
+
|
| 23 |
+
except Exception as error:
|
| 24 |
+
pass
|
| 25 |
+
configure_logging_libs()
|
| 26 |
|
| 27 |
+
from main import song_cover_pipeline, yt_download
|
| 28 |
|
| 29 |
BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
|
| 30 |
+
IS_ZERO_GPU = os.getenv("SPACES_ZERO_GPU")
|
| 31 |
|
| 32 |
mdxnet_models_dir = os.path.join(BASE_DIR, 'mdxnet_models')
|
| 33 |
rvc_models_dir = os.path.join(BASE_DIR, 'rvc_models')
|
|
|
|
| 96 |
if 'pixeldrain.com' in url:
|
| 97 |
url = f'https://pixeldrain.com/api/file/{zip_name}'
|
| 98 |
|
| 99 |
+
|
| 100 |
+
if "," in url:
|
| 101 |
+
urls = [u.strip() for u in url.split(",") if u.strip()]
|
| 102 |
+
os.makedirs(extraction_folder, exist_ok=True)
|
| 103 |
+
for u in urls:
|
| 104 |
+
u = u.replace("?download=true", "")
|
| 105 |
+
file_name = u.split('/')[-1]
|
| 106 |
+
file_path = os.path.join(extraction_folder, file_name)
|
| 107 |
+
if not os.path.exists(file_path): # avoid re-downloading
|
| 108 |
+
urllib.request.urlretrieve(u, file_path)
|
| 109 |
+
else:
|
| 110 |
+
urllib.request.urlretrieve(url, zip_name)
|
| 111 |
+
|
| 112 |
+
progress(0.5, desc='[~] Extracting zip...')
|
| 113 |
+
extract_zip(extraction_folder, zip_name)
|
| 114 |
return f'[+] {dir_name} Model successfully downloaded!'
|
| 115 |
|
| 116 |
except Exception as e:
|
|
|
|
| 186 |
if __name__ == '__main__':
|
| 187 |
parser = ArgumentParser(description='Generate a AI cover song in the song_output/id directory.', add_help=True)
|
| 188 |
parser.add_argument("--share", action="store_true", dest="share_enabled", default=False, help="Enable sharing")
|
| 189 |
+
parser.add_argument("--builtin-player", action="store_true", default=False, help="Use the builtin audio player")
|
| 190 |
parser.add_argument("--listen", action="store_true", default=False, help="Make the WebUI reachable from your local network.")
|
| 191 |
parser.add_argument('--listen-host', type=str, help='The hostname that the server will use.')
|
| 192 |
parser.add_argument('--listen-port', type=int, help='The listening port that the server will use.')
|
| 193 |
+
parser.add_argument('--theme', type=str, default="NoCrypt/miku", help='Set the theme (default: NoCrypt/miku)')
|
| 194 |
+
parser.add_argument("--ssr", action="store_true", help="Enable SSR (Server-Side Rendering)")
|
| 195 |
args = parser.parse_args()
|
| 196 |
|
| 197 |
voice_models = get_current_models(rvc_models_dir)
|
| 198 |
with open(os.path.join(rvc_models_dir, 'public_models.json'), encoding='utf8') as infile:
|
| 199 |
public_models = json.load(infile)
|
| 200 |
|
| 201 |
+
with gr.Blocks(title='AICoverGenWebUI', theme=args.theme, fill_width=True, fill_height=False) as app:
|
| 202 |
+
|
| 203 |
+
gr.Label(f'AICoverGen WebUI {"ZeroGPU mode" if IS_ZERO_GPU else ""} created with ❤️', show_label=False)
|
| 204 |
+
if IS_ZERO_GPU:
|
| 205 |
+
gr.Markdown(
|
| 206 |
+
"""
|
| 207 |
+
<details>
|
| 208 |
+
<summary style="font-size: 1.5em;">⚠️ Important (click to expand)</summary>
|
| 209 |
+
<ul>
|
| 210 |
+
<li>🚀 This demo use a Zero GPU, which is available only for a limited time. It's recommended to use audio files that are no longer than 5 minutes. If you want to use it without time restrictions, you can duplicate the 'old CPU space'. ⏳</li>
|
| 211 |
+
</ul>
|
| 212 |
+
</details>
|
| 213 |
+
"""
|
| 214 |
+
)
|
| 215 |
+
gr.Markdown("Duplicate the old CPU space for use in private: [](https://huggingface.co/spaces/r3gm/AICoverGen_old_stable_cpu?duplicate=true)\n\n")
|
| 216 |
|
| 217 |
# main tab
|
| 218 |
with gr.Tab("Generate"):
|
| 219 |
+
|
| 220 |
with gr.Accordion('Main Options'):
|
| 221 |
with gr.Row():
|
| 222 |
with gr.Column():
|
|
|
|
| 224 |
ref_btn = gr.Button('Refresh Models 🔁', variant='primary')
|
| 225 |
|
| 226 |
with gr.Column(visible=False) as yt_link_col:
|
| 227 |
+
song_input = gr.Text(label='Song input', info='Link to a song on YouTube or full path to a local file. For file upload, click the button below.')
|
| 228 |
show_file_upload_button = gr.Button('Upload file instead')
|
| 229 |
|
| 230 |
with gr.Column(visible=True) as file_upload_col:
|
| 231 |
audio_extensions = ['.mp3', '.m4a', '.flac', '.wav', '.aac', '.ogg', '.wma', '.alac', '.aiff', '.opus', 'amr']
|
| 232 |
+
local_file = gr.File(label='Audio file', interactive=True, type="filepath", file_types=audio_extensions, height=150)
|
| 233 |
+
if not IS_ZERO_GPU:
|
| 234 |
+
with gr.Row():
|
| 235 |
+
with gr.Row(scale=2):
|
| 236 |
+
url_media_gui = gr.Textbox(value="", label="Enter URL", placeholder="www.youtube.com/watch?v=g_9rPvbENUw", lines=1)
|
| 237 |
+
with gr.Row(scale=1):
|
| 238 |
+
url_button_gui = gr.Button("Process URL", variant="secondary")
|
| 239 |
+
url_button_gui.click(yt_download, [url_media_gui], [local_file])
|
| 240 |
song_input_file = gr.UploadButton('Upload 📂', file_types=['audio'], variant='primary', visible=False)
|
| 241 |
show_yt_link_button = gr.Button('Paste YouTube link/Path to local file instead', visible=False)
|
| 242 |
song_input_file.upload(process_file_upload, inputs=[song_input_file], outputs=[local_file, song_input])
|
|
|
|
| 257 |
f0_method = gr.Dropdown(['rmvpe+', 'rmvpe', 'mangio-crepe'], value='rmvpe+', label='Pitch detection algorithm', info='Best option is rmvpe (clarity in vocals), then mangio-crepe (smoother vocals), rmvpe+ use a minimum and maximum allowed pitch values.')
|
| 258 |
crepe_hop_length = gr.Slider(32, 320, value=128, step=1, visible=False, label='Crepe hop length', info='Lower values leads to longer conversions and higher risk of voice cracks, but better pitch accuracy.')
|
| 259 |
f0_method.change(show_hop_slider, inputs=f0_method, outputs=crepe_hop_length)
|
| 260 |
+
with gr.Row():
|
| 261 |
+
with gr.Row():
|
| 262 |
+
steps = gr.Slider(minimum=1, maximum=3, label="Steps", value=1, step=1, interactive=True)
|
| 263 |
+
with gr.Row():
|
| 264 |
+
extra_denoise = gr.Checkbox(True, label='Denoise', info='Apply an additional noise reduction step to clean up the audio further.')
|
| 265 |
+
keep_files = gr.Checkbox((False if IS_ZERO_GPU else True), label='Keep intermediate files', info='Keep all audio files generated in the song_output/id directory, e.g. Isolated Vocals/Instrumentals. Leave unchecked to save space', interactive=(False if IS_ZERO_GPU else True))
|
| 266 |
|
| 267 |
with gr.Accordion('Audio mixing options', open=False):
|
| 268 |
gr.Markdown('### Volume Change (decibels)')
|
|
|
|
| 284 |
with gr.Row():
|
| 285 |
clear_btn = gr.ClearButton(value='Clear', components=[song_input, rvc_model, keep_files, local_file])
|
| 286 |
generate_btn = gr.Button("Generate", variant='primary')
|
| 287 |
+
ai_cover = (
|
| 288 |
+
gr.Audio(label='AI Cover', show_share_button=True)
|
| 289 |
+
if args.builtin_player else
|
| 290 |
+
gr.File(label="AI Cover", interactive=False)
|
| 291 |
+
)
|
| 292 |
+
gr.Markdown("- You can also try `AICoverGen❤️` in Colab’s free tier, which provides free GPU [link](https://github.com/R3gm/AICoverGen?tab=readme-ov-file#aicovergen).")
|
| 293 |
|
| 294 |
ref_btn.click(update_models_list, None, outputs=rvc_model)
|
| 295 |
is_webui = gr.Number(value=1, visible=False)
|
|
|
|
| 297 |
inputs=[local_file, rvc_model, pitch, keep_files, is_webui, main_gain, backup_gain,
|
| 298 |
inst_gain, index_rate, filter_radius, rms_mix_rate, f0_method, crepe_hop_length,
|
| 299 |
protect, pitch_all, reverb_rm_size, reverb_wet, reverb_dry, reverb_damping,
|
| 300 |
+
output_format, extra_denoise, steps],
|
| 301 |
outputs=[ai_cover])
|
| 302 |
+
clear_btn.click(lambda: [0, 0, 0, 0, 0.5, 3, 0.25, 0.33, 'rmvpe+', 128, 0, 0.15, 0.2, 0.8, 0.7, 'mp3', None, True, 1],
|
| 303 |
outputs=[pitch, main_gain, backup_gain, inst_gain, index_rate, filter_radius, rms_mix_rate,
|
| 304 |
protect, f0_method, crepe_hop_length, pitch_all, reverb_rm_size, reverb_wet,
|
| 305 |
+
reverb_dry, reverb_damping, output_format, ai_cover, extra_denoise, steps])
|
| 306 |
|
| 307 |
# Download tab
|
| 308 |
with gr.Tab('Download model'):
|
|
|
|
| 321 |
gr.Markdown('## Input Examples')
|
| 322 |
gr.Examples(
|
| 323 |
[
|
| 324 |
+
['https://huggingface.co/MrDawg/ToothBrushing/resolve/main/ToothBrushing.zip?download=true', 'ToothBrushing'],
|
| 325 |
+
['https://huggingface.co/sail-rvc/Aldeano_Minecraft__RVC_V2_-_500_Epochs_/resolve/main/model.pth?download=true, https://huggingface.co/sail-rvc/Aldeano_Minecraft__RVC_V2_-_500_Epochs_/resolve/main/model.index?download=true', 'Minecraft_Villager'],
|
| 326 |
['https://huggingface.co/phant0m4r/LiSA/resolve/main/LiSA.zip', 'Lisa'],
|
| 327 |
['https://pixeldrain.com/u/3tJmABXA', 'Gura'],
|
| 328 |
['https://huggingface.co/Kit-Lemonfoot/kitlemonfoot_rvc_models/resolve/main/AZKi%20(Hybrid).zip', 'Azki']
|
|
|
|
| 381 |
|
| 382 |
app.launch(
|
| 383 |
share=args.share_enabled,
|
| 384 |
+
debug=args.share_enabled,
|
| 385 |
+
show_error=True,
|
| 386 |
# enable_queue=True,
|
| 387 |
server_name=None if not args.listen else (args.listen_host or '0.0.0.0'),
|
| 388 |
server_port=args.listen_port,
|
| 389 |
+
ssr_mode=args.ssr
|
| 390 |
)
|