mirror of
https://github.com/babysor/Realtime-Voice-Clone-Chinese.git
synced 2026-02-09 05:24:19 +08:00
Compare commits
11 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
9c075e6b94 | ||
|
|
72b37a5a21 | ||
|
|
1cde29d5f3 | ||
|
|
4b8fa992b7 | ||
|
|
42789babd8 | ||
|
|
2354bb42d1 | ||
|
|
4358f6f353 | ||
|
|
5971555319 | ||
|
|
6f84026c51 | ||
|
|
a30657ecf5 | ||
|
|
cc250af1f6 |
@@ -29,6 +29,7 @@
|
|||||||
> 如果在用 pip 方式安装的时候出现 `ERROR: Could not find a version that satisfies the requirement torch==1.9.0+cu102 (from versions: 0.1.2, 0.1.2.post1, 0.1.2.post2)` 这个错误可能是 python 版本过低,3.9 可以安装成功
|
> 如果在用 pip 方式安装的时候出现 `ERROR: Could not find a version that satisfies the requirement torch==1.9.0+cu102 (from versions: 0.1.2, 0.1.2.post1, 0.1.2.post2)` 这个错误可能是 python 版本过低,3.9 可以安装成功
|
||||||
* 安装 [ffmpeg](https://ffmpeg.org/download.html#get-packages)。
|
* 安装 [ffmpeg](https://ffmpeg.org/download.html#get-packages)。
|
||||||
* 运行`pip install -r requirements.txt` 来安装剩余的必要包。
|
* 运行`pip install -r requirements.txt` 来安装剩余的必要包。
|
||||||
|
> 这里的环境建议使用 `Repo Tag 0.0.1` `Pytorch1.9.0 with Torchvision0.10.0 and cudatoolkit10.2` `requirements.txt` `webrtcvad-wheels` 因为 `requiremants.txt` 是在几个月前导出的,所以不适配新版本
|
||||||
* 安装 webrtcvad `pip install webrtcvad-wheels`。
|
* 安装 webrtcvad `pip install webrtcvad-wheels`。
|
||||||
|
|
||||||
或者
|
或者
|
||||||
|
|||||||
@@ -1,5 +1,7 @@
|
|||||||
|
> 🚧 While I no longer actively update this repo, you can find me continuously pushing this tech forward to good side and open-source. I'm also building an optimized and cloud hosted version: https://noiz.ai/ and [we're hiring](https://github.com/babysor/MockingBird/issues/1029).
|
||||||
|
>
|
||||||

|

|
||||||
|
<a href="https://trendshift.io/repositories/3869" target="_blank"><img src="https://trendshift.io/api/badge/repositories/3869" alt="babysor%2FMockingBird | Trendshift" style="width: 250px; height: 55px;" width="250" height="55"/></a>
|
||||||
|
|
||||||
[](http://choosealicense.com/licenses/mit/)
|
[](http://choosealicense.com/licenses/mit/)
|
||||||
|
|
||||||
@@ -29,6 +31,7 @@
|
|||||||
> If you get an `ERROR: Could not find a version that satisfies the requirement torch==1.9.0+cu102 (from versions: 0.1.2, 0.1.2.post1, 0.1.2.post2 )` This error is probably due to a low version of python, try using 3.9 and it will install successfully
|
> If you get an `ERROR: Could not find a version that satisfies the requirement torch==1.9.0+cu102 (from versions: 0.1.2, 0.1.2.post1, 0.1.2.post2 )` This error is probably due to a low version of python, try using 3.9 and it will install successfully
|
||||||
* Install [ffmpeg](https://ffmpeg.org/download.html#get-packages).
|
* Install [ffmpeg](https://ffmpeg.org/download.html#get-packages).
|
||||||
* Run `pip install -r requirements.txt` to install the remaining necessary packages.
|
* Run `pip install -r requirements.txt` to install the remaining necessary packages.
|
||||||
|
> The recommended environment here is `Repo Tag 0.0.1` `Pytorch1.9.0 with Torchvision0.10.0 and cudatoolkit10.2` `requirements.txt` `webrtcvad-wheels` because `requirements. txt` was exported a few months ago, so it doesn't work with newer versions
|
||||||
* Install webrtcvad `pip install webrtcvad-wheels`(If you need)
|
* Install webrtcvad `pip install webrtcvad-wheels`(If you need)
|
||||||
|
|
||||||
or
|
or
|
||||||
|
|||||||
@@ -116,14 +116,13 @@ def preprocess_general(speaker_dir, out_dir: Path, skip_existing: bool, hparams,
|
|||||||
print(f"No word found in dict_info for {wav_fpath.name}, skip it")
|
print(f"No word found in dict_info for {wav_fpath.name}, skip it")
|
||||||
continue
|
continue
|
||||||
sub_basename = "%s_%02d" % (wav_fpath.name, 0)
|
sub_basename = "%s_%02d" % (wav_fpath.name, 0)
|
||||||
mel_fpath = out_dir.joinpath("mels", f"mel-{sub_basename}.npy")
|
mel_fpath_out = out_dir.joinpath("mels", f"mel-{sub_basename}.npy")
|
||||||
wav_fpath = out_dir.joinpath("audio", f"audio-{sub_basename}.npy")
|
wav_fpath_out = out_dir.joinpath("audio", f"audio-{sub_basename}.npy")
|
||||||
|
|
||||||
if skip_existing and mel_fpath.exists() and wav_fpath.exists():
|
if skip_existing and mel_fpath_out.exists() and wav_fpath_out.exists():
|
||||||
continue
|
continue
|
||||||
wav, text = _split_on_silences(wav_fpath, words, hparams)
|
wav, text = _split_on_silences(wav_fpath, words, hparams)
|
||||||
result = _process_utterance(wav, text, out_dir, sub_basename,
|
result = _process_utterance(wav, text, out_dir, sub_basename, mel_fpath_out, wav_fpath_out, hparams, encoder_model_fpath)
|
||||||
False, hparams, encoder_model_fpath) # accelarate
|
|
||||||
if result is None:
|
if result is None:
|
||||||
continue
|
continue
|
||||||
wav_fpath_name, mel_fpath_name, embed_fpath_name, wav, mel_frames, text = result
|
wav_fpath_name, mel_fpath_name, embed_fpath_name, wav, mel_frames, text = result
|
||||||
|
|||||||
@@ -2,7 +2,8 @@ umap-learn
|
|||||||
visdom
|
visdom
|
||||||
librosa
|
librosa
|
||||||
matplotlib>=3.3.0
|
matplotlib>=3.3.0
|
||||||
numpy
|
numpy==1.19.3; platform_system == "Windows"
|
||||||
|
numpy==1.20.3; platform_system != "Windows"
|
||||||
scipy>=1.0.0
|
scipy>=1.0.0
|
||||||
tqdm
|
tqdm
|
||||||
sounddevice
|
sounddevice
|
||||||
@@ -12,8 +13,8 @@ inflect
|
|||||||
PyQt5
|
PyQt5
|
||||||
multiprocess
|
multiprocess
|
||||||
numba
|
numba
|
||||||
webrtcvad
|
webrtcvad; platform_system != "Windows"
|
||||||
pypinyin
|
pypinyin==0.44.0
|
||||||
flask
|
flask
|
||||||
flask_wtf
|
flask_wtf
|
||||||
flask_cors
|
flask_cors
|
||||||
@@ -25,9 +26,9 @@ PyYAML
|
|||||||
torch_complex
|
torch_complex
|
||||||
espnet
|
espnet
|
||||||
PyWavelets
|
PyWavelets
|
||||||
monotonic-align==0.0.3
|
|
||||||
transformers
|
|
||||||
fastapi
|
fastapi
|
||||||
loguru
|
loguru
|
||||||
typer[all]
|
click==8.0.4
|
||||||
click
|
typer
|
||||||
|
monotonic-align==1.0.0
|
||||||
|
transformers
|
||||||
|
|||||||
Reference in New Issue
Block a user