Skip to content

Latest commit

 

History

History
135 lines (76 loc) · 7.41 KB

README_EN.md

File metadata and controls

135 lines (76 loc) · 7.41 KB

简体中文 / Discord / Buy me a coffee / Twitter

CV Voice Clone Tool

All the models used in this project are from https://github.com/coqui-ai/TTS, the model protocol is CPML and can only be used for study and research, not for commercial use.

This is a voice cloning tool that can use any human voice to synthesize a piece of text into a voice using that voice, or to convert one voice into another using that voice.

It's very easy to use, even without an N-series GPU. Download the precompiled version and double click on app.exe to open a web interface, and it can be used with a few mouse clicks.

Supports Chinese English Japanese Korean eg. total 16 languages, and can record voices online through a microphone.

To ensure the synthesized effect, it's recommended to record for 5 to 20 seconds, pronounce clearly and accurately, and don't have background noise.

Video Demonstration

clone-en.mp4

image

How to use the precompiled version under win (other systems can deploy source code)

  1. Download the 'precompiled version of the main file(1.7G) and Model(3G) separately from Releases on the right.

  2. After downloading, unzip it to somewhere, for example E:/clone-voice.

  3. Double click app.exe, wait for the web window to open automatically, Please read the text prompts in the CMD window carefully, if there are errors, they will be displayed here. image

  4. After the model download, unzip it to the tts folder under the software directory, the effect after unzipping is as shown in the picture

image

  1. Conversion operation steps:

    • Enter the text in the text box, or import the SRT file, or select "Voice-> Voice", choose the voice wav format file you want to convert.

    • Then select the voice you want to use from the drop-down box under "Voice wav file to use", if you are not satisfied, you can also click the "Upload locally" button, select a recorded 5-20s wav voice file. Or click the "Start recording" button to record your own voice for 5-20 seconds online, after recording, click to use.

    • Click the "Start Generating Now" button and wait patiently for completion.

  2. If the machine has an N card GPU and CUDA environment is correctly configured, CUDA acceleration will be used automatically.

Source Code Deployment (linux mac window) / Example: window

If your area can't access google and huggingface, you'll need a global proxy because models need to be downloaded from github and huggingface

  1. Required python 3.9-> 3.11, and enable a global proxy, ensure the proxy is stable

  2. Create an empty directory, such as E:/clone-voice, open a cmd window in this directory, the method is to type cmd in the address bar, then press Enter. and exec git pull source code git clone [email protected]:jianchang512/clone-voice.git .

  3. Create a virtual environment python -m venv venv

  4. Activate the environment E:/clone-voice/venv/scripts/activate

  5. Install dependencies: pip install -r requirements.txt

  6. Unzip the ffmpeg.7z to the project root directory

  7. First run python code_dev.py, enter y when prompted to accept the agreement, then wait for the model to be downloaded completely.

  8. After downloading, restart python app.py.

  9. Every startup will connect to the foreign Internet to check or update the model, please be patient and wait. If you don't want to check or update every time you start, you need to manually modify the files under the dependent package, open \venv\Lib\site-packages\TTS\utils\manage.py, around line 389, def download_model method, comment out the following code.

if md5sum is not None:
	md5sum_file = os.path.join(output_path, "hash.md5")
	if os.path.isfile(md5sum_file):
	    with open(md5sum_file, mode="r") as f:
		if not f.read() == md5sum:
		    print(f" > {model_name} has been updated, clearing model cache...")
		    self.create_dir_and_download_model(model_name, model_item, output_path)
		else:
		    print(f" > {model_name} is already downloaded.")
	else:
	    print(f" > {model_name} has been updated, clearing model cache...")
	    self.create_dir_and_download_model(model_name, model_item, output_path)
  1. The startup of the source code version may frequently encounter errors, which are basically due to proxy problems that prevent the download of models from the walls or the download is interrupted and not complete. It is recommended to use a stable proxy and open it globally. If you can't download completely all the time, it's recommended to use the precompiled version.

CUDA Acceleration Support

Installation of CUDA tools

If your computer has Nvidia graphics card, upgrade the graphics card driver to the latest, then go to install the corresponding CUDA Toolkit 11.8 and cudnn for CUDA11.X.

When installation is complete, press Win + R, type cmd then press Enter, in the pop-up window type nvcc --version, confirm the version information display, similar to this image image

Then continue to type nvidia-smi, confirm there's output information, and you can see the cuda version number, similar to this image image

That means the installation is correct, you can cuda accelerate now, otherwise you need to reinstall.

Precautions

The model xtts can only be used for study and research, not for commerical use

  1. The source code version requires global proxy, because it needs to download models from https://huggingface.co, and this website can't be accessed in China, the source code version may frequently encounter errors when starting, basically proxy problems lead to unable to download models from overseas or download interruption incomplete. It's recommended to use a stable proxy, open it globally. If you can't download completely all the time, it's recommended to use the precompiled version.

  2. It will consume some time to load the model coldly after starting, please wait patiently for http://127.0.0.1:9988 to be displayed, and automatically open the browser page, wait for two or three minutes before converting.

  3. Functions include:

     Text to voice: that is, enter the text, generate voice with the selected voice.
     
     Voice to Voice: that is, select an audio file from the local area, generate another audio file with the selected voice.
    
  4. If the cmd window opened for a long time doesn't move, you need to press Enter on it to continue output, please click on the icon in the upper left corner of cmd, select "Properties", then uncheck the "Quick Edit" and "Insert Mode" checkboxes