|
--- |
|
license: mit |
|
--- |
|
|
|
# Amphion Vocoder Pretrained Models |
|
|
|
We provide a [DiffWave](https://github.com/open-mmlab/Amphion/tree/main/egs/vocoder/diffusion) pretrained checkpoint, which is trained on 125 hours of speech data and 80 hours of singing voice data. |
|
|
|
## Quick Start |
|
|
|
To utilize these pretrained vocoders, just run the following commands: |
|
|
|
### Step1: Download the checkpoint |
|
```bash |
|
git lfs install |
|
git clone https://huggingface.co/amphion/diffwave |
|
``` |
|
|
|
### Step2: Clone the Amphion's Source Code of GitHub |
|
```bash |
|
git clone https://github.com/open-mmlab/Amphion.git |
|
``` |
|
|
|
### Step3: Specify the checkpoint's path |
|
Use the soft link to specify the downloaded checkpoint in the first step: |
|
|
|
```bash |
|
cd Amphion |
|
mkdir -p ckpts/vocoder |
|
ln -s "$(realpath ../diffwave/diffwave)" pretrained/diffwave |
|
``` |
|
|
|
### Step4: Inference |
|
For analysis synthesis on the processed dataset, raw waveform, or predicted mel spectrograms, you can follow the inference part of [this recipe](https://github.com/open-mmlab/Amphion/tree/main/egs/vocoder/diffusion). |
|
|
|
```bash |
|
sh egs/vocoder/diffusion/diffwave/run.sh --stage 3 \ |
|
--infer_mode [Your chosen inference mode] \ |
|
--infer_datasets [Datasets you want to inference, needed when infer_from_dataset] \ |
|
--infer_feature_dir [Your path to your predicted acoustic features, needed when infer_from_feature] \ |
|
--infer_audio_dir [Your path to your audio files, needed when infer_form_audio] \ |
|
--infer_expt_dir Amphion/ckpts/vocoder/[YourExptName] \ |
|
--infer_output_dir Amphion/ckpts/vocoder/[YourExptName]/result \ |
|
``` |