Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up

FastSAG: Towards Fast Non-Autoregressive Singing Accompaniment Generation

NotificationsYou must be signed in to change notification settings

chenjianyi/fastsag

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

13 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

This is a PyTorch/GPU implementation of the IJCAI 2024 paperFastSAG: Towards Fast Non-Autoregressive Singing Accompaniment Generation.Demo page can be found atdemo.

@article{chen2024fastsag,  title={FastSAG: Towards Fast Non-Autoregressive Singing Accompaniment Generation},  author={Chen, Jianyi and Xue, Wei and Tan, Xu and Ye, Zhen and Liu, Qifeng and Guo, Yike},  journal={arXiv preprint arXiv:2405.07682},  year={2024}}

Preparation

  1. Download this code:
git clone https://github.com/chenjianyi/fastsag/cd fastsag
  1. Download fastsag checkpoint fromhereand put all weights in fastsag/weights

    BigvGAN checkpoints could be downloaded fromBigvGAN. The checkpoints we used is "bigvgan_24khz_100band". I upgrade BigvGAN to BigvGAN-v2, and the checkpoints would be downloaded automatically.

    MERT pretrained checkpoints would be downloaded automatically from huggingface. Please make sure your sever could access huggingface.

Dataset

  1. Source seperation:
cd preprocessingpython3 demucs_processing.py  # you may need to change root_dir and out_dir in this file
  1. cliping to 10s and filtering salient clips
python3 clip_to_10s.py  # change src_root and des_root for your dataset

Training

cd ../sde_diffusionpython3 train.py --data_dir YOUR_TRAIN_DATA --data_dir_testset YOUR_TEST_DATA --results_folder RESULTS

Generation

python3 generate.py --ckpt TRAINED_MODEL --data_dir DATA_DIR --result_dir OUTPUT

Acknowledgement and reference

  1. Grad-TTS.
  2. CoMoSpeech

[8]ページ先頭

©2009-2025 Movatter.jp