AnyAccomp: Generalizable Accompaniment Generation via Quantized Melodic Bottleneck

This is the official Hugging Face model repository for AnyAccomp, an accompaniment generation framework from the paper AnyAccomp: Generalizable Accompaniment Generation via Quantized Melodic Bottleneck.

AnyAccomp addresses two critical challenges in accompaniment generation: generalization to in-the-wild singing voices and versatility in handling solo instrumental inputs.

The core of our framework is a quantized melodic bottleneck, which extracts robust melodic features. A subsequent flow matching model then generates a matching accompaniment based on these features.

For more details, please visit our GitHub Repository.

framework

Model Checkpoints

This repository contains the three pretrained components of the AnyAccomp framework:

Model Name Directory Description
VQ ./pretrained/vq Extracts core melodic features from audio.
Flow Matching ./pretrained/flow_matching Generates accompaniments from melodic features.
Vocoder ./pretrained/vocoder Converts generated features into audio waveforms.

How to use

To run this model, you need to follow the steps below:

  1. Clone the repository and install the environment.
  2. Run the Gradio demo / Inference script.

1. Clone and Environment

In this section, follow the steps below to clone the repository and install the environment.

  1. Clone the repository.
  2. Install the environment following the guide below.
git clone https://github.com/AmphionTeam/AnyAccomp.git

# enter the repositry directory
cd AnyAccomp

2. Download the Pretrained Models

We provide a simple Python script to download all the necessary pretrained models from Hugging Face into the correct directory.

Before running the script, make sure you are in the AnyAccomp root directory.

Run the following command:

python -c "from huggingface_hub import snapshot_download; snapshot_download(repo_id='amphion/anyaccomp', local_dir='./pretrained', repo_type='model')"

If you have trouble connecting to Hugging Face, you can try switching to a mirror endpoint before running the command:

export HF_ENDPOINT=https://hf-mirror.com

3. Install the Environment

Before start installing, make sure you are under the AnyAccomp directory. If not, use cd to enter.

conda create -n anyaccomp python=3.9
conda activate anyaccomp
conda install -c conda-forge ffmpeg=4.0
pip install -r requirements.txt 

Run the Model

Once the setup is complete, you can run the model using either the Gradio demo or the inference script.

Run Gradio πŸ€— Playground Locally

You can run the following command to interact with the playground:

python gradio_app.py

Inference Script

If you want to infer several audios, you can use the python inference script from folder.

python infer_from_folder.py

By default, the script loads input audio from ./example/input and saves the results to ./example/output. You can customize these paths in the inference script.

Citation

If you use AnyAccomp in your research, please cite our paper:

@article{zhang2025anyaccomp,
  title={AnyAccomp: Generalizable Accompaniment Generation via Quantized Melodic Bottleneck},
  author={Zhang, Junan and Zhang, Yunjia and Zhang, Xueyao and Wu, Zhizheng},
  journal={arXiv preprint arXiv:2509.14052},
  year={2025}
}
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ 2 Ask for provider support