diff --git a/.github/actions/audiocraft_build/action.yml b/.github/actions/audiocraft_build/action.yml index 1d32cddc..9a797723 100644 --- a/.github/actions/audiocraft_build/action.yml +++ b/.github/actions/audiocraft_build/action.yml @@ -21,7 +21,7 @@ runs: python3 -m venv env . env/bin/activate python -m pip install --upgrade pip - pip install 'numpy==1.17.3' 'scipy==1.5.4' torch==2.1.0 torchvision==0.16.0 torchaudio==2.1.0 + pip install numpy==1.23.5 torch==2.1.0 torchvision==0.16.0 torchaudio==2.1.0 pip install xformers==0.0.22.post7 pip install -e '.[dev,wm]' - name: System Dependencies diff --git a/docs/MUSICGEN_STYLE.md b/docs/MUSICGEN_STYLE.md index 9ed9b0f4..6814dcbe 100644 --- a/docs/MUSICGEN_STYLE.md +++ b/docs/MUSICGEN_STYLE.md @@ -26,6 +26,8 @@ MusicGen-Stem requires a GPU with at least 16 GB of memory for running inference ## Usage 1. You can play with MusicGen-Style by running the jupyter notebook at [`demos/musicgen_style_demo.ipynb`](../demos/musicgen_style_demo.ipynb) locally (if you have a GPU). +2. You can use the gradio demo locally by running python -m demos.musicgen_style_app --share. +3. You can play with MusicGen by running the jupyter notebook at demos/musicgen_style_demo.ipynb locally (if you have a GPU). ## API diff --git a/requirements.txt b/requirements.txt index 120f9085..1ed24305 100644 --- a/requirements.txt +++ b/requirements.txt @@ -16,7 +16,7 @@ tqdm transformers>=4.31.0 # need Encodec there. xformers<0.0.23 demucs -librosa +librosa==0.10.1 soundfile gradio torchmetrics