-
Notifications
You must be signed in to change notification settings - Fork 596
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Help]: Gradio playground for MaskGCT load fails with less than 16GB VRAM #328
Comments
Hello, I had the same problem. |
My experience of trying this model on 3090Ti: Merely loading the model takes up about 10G VRAM, then when doing normal inference(prompt audio of reasonable length like 10+seconds), it takes up about 12G VRAM, so I am afraid your hardware is not adequate for this model without further optimization. |
Try Nvidia Control Panel > Manage 3D Settings > CUDA - Sysmem Fallback Policy > Set Prefer Sysmem Fallback and when you exceed your VRAM it will start using your system ram. |
I uploaded a modified version of the original huggingface MaskGCT gradio app that uses 10GB and 370MB of VRAM. The file is called app-old-target-duration-correction.py |
Problem Overview
Attempting to test the MaskGCT Gradio playground locally
https://github.com/open-mmlab/Amphion/tree/main/models/tts/maskgct
Steps Taken
Ran the steps in the README for the full install
when executing
python -m models.tts.maskgct.gradio_demo
, fails after loadingfacebook/w2v-bert-2.0
with error:torch.cuda.OutOfMemoryError: tried to allocate... 7.75 total capacity ... 6.64 already allocated.
Expected Outcome
(A clear and concise description of what you expected to happen.)
Is there another set of models that can make this possible to test with 8GB of VRAM?
Environment Information
The text was updated successfully, but these errors were encountered: