-
Notifications
You must be signed in to change notification settings - Fork 5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
EXAMPLE Lora training and EXAMPLE lora #7
Comments
Hi there,
I'm still experimenting a bunch with this myself, but I'll prioritize adding some examples. As for the dataset preparation, I would defer to the examples in stable-audio-tools for the time being. I'll add some links to those.
train.py has all the same functionality as train.py in stable-audio-tools. Adding the --use-lora argument will instruct it to freeze the base model and train lora weights only.
I'm still figuring out the best hyperparameters to use. With Stable Audio Open, a rank 16 lora is ~59 MB and a rank 128 lora is ~472 MB. With a batch size of 32 and sample length of 10s, I am getting good results in several-thousand steps range (a few hours on my setup)
I am fine-tuning Stable Audio Open with a batch size of 32 and a sample length of 10 seconds. On a 4070 ti super in Windows 11, it is taking a bit less that 10GB for rank 16 and 13GB for rank 128. It may be more efficient on linux -- I was getting less than 8GB while experimenting with rank 16 on Colab. |
Any new about examples? |
The sticking point in the README is this line: https://github.com/NeuralNotW0rk/LoRAW/blob/main/README.md?plain=1#L21 I have tried the configs here with the Stable Audio Open 1.0 checkpoint as the starting point with the command:
But it fails to run entirely in this repo. In the main repo the finetuned model returns noise when using |
I think I can share some audio examples in my next commit. In your original question are you asking for an actual dataset and a lora checkpoint? |
The config in https://github.com/NeuralNotW0rk/LoRAW/blob/main/examples/model_config.json is what I use with Stable Audio Open. You should be able to add a "lora" section to the end of any model config. |
Ah yes, it wasn't originally clear which config was used to me to train stable audio open but I actually found the model_config.json file on hugging face independently just now! |
What is the maximum length recommended for each wav in the data? |
Where I can find trained loras for testing? thax all! |
HI!
The text was updated successfully, but these errors were encountered: