-
Notifications
You must be signed in to change notification settings - Fork 5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Brainstorm ISMIR tasks #3
Comments
Generation -- could try taking a pre-trained Transformer (trained on a large multi-genre midi collection) then fine-tune it on this dataset [like the LakhNES training method] |
Hey CJ, that's great idea ... so to summarize (1) preprocess data into trainable format (2) finetune on something like the Transformer-XL checkpoints (3) generate more basslines. What did I miss? |
That sounds about right. Convert your dataset into the format the pretrained model used. Finetune train (swap in this new dataset and continue training). Generate tons of output examples as you train. Not sure if you've done other generative Transformer finetuning experiments, but I've done it on text with GPT2, as far as output quality, there's usually a sweet range. Too little iterations and it doesn't fit the aesthetic of your dataset yet. Too many iterations and it plagiarizes too much. The sweet range may last several thousand iterations. Every iteration you sample from has a different personality. If you only generate from your best/last iteration, the output won't be as diverse as if you were to generate from many different iterations in the sweet range. |
Thanks CJ, I appreciate the tips! Once I'm going on the dataset, and have some initial generated basslines. I'll share it w/ you, perhaps there's a vector of collaboration here. I haven't fine-tuned generative Transformers. But I have built/iterated/refined lots of other audio/dsp/speech models mostly on audio-signals though. Do you have any recommendations on MIDI python libs/frameworks that'd be useful in whipping the MIDI data into shape? I fear I may have to manually pull the "riddim"/bassline out of the songs that are multitrack, others are just riddims + drum. Also debating whether to keep the "skank" with the bassline 🙇 🤔 |
cool! :D keep me in the loop
hmm try pretty-midi http://craffel.github.io/pretty-midi/
…On Sat, Jan 25, 2020 at 1:46 AM Ruoho Ruotsi ***@***.***> wrote:
Thanks CJ, I appreciate the tips! Once I'm going on the dataset, and have
some initial generated basslines. I'll share it w/ you, perhaps there's a
vector of collaboration here.
I haven't fine-tuned generative Transformers. But I have
built/iterated/refined lots of other audio/dsp/speech models mostly on
audio-signals though.
Do you have any recommendations on MIDI python libs/frameworks that'd be
useful in whipping the MIDI data into shape? I fear I may have to manually
pull the "riddim"/bassline out of the songs that are multitrack, others are
just riddims + drum. Also debating whether to keep the "skank" with the
bassline 🙇 🤔
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
<#3?email_source=notifications&email_token=AASXCZN6L6L2KSZ2EKVWZV3Q7PN2TA5CNFSM4IEVUD62YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOEJ4WW7A#issuecomment-578382716>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AASXCZO4LLQJEGKNVC4VXK3Q7PN2TANCNFSM4IEVUD6Q>
.
|
Brainstorm ISMIR 2020 tasks that would be enlightened, liberated, illuminated with a proper MIDI dataset of Caribbean riddims (some with melodies and skank chord progressions)
The text was updated successfully, but these errors were encountered: