Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Brainstorm ISMIR tasks #3

Open
ruohoruotsi opened this issue Jul 17, 2019 · 6 comments
Open

Brainstorm ISMIR tasks #3

ruohoruotsi opened this issue Jul 17, 2019 · 6 comments

Comments

@ruohoruotsi
Copy link
Member

Brainstorm ISMIR 2020 tasks that would be enlightened, liberated, illuminated with a proper MIDI dataset of Caribbean riddims (some with melodies and skank chord progressions)

  • Release MIDI Dataset
  • Use it on one or more tasks. See MIREX for more info
    • Preliminary ideas include: generative symbolic music systems, VAE latent space explorers, bassline-melody embeddings, other symbolic music information retrieval/generation tasks?
@ruohoruotsi
Copy link
Member Author

@Cortexelus
Copy link

Generation -- could try taking a pre-trained Transformer (trained on a large multi-genre midi collection) then fine-tune it on this dataset [like the LakhNES training method]

@ruohoruotsi
Copy link
Member Author

Hey CJ, that's great idea ... so to summarize (1) preprocess data into trainable format (2) finetune on something like the Transformer-XL checkpoints (3) generate more basslines. What did I miss?

@Cortexelus
Copy link

That sounds about right. Convert your dataset into the format the pretrained model used. Finetune train (swap in this new dataset and continue training). Generate tons of output examples as you train.

Not sure if you've done other generative Transformer finetuning experiments, but I've done it on text with GPT2, as far as output quality, there's usually a sweet range. Too little iterations and it doesn't fit the aesthetic of your dataset yet. Too many iterations and it plagiarizes too much. The sweet range may last several thousand iterations. Every iteration you sample from has a different personality. If you only generate from your best/last iteration, the output won't be as diverse as if you were to generate from many different iterations in the sweet range.

@ruohoruotsi
Copy link
Member Author

Thanks CJ, I appreciate the tips! Once I'm going on the dataset, and have some initial generated basslines. I'll share it w/ you, perhaps there's a vector of collaboration here.

I haven't fine-tuned generative Transformers. But I have built/iterated/refined lots of other audio/dsp/speech models mostly on audio-signals though.

Do you have any recommendations on MIDI python libs/frameworks that'd be useful in whipping the MIDI data into shape? I fear I may have to manually pull the "riddim"/bassline out of the songs that are multitrack, others are just riddims + drum. Also debating whether to keep the "skank" with the bassline 🙇 🤔

@Cortexelus
Copy link

Cortexelus commented Jan 27, 2020 via email

@ruohoruotsi ruohoruotsi changed the title Brainstorm ISMIR 2020 tasks Brainstorm ISMIR tasks Nov 10, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants