Make some music with an RNN
Let’s make some music with an LSTM-RNN!
For this exercise, you can start with the “music RNN” example here on your computer.
This example works on Colab as well, but you won’t be able to view or hear the generated music as easily so it’s not quite as much fun.
Notebook: Zeldic RNN (Github) (open in Colab)
The melody RNN needs a few extra pieces of software to work correctly.
You’ll need to install the Python library music21
to read and
write MIDI files, and musescore
to play them back in Jupyter
Notebooks.
- The melody RNN code is designed to train a small RNN on a corpus of MIDI files. Just as the charRNN generates text letter by letter, the melody RNN generates music note-by-note. There’s a small corpus of MIDI data (already processed) provided with the example, try running the code to see if you can train an RNN with this data.
- Find some of your own MIDI files to load into the melody-RNN. You’ll have to run the dataset generator code to get them into a format that the RNN can train from.
- The musical representation for this RNN is very simple: Integers such that 0-127 means a note-on at that pitch, 128 means do nothing, and 129 means note-off. Can you think of a different way of representing the notes that would work with this RNN?
- Think about how a melody generator RNN could be integrated into an instrument or artwork. How would it challenge or extend the role of the performer, and of the listener?
Notes on the MIDI representation:
This script uses a very simple method to represent MIDI melodies as a sequence of integers.
- 0-127 play a note at that MIDI note number. (
MELODY_NOTE_ON
) - 128 stop whatever note was playing. (
MELODY_NOTE_OFF
) - 129 do nothing. (
MELODY_NO_EVENT
)
This encoding is inspired (and borrowed) from Magenta’s Melody_RNN model.
Note that this is an early and (now) superseded practice for
representing MIDI data to a deep neural network. The most modern
models
like
Music Transformer
and PerformanceRNN tend to use more complicated representations. One
downside of this representation is that 128
and 129
are extremely
over-represented in the data which can cause issues with learning very
large amounts of data or more complicated models.
However, this one is simple and easy to calculate. I have a more advanced script for converting a lot of MIDI files if you’re interested.