0.1 - Literature

My first thought with the whole CNN structure is that it doesn't necessarily represent how the ears take in information, but they are very good at mimicking a visual retina. Therefore after looking into different directions as to what optimisations might be possible for the CNN's to further help them to make sense of music, found a recent paper about MIDInet. Which states that to guide the process of music generation by a GAN coupled CNN has to have a time constraint in which it generates the following note based on the note(s) that came before as to make sure there is more structure.

I am still looking into the idea of how to replicate the inner workings of the human ear to make sure that better suits the whole interpretative part of the music generation process.

Comments

Popular posts from this blog

4.0 - More data, more data!

4.1 - Redefining overall structure

2.0 - Literature draft