This is an interesting and very detailed examination of attempts to create music using artificial intelligence. It tracks what are (to my mind) two major stages in the evolution of this work: first, the shift from symbolic representations of music to actual samples of music; and second, the shift to convolutional neural networks: "Convolutional networks learn combinations of filters. They're normally used for processing images, but WaveNet treats time like a spatial dimension." It makes me think: that's why humans haave short-term memory (STM). Not as a staging area for long-term memory (LTM) but as a way of treating time as a spatial dimension. There's the obligatory question of whether these will replace humans, posed at the very end of the article (to no effect whatsoever) and a look at the use of these techniques to generate spoken word audio.
Today: 0 Total: 95 [Share]
] [