Wednesday, October 21, 2009

Modelling Music as a Dynamic Texture

My paper "Modeling Music as a Dynamic Texture" has been accepted for publication in IEEE Transactions on Audio, Speech and Language Processing! This is work with my colleague and good buddy, Prof. Antoni Chan, and my advisor and good buddy, Prof. Gert Lanckriet and is awesome for at least 5 reasons:
  1. It presents a new, statistical model of musical audio that really takes time into account
  2. It introduces - and solves - the "Bohemian Rhapsody problem" (in case you don't know what this is - try describing the sound of BoRhap in 5 words or less...)
  3. Our model can automatically separate the verse, chorus, solo, bridge, etc parts of any song as well or better than anything else
  4. It will be a chapter in my thesis
  5. I get to talk about Freddie Mercury's falsetto, Bjork and the Beatles in an academic paper
Lookout for the paper in a library near you before the end of the year! We're working on extending the model to more Music Information Retrieval tasks like tagging and similarity. In the meantime, check out some of our results on automatically segmenting music.

Listen to the 35-second version of Bohemian Rhapsody!

No comments: