The ever-increasing size and accessibility of vast music libraries has created a demand more than ever for artificial systems that are capable of understanding, organizing, or even generating such complex data. While this topic has received relatively marginal attention within the machine learning community, it has been an area of intense focus within the community of Music Information Retrieval (MIR). While significant progress has been made, these problems remain far from solved.
Furthermore, the recommender systems community has made great advances in terms of collaborative feedback recommenders, but these approaches suffer strongly from the cold-start problem. As such, recommendation techniques often fall back on content-based machine learning systems, but defining musical similarity is extremely challenging as myriad features all play some role (e.g., cultural, emotional, timbral, rhythmic). Thus, for machines must actually understand music to achieve an expert level of music recommendation.
On the other side of this problem sits the recent explosion of work in the area of machine creativity. Relevant examples are both Google Magenta and the startup Jukedeck, who seek to develop algorithms capable of composing and performing completely original (and compelling) works of music. These algorithms require a similar deep understanding of music and present challenging new problems for the machine learning and AI community at large.
This workshop proposal is timely in that it will bridge these separate pockets of otherwise very related research. And in addition to making progress on the challenges above, we hope to engage the wide AI and machine learning community with our nebulous problem space, and connect them with the many available datasets the MIR community has to offer (e.g., Audio Set, AcousticBrainz, Million Song Dataset), which offer near commercial scale to the academic research community.
|09:00 AM||From Listening to Watching, A Recommender Systems Perspective (Keynote)|| Video »||Yves Raimond|
|10:00 AM||Poster Presentations (Part 1) (Poster Session)||||Ruchit Agrawal, Jeong Choi, Siddharth Gururani Gururani, Nima Hamidi, Harsh Jhamtani, Radha Manisha Kopparti, Ben Krause, Jongpil Lee, Ashis Pati, Fedor Zhdanov|
|11:00 AM||Making Efficient use of Musical Annotations (Invited Talk)|| Video »||Brian McFee|
|11:20 AM||Two-level Explanations in Music Emotion Recognition (Accepted Talk)|| Video »||Verena Haunschmid|
|11:40 AM||Characterizing Musical Correlates of Large-Scale Discovery Behavior (Invited Talk)|| Video »||Blair Kaneshiro|
|12:00 PM||NPR: Neural Personalised Ranking for Song Selection (Accepted Talk)|| Video »||Mark Levy|
|02:00 PM||Personalization at Amazon Music (Invited Talk)|| Video »||Kat Ellis|
|02:20 PM||A Model-Driven Exploration of Accent Within the Amateur Singing Voice (Accepted Talk)|| Video »||Camille Noufi|
|02:40 PM||What’s Broken in Music Informatics Research? Three Uncomfortable Statements (Invited Talk)||||Justin Salamon|
|03:00 PM||Poster Presentations (Part 2) (Poster Session)||||Ruchit Agrawal, Jeong Choi, Siddharth Gururani Gururani, Nima Hamidi, Harsh Jhamtani, Radha Manisha Kopparti, Ben Krause, Jongpil Lee, Ashis Pati, Fedor Zhdanov|
|04:30 PM||User-curated shaping of expressive performances (Invited Talk)|| Video »||Zhengshan Shi|
|04:50 PM||Interactive Neural Audio Synthesis (Accepted Talk)|| Video »||Hanoi Hantrakul|
|05:10 PM||Visualizing and Understanding Self-attention based Music Tagging (Accepted Talk)|| Video »||Minz Won|
|05:30 PM||A CycleGAN for style transfer between drum & bass subgenres (Accepted Talk)|| Video »||Len Vande Veire|