Visualizing Ensemble Predictions of Music Mood

Zelin Ye, Min Chen

View presentation: 2022-10-20T16:45:00Z GMT-0600 Change your timezone on the schedule page
2022-10-20T16:45:00Z
Exemplar figure, described by caption below
An ensemble of 210 machine learning (ML) models are visualized by the dual-flux ThemeRiver, accompanied with the music score for bar 30~34 of Bach: Paritita V in G, BWV829: Gigue. The upper, lower flux and the dashed line depict the dominant mood, the other moods in the descending order, and the 50% threshold over time.

Prerecorded Talk

The live footage of the talk, including the Q&A, can be viewed on the session page, Interpreting Machine Learning.

Fast forward
Abstract

Music mood classification has been a challenging problem in comparison with other music classification problems (e.g., genre, composer, or period). One solution for addressing this challenge is to use an ensemble of machine learning models. In this paper, we show that visualization techniques can effectively convey the popular prediction as well as uncertainty at different music sections along the temporal axis while enabling the analysis of individual ML models in conjunction with their application to different musical data. In addition to the traditional visual designs, such as stacked line graph, ThemeRiver, and pixel-based visualization, we introduce a new variant of ThemeRiver, called ``dual-flux ThemeRiver'', which allows viewers to observe and measure the most popular prediction more easily than stacked line graph and ThemeRiver. Together with pixel-based visualization, dual-flux ThemeRiver plots can also assist in model-development workflows, in addition to annotating music using ensemble model predictions.