With so much music readily available for consumption today, it has never been more important to study music perception. In this paper, we represent lyrics and chords in a shared vector space using a phrase-aligned lyrics-and-chords corpus and show that models that use these shared representations can predict musical genre of songs—a perceptual construct of music listening—better than models that do not use these representations. This work adds to our understanding of how lyrics and chords interact with one another in music and has applications in multimodal perception and music information retrieval.
Cite as: Greer, T., Narayanan, S. (2019) Using Shared Vector Representations of Words and Chords in Music for Genre Classification. Proc. Workshop on Speech, Music and Mind (SMM 2019), 46-50, doi: 10.21437/SMM.2019-10
@inproceedings{greer19_smm, author={Timothy Greer and Shrikanth Narayanan}, title={{Using Shared Vector Representations of Words and Chords in Music for Genre Classification}}, year=2019, booktitle={Proc. Workshop on Speech, Music and Mind (SMM 2019)}, pages={46--50}, doi={10.21437/SMM.2019-10} }