Chord Recognition in Beatles Songs

Chord Recognition in Beatles Songs

While a graduate student at MIT’s Media Lab, I collaborated with office-mate Vi­ctor Adán to explore how if we might train a machine to recognize chord changes in music. We tried multiple models to solve the problem, including Support Vector Machines, Neural Networks, Hidden Markov Models, and a few variations of Maximum Likelihood systems.

We chose Beatles tunes as a subset of the larger problem and trained our systems with 16 songs from three of their albums. Our systems processed 2700 training samples, 150 validation samples, and 246 testing samples. Our most successful system, a Support Vector Machine, achieved 68% accuracy in testing.

Our intention was to further the research which will lead to applications such as automatic transcription, live tracking for improvisation, and computer-assisted (synthetic) performers. Our models were an extension of the research provided by the following papers:

  • Musical Key Extraction from Audio, Steffen Pauws
  • Chord Segmentation and Recognition using EM-Trained Hidden Markov Models, Alexander Sheh and Daniel P.W. Ellis
  • SmartMusicKIOSK: Music Listening Station with Chorus-Search Function, Masataka Goto
  • A Chorus-Section Detecting Method for Musical Audio Signals, Masataka Goto

Main Website

Similar Posts

  • TV Mountain

    TV Mountain is an art piece which was initially installed as part of TechArtICT’s Dreamgrass installation at Exploration Place Fall 2022. It consisted of 12 TVs at various angles and orientations, together forming a jumbotron. As part of Dreamgrass, it subjected faceless child-like objects to a litany of 80s commercials. The working remote enjoyed a…

  • SoundBlocks

    SoundBlocks is a tangible environment where youth connect blocks to describe network dataflow. The environment explores digital sound manipulation as a personal, meaningful and fun artistic endeavor, rather than as a venture into mathematical, electronic or networking relationships.

  • Digital Puppetry

    I worked with a team of colleagues, community members, and urban youth. Our intention was to help the youth learn in a playful environment, find personal self-expression, and have their voices heard by communities in Boston. To do this, we adapted commercially available technology to provide a unique medium: digital puppetry.

  • Still Life

    In 2011, as part of Hack.Art.Lab, I collaborated with composer Mary Ellen Childs and percussionist Michael Holland to create live animation triggered by live performance of Mary Ellen Childs’ composition “Still Life.” We analyzed the piece into 11 sections and created algorithmic video triggered by sound and motion to match each of the 11 sections. The video was projected…

  • TechArtICT: Whispering Woodlands

    Whispering Woodlands was an outdoor installation created by TechArtICT. It was installed at Exploration Place in Wichita, Kansas from November 2023 through January 2024. The work featured 24 independently controlled sets of speakers and LEDs, all synchronized to create an immersive sound and lightscape. Using eclectic audio ranging from thunder and rain to spaceships and…