• Login
    JavaScript is disabled for your browser. Some features of this site may not work without it.
    Integrating Additional Chord Information Into HMM-Based Lyrics-to-Audio Alignment 
    •   QMRO Home
    • School of Electronic Engineering and Computer Science
    • Electronic Engineering and Computer Science
    • Integrating Additional Chord Information Into HMM-Based Lyrics-to-Audio Alignment
    •   QMRO Home
    • School of Electronic Engineering and Computer Science
    • Electronic Engineering and Computer Science
    • Integrating Additional Chord Information Into HMM-Based Lyrics-to-Audio Alignment
    ‌
    ‌

    Browse

    All of QMROCommunities & CollectionsBy Issue DateAuthorsTitlesSubjectsThis CollectionBy Issue DateAuthorsTitlesSubjects
    ‌
    ‌

    Administrators only

    Login
    ‌
    ‌

    Statistics

    Most Popular ItemsStatistics by CountryMost Popular Authors

    Integrating Additional Chord Information Into HMM-Based Lyrics-to-Audio Alignment

    View/Open
    Published Version
    Embargoed until: 2100-01-01
    Reason: VoR
    Volume
    20
    Pagination
    200 - 210 (11)
    DOI
    10.1109/TASL.2011.2159595
    Journal
    IEEE Transactions on Audio, Speech and Language Processing
    Issue
    1
    Metadata
    Show full item record
    Abstract
    Aligning lyrics to audio has a wide range of applications such as the automatic generation of karaoke scores, song-browsing by lyrics, and the generation of audio thumbnails. Existing methods are restricted to using only lyrics and match them to phoneme features extracted from the audio (usually mel-frequency cepstral coefficients). Our novel idea is to integrate the textual chord information provided in the paired chords-lyrics format known from song books and Internet sites into the inference procedure. We propose two novel methods that implement this idea: First, assuming that all chords of a song are known, we extend a hidden Markov model (HMM) framework by including chord changes in the Markov chain and an additional audio feature (chroma) in the emission vector; second, for the more realistic case in which some chord information is missing, we present a method that recovers the missing chord information by exploiting repetition in the song. We conducted experiments with five changing parameters and show that with accuracies of 87.5% and 76.7%, respectively, both methods perform better than the baseline with statistical significance. We introduce the new accompaniment interface Song Prompter, which uses the automatically aligned lyrics to guide musicians through a song. It demonstrates that the automatic alignment is accurate enough to be used in a musical performance.
    Authors
    MAUCH, M; Fujihara, H; Goto, M
    URI
    http://qmro.qmul.ac.uk/xmlui/handle/123456789/2875
    Collections
    • Electronic Engineering and Computer Science [2816]
    Copyright statements
    © Copyright 2011 IEEE - All rights reserved.
    Twitter iconFollow QMUL on Twitter
    Twitter iconFollow QM Research
    Online on twitter
    Facebook iconLike us on Facebook
    • Site Map
    • Privacy and cookies
    • Disclaimer
    • Accessibility
    • Contacts
    • Intranet
    • Current students

    Modern Slavery Statement

    Queen Mary University of London
    Mile End Road
    London E1 4NS
    Tel: +44 (0)20 7882 5555

    © Queen Mary University of London.