Difference between revisions of "2021:Lyrics Transcription (former: Automatic Lyrics-to-Audio Alignment)"

From MIREX Wiki
(Potential Participants)
(Evaluation)
 
(50 intermediate revisions by the same user not shown)
Line 1: Line 1:
==Description==
+
= Description =
  
This year we host the Automatic Lyrics Transcription challenge. You are free to participate in one of the tasks or both of them.  
+
This year we host the '''MIREX2021: Automatic Lyrics Transcription''' challenge. You are free to participate in one of the tasks or both of them.  
The task of Lyrics Transcription aims to identify the words from sung music audio, in the same way as in automatic speech recognition. 
 
  
 +
The task of Lyrics Transcription aims to identify the words from sung utterances, in the same way as in automatic speech recognition. This can be mathematically expressed as follows:
  
 +
  Prediction('''w''') = argmax P('''w'''|'''X''')
  
 +
where '''w''' and '''X''' are the word and acoustic features respectively.
  
 +
Ideally, the lyrics transcriber should return meaningful word sequences:
  
 +
  Prediction('''w''')  = [ <w_1>, <w_2>, ..., <w_N> ]
  
The algorithm receives either monophonic singing performances or a polyphonic mix (singing voice + musical accompaniment). Ideally, the lyrics transcriber should output meaningful word sequences:
+
The algorithm receives either monophonic singing performances or a polyphonic mix (singing voice + musical accompaniment). Both cases are evaluated separately in this challenge.
  
 +
= Submission Format =
  
==Training Datasets==
+
Submissions should be packaged in a compressed file (.zip or .rar, etc.) which contains at least two files:
  
Datasets within automatic lyrics transcription research can be categorised under two domains: Monophonic and polyphonic recordings. The former is considered to have only one singer singing the lyrics, and the latter is when there is music accompaniment. In this challenge, we recommend using the open source datasets below:
+
=== A) The main transcription script ===
  
==== DAMP dataset ====
+
The main transcription script to execute. This should be a one-line executable in one of the following formats: a bash (.sh) a python (.py) script, or a binary file.
 +
 
 +
===  I / O ===
 +
 
 +
The submitted algorithm must take as arguments an audio file and the full output path to save the transcriptions. The ability to specify the output path and file name is essential.
 +
 
 +
Denoting the input audio filename path as $[input_audio_path} and the output file path and name as ${output}, a program called `foobar' will be called from the command-line as follows:
 +
 
 +
foobar ${input_audio_path}  ${output}
 +
 
 +
OR with flags:
 +
 
 +
foobar -i ${input_audio_path}  -o ${output}
 +
 
 +
==== Input Audio ====
 +
 
 +
Participating algorithms will have to receive the following input format:
 +
 
 +
* Audio format : WAV / MP3
 +
* CD-quality (PCM, 16-bit, 44100 Hz)
 +
* single channel (mono) for a cappella (Hansen) and two channels for original
 +
 
 +
==== Output File Format ====
 +
 
 +
A text file (per song) containing list of words separated by white space:
 +
 
 +
  <word_1> <word_2> ... <word_N>
 +
 
 +
Any non-word items (e.g. silence, music, noise or end of the sentence tokens) should be removed from the final output.
 +
 
 +
This file should ideally be located at:
 +
 +
  ${output}/${input_song_id}.txt
 +
 
 +
 
 +
=== B) The README file ===
 +
 
 +
This file must contain detailed installation instructions, the use of the main script and contact information.
 +
 
 +
----
 +
 
 +
Any submission that is failed to meet above requirements will not be considered in evaluation!
 +
 
 +
= Training Datasets =
 +
 
 +
Datasets within automatic lyrics transcription research can be categorised under two domains in regards to the presence of music instruments accompanying the singer: Monophonic and polyphonic datasets.
 +
 
 +
The former is considered to have only one singer singing the lyrics, and the latter is when there is music accompaniment.
 +
 
 +
In this challenge, the participants are encouraged but '''not obliged''' to use the open source datasets below, which are also commonly used in the literature for benchmarking ALT results:
 +
 
 +
=== DAMP dataset ===
 
The [https://zenodo.org/record/2747436#.Xyge4xMzZ0s DAMP - Sing!300x30x2 dataset] which consists of solo singing recordings (monophonic) performed by amateur singers, collected via a mobile Karaoke application. The data is curated to be gender-wise balanced and contains performers from 30 different countries, which introduces a good amount of variability in terms of accents and pronunciation.   
 
The [https://zenodo.org/record/2747436#.Xyge4xMzZ0s DAMP - Sing!300x30x2 dataset] which consists of solo singing recordings (monophonic) performed by amateur singers, collected via a mobile Karaoke application. The data is curated to be gender-wise balanced and contains performers from 30 different countries, which introduces a good amount of variability in terms of accents and pronunciation.   
 
[https://docs.google.com/spreadsheets/d/1YwhPhXU6t-BMZfdEODS_pNW_umFIsciYL62kh-fiBWI/edit?usp=sharing list of recordings]. For more details see the paper.  
 
[https://docs.google.com/spreadsheets/d/1YwhPhXU6t-BMZfdEODS_pNW_umFIsciYL62kh-fiBWI/edit?usp=sharing list of recordings]. For more details see the paper.  
Line 24: Line 80:
 
* Or annotations can be directly retrieved in the Kaldi form [https://github.com/emirdemirel/ALTA/s5/data here]
 
* Or annotations can be directly retrieved in the Kaldi form [https://github.com/emirdemirel/ALTA/s5/data here]
  
==== DALI Dataset ====
+
=== DALI Dataset ===
  
 
DALI (a large '''D'''ataset of synchronised '''A'''udio, '''L'''yr'''I'''cs and notes) is the benchmark dataset for building an acoustic model on polyphonic recordings and it contains over 5000 songs with semi-automatically aligned lyrics annotations. The songs are commercial recordings in full-duration, whereas the lyrics are described according to different levels of granularity including words and notes (and syllables underlying a given note). For each song DALI provides a link to a matched youtube video, from which the audio could be retrieved.
 
DALI (a large '''D'''ataset of synchronised '''A'''udio, '''L'''yr'''I'''cs and notes) is the benchmark dataset for building an acoustic model on polyphonic recordings and it contains over 5000 songs with semi-automatically aligned lyrics annotations. The songs are commercial recordings in full-duration, whereas the lyrics are described according to different levels of granularity including words and notes (and syllables underlying a given note). For each song DALI provides a link to a matched youtube video, from which the audio could be retrieved.
Line 30: Line 86:
 
* For more details how, see its full description [https://github.com/gabolsgabs/DALI here].
 
* For more details how, see its full description [https://github.com/gabolsgabs/DALI here].
  
==Evaluation Datasets==
+
= Evaluation Datasets =
 +
 
 +
The following datasets are used for evaluation and so '''cannot''' be used by participants to train their models under any circumstance.
  
The following datasets are used for evaluation and so cannot be used by participants to train their models under any circumstance. Note that the evaluation sets listed below consist of popular songs in English language, and have overlapping samples with DALI. In case using DALI for training, you MUST exclude the songs listed above in training the model for a scientific evaluation.  
+
Note that the evaluation sets listed below consist of popular songs in English language, and have overlapping samples with DALI. In case using DALI for training, you '''MUST''' exclude the songs listed above in training the model for a scientific evaluation.  
  
==== Hansen's Dataset ====
+
=== Hansen's Dataset ===
 
The dataset contains 9 pop music songs released in early 2010s.
 
The dataset contains 9 pop music songs released in early 2010s.
  
Line 44: Line 102:
 
* 3590 words annotated in total
 
* 3590 words annotated in total
  
==== Mauch's Dataset ====
+
=== Mauch's Dataset ===
  
 
The dataset contains 20 pop music songs with annotations of beginning-timestamps of each word.
 
The dataset contains 20 pop music songs with annotations of beginning-timestamps of each word.
Line 54: Line 112:
 
* 5050 words annotated in total
 
* 5050 words annotated in total
  
==== Jamendo Dataset ====
+
=== Jamendo Dataset ===
 +
 
 +
This dataset contains 20 recordings with varying Western music genres, annotated with start-of-word timestamps. All songs have instrumental accompaniment.
  
This dataset contains 20 recordings with varying Western music genres, annotated with start-of-word timestamps. All songs have instrumental accompaniment. It is available online on [https://github.com/f90/jamendolyrics Github], although note that we do not allow tuning model parameters using this data, it can only be used to gain insight into the general structure of the test data. For more information also refer to [https://arxiv.org/abs/1902.06797 this paper].
+
It is available online on [https://github.com/f90/jamendolyrics Github], although note that we do not allow tuning model parameters using this data, it can only be used to gain insight into the general structure of the test data. For more information also refer to [https://arxiv.org/abs/1902.06797 this paper].
  
 
* file duration up to 4:43 (total time: 1h 12m)
 
* file duration up to 4:43 (total time: 1h 12m)
 
* 5677 words annotated in total
 
* 5677 words annotated in total
  
==Evaluation==
+
= Evaluation =
  
 
Word Error Rate (WER) : the standard metric use in Automatic Speech Recognition.
 
Word Error Rate (WER) : the standard metric use in Automatic Speech Recognition.
  
 +
  WER = (S + I + D) / (C + S + I + D)
  
 
+
where;
 +
C : correctly predicted words
 +
S : substitution errors
 +
I : insertion errors
 +
D : deletion errors
  
  
 
Character Error Rate (CER) : the above computation can also be done on the character level. This metric penalises the partially correctly predicted / incorrectly spelled words less than WER.
 
Character Error Rate (CER) : the above computation can also be done on the character level. This metric penalises the partially correctly predicted / incorrectly spelled words less than WER.
  
== Submission Format ==
+
= Submission closing dates =
 
 
Submissions should be packaged and contain at least two files: The algorithm itself (as a binary or source code) and a README containing contact information and detailing, in full, the use of the algorithm.
 
 
 
=== Input Data ===
 
 
 
Participating algorithms will have to receive the following input format:
 
 
 
* Audio format : WAV / MP3
 
* CD-quality (PCM, 16-bit, 44100 Hz)
 
* single channel (mono) for a cappella (Hansen) and two channels for original
 
 
 
=== Output File Format ===
 
A text file (per song) containing list of words separated by white space:
 
 
 
 
 
  <word1> <word2> ... <wordN>
 
 
 
 
 
Any non-word items (e.g. silence, music, noise or end of the sentence tokens) should be removed from the final output.
 
 
 
This file should ideally be located at ${output}/${input_song_id}.txt .
 
 
 
=== Command line calling format ===
 
 
 
The submitted algorithm must take as arguments .wav file, .txt file as well as the full output path and filename of the output file. The ability to specify the output path and file name is essential. Denoting the input .wav file path and name as %input_audio; the lyrics .txt file as %input_txt and the output file path and name as %output, a program called foobar could be called from the command-line as follows:
 
 
 
foobar ${input_audio_path}  ${output}
 
 
 
OR with flags:
 
 
 
foobar -i ${input_audio_path}  -o ${output}
 
 
 
=== README File ===
 
 
 
A README file accompanying each submission should contain clear instructions on how to run the program (as well as contact information, etc.). In particular, each command line to run should be specified, using %input for the input sound file and %output for the resulting text file.
 
 
 
== Submission closing dates ==
 
  
Closing date: December 9, 2021
+
Closing date: '''December 9, 2021'''
  
== Questions? ==
+
= Questions? =
  
 
* send us an email - e.demirel@qmul.ac.uk (Emir Demirel) or info@voicemagix.com (Georgi Dzhambazov)
 
* send us an email - e.demirel@qmul.ac.uk (Emir Demirel) or info@voicemagix.com (Georgi Dzhambazov)
Line 128: Line 156:
  
 
Stoller, D. and Durand, S. and Ewert, S. (2019) End-to-end Lyrics Alignment for Polyphonic Music Using An Audio-to-Character Recognition Model. ICASSP 2019.
 
Stoller, D. and Durand, S. and Ewert, S. (2019) End-to-end Lyrics Alignment for Polyphonic Music Using An Audio-to-Character Recognition Model. ICASSP 2019.
 
Sharma B, Gupta C. (2019) Automatic Lyrics-to-audio Alignment on Polyphonic Music Using Singing-adapted Acoustic Models. ICASSP 2019
 
 
Lee S. W., Scott, J. (2017) Word-level lyrics-audio synchronization using separated vocals", Acoustics Speech and Signal Processing, ICASSP IEEE International Conference on, pp. 646-650
 
 
Chang, S., & Lee, K. (2017). Lyrics-to-Audio Alignment by Unsupervised Discovery of Repetitive Patterns in Vowel Acoustics. arXiv preprint arXiv:1701.06078.
 
 
Pons, J. Gong, R. and Serra, X. (2017). Score-informed syllable segmentation for a cappella singing voice with convolutional neural networks. ISMIR 2017
 
 
Kruspe, A. (2016). Bootstrapping a System for Phoneme Recognition and Keyword Spotting in Unaccompanied Singing, ISMIR 2016
 
 
Dzhambazov, G. and Serra, X. (2015) Modeling of phoneme durations for alignment between polyphonic audio and lyrics, in 12th Sound and Music Computing Conference
 
 
Fujihara, H., & Goto, M. (2012). Lyrics-to-audio alignment and its application. In Dagstuhl Follow-Ups (Vol. 3). Schloss Dagstuhl-Leibniz-Zentrum fuer Informatik.
 
  
 
Mauch, M., Fujihara, H., & Goto, M. (2012). Integrating additional chord information into HMM-based lyrics-to-audio alignment. IEEE Transactions on Audio, Speech, and Language Processing, 20(1), 200-210.
 
Mauch, M., Fujihara, H., & Goto, M. (2012). Integrating additional chord information into HMM-based lyrics-to-audio alignment. IEEE Transactions on Audio, Speech, and Language Processing, 20(1), 200-210.
 
Fujihara, H. Goto, M. Ogata, J. and Okuno, H. G. (2011) Lyricsynchronizer: Automatic synchronization system between musical audio signals and lyrics, IEEE Journal of Selected Topics in Signal Processing
 
 
Mesaros, A. and Virtanen, T. (2008), Automatic alignment of music audio and lyrics, in Proceedings of the 11th Int. Conference on Digital Audio Effects (DAFx-08), Espoo, Finland, 2008.
 

Latest revision as of 16:56, 26 October 2021

Description

This year we host the MIREX2021: Automatic Lyrics Transcription challenge. You are free to participate in one of the tasks or both of them.

The task of Lyrics Transcription aims to identify the words from sung utterances, in the same way as in automatic speech recognition. This can be mathematically expressed as follows:

 Prediction(w) = argmax P(w|X)

where w and X are the word and acoustic features respectively.

Ideally, the lyrics transcriber should return meaningful word sequences:

 Prediction(w)  = [ <w_1>, <w_2>, ..., <w_N> ]

The algorithm receives either monophonic singing performances or a polyphonic mix (singing voice + musical accompaniment). Both cases are evaluated separately in this challenge.

Submission Format

Submissions should be packaged in a compressed file (.zip or .rar, etc.) which contains at least two files:

A) The main transcription script

The main transcription script to execute. This should be a one-line executable in one of the following formats: a bash (.sh) a python (.py) script, or a binary file.

I / O

The submitted algorithm must take as arguments an audio file and the full output path to save the transcriptions. The ability to specify the output path and file name is essential.

Denoting the input audio filename path as $[input_audio_path} and the output file path and name as ${output}, a program called `foobar' will be called from the command-line as follows:

foobar ${input_audio_path}  ${output}

OR with flags:

foobar -i ${input_audio_path}  -o ${output}

Input Audio

Participating algorithms will have to receive the following input format:

  • Audio format : WAV / MP3
  • CD-quality (PCM, 16-bit, 44100 Hz)
  • single channel (mono) for a cappella (Hansen) and two channels for original

Output File Format

A text file (per song) containing list of words separated by white space:

 <word_1> <word_2> ... <word_N>

Any non-word items (e.g. silence, music, noise or end of the sentence tokens) should be removed from the final output.

This file should ideally be located at:

 ${output}/${input_song_id}.txt


B) The README file

This file must contain detailed installation instructions, the use of the main script and contact information.


Any submission that is failed to meet above requirements will not be considered in evaluation!

Training Datasets

Datasets within automatic lyrics transcription research can be categorised under two domains in regards to the presence of music instruments accompanying the singer: Monophonic and polyphonic datasets.

The former is considered to have only one singer singing the lyrics, and the latter is when there is music accompaniment.

In this challenge, the participants are encouraged but not obliged to use the open source datasets below, which are also commonly used in the literature for benchmarking ALT results:

DAMP dataset

The DAMP - Sing!300x30x2 dataset which consists of solo singing recordings (monophonic) performed by amateur singers, collected via a mobile Karaoke application. The data is curated to be gender-wise balanced and contains performers from 30 different countries, which introduces a good amount of variability in terms of accents and pronunciation. list of recordings. For more details see the paper.

  • The audio can be downloaded from the Smule web site
  • Lyrics boundary annotations can be generated from raw annotations using this repository.
  • Or annotations can be directly retrieved in the Kaldi form here

DALI Dataset

DALI (a large Dataset of synchronised Audio, LyrIcs and notes) is the benchmark dataset for building an acoustic model on polyphonic recordings and it contains over 5000 songs with semi-automatically aligned lyrics annotations. The songs are commercial recordings in full-duration, whereas the lyrics are described according to different levels of granularity including words and notes (and syllables underlying a given note). For each song DALI provides a link to a matched youtube video, from which the audio could be retrieved.

  • For more details how, see its full description here.

Evaluation Datasets

The following datasets are used for evaluation and so cannot be used by participants to train their models under any circumstance.

Note that the evaluation sets listed below consist of popular songs in English language, and have overlapping samples with DALI. In case using DALI for training, you MUST exclude the songs listed above in training the model for a scientific evaluation.

Hansen's Dataset

The dataset contains 9 pop music songs released in early 2010s.

The audio has two versions: the original mix with instrumental accompaniment and a cappella singing voice only one. An example song can be seen here.

You can read in detail about how the dataset was made here: Recognition of Phonemes in A-cappella Recordings using Temporal Patterns and Mel Frequency Cepstral Coefficients. The recordings have been provided by Jens Kofod Hansen for public evaluation.

  • file duration up to 4:40 minutes (total time: 35:33 minutes)
  • 3590 words annotated in total

Mauch's Dataset

The dataset contains 20 pop music songs with annotations of beginning-timestamps of each word. The audio has instrumental accompaniment. An example song can be seen here.

You can read in detail about how the dataset was used for the first time here: Integrating Additional Chord Information Into HMM-Based Lyrics-to-Audio Alignment. The dataset has been kindly provided by Sungkyun Chang.

  • file duration up to 5:40 minutes (total time: 1h 19m)
  • 5050 words annotated in total

Jamendo Dataset

This dataset contains 20 recordings with varying Western music genres, annotated with start-of-word timestamps. All songs have instrumental accompaniment.

It is available online on Github, although note that we do not allow tuning model parameters using this data, it can only be used to gain insight into the general structure of the test data. For more information also refer to this paper.

  • file duration up to 4:43 (total time: 1h 12m)
  • 5677 words annotated in total

Evaluation

Word Error Rate (WER) : the standard metric use in Automatic Speech Recognition.

 WER = (S + I + D) / (C + S + I + D)

where;

C : correctly predicted words
S : substitution errors
I : insertion errors
D : deletion errors


Character Error Rate (CER) : the above computation can also be done on the character level. This metric penalises the partially correctly predicted / incorrectly spelled words less than WER.

Submission closing dates

Closing date: December 9, 2021

Questions?

  • send us an email - e.demirel@qmul.ac.uk (Emir Demirel) or info@voicemagix.com (Georgi Dzhambazov)

Potential Participants

Chitralekha Gupta

Emir Demirel

Gerardo Roa Dabike

Jiawen Huang

Bibliography

Stoller, D. and Durand, S. and Ewert, S. (2019) End-to-end Lyrics Alignment for Polyphonic Music Using An Audio-to-Character Recognition Model. ICASSP 2019.

Mauch, M., Fujihara, H., & Goto, M. (2012). Integrating additional chord information into HMM-based lyrics-to-audio alignment. IEEE Transactions on Audio, Speech, and Language Processing, 20(1), 200-210.