Difference between revisions of "2009:Audio Beat Tracking Results"
(→McKinney Collection) |
(→Understanding the Metrics) |
||
(23 intermediate revisions by 4 users not shown) | |||
Line 1: | Line 1: | ||
+ | == Introduction == | ||
+ | This year was the first year since MIREX 2006 that we have run the Audio Beat Track (ABT) task. ABT 2009 saw two collection used, the original ''McKinney Collection'', and the new ''Sapp's Mazurka Collection''. A number of new scoring metrics were also introduced this year beyond McKinney's P-Score. | ||
+ | |||
+ | ===McKinney Collection Information=== | ||
+ | The McKinney Collection comprises 140 files of mixed genres. There are 40 ground-truth sets per music file across which the beats were averaged. | ||
+ | |||
+ | ===Sapp's Mazurka Collection Information=== | ||
+ | New this year, we used 322 files drawn from the Mazurka.org dataset put together by Craig Sapp. Craig Sapp was also responsible for creating the high-quality ground-truth files. | ||
+ | |||
+ | ===Understanding the Metrics=== | ||
+ | The evaluation methods were taken from the beat evaluation toolbox and | ||
+ | are described in the following technical report: | ||
+ | |||
+ | M. E. P. Davies, N. Degara and M. D. Plumbley. "Evaluation methods for musical audio beat tracking algorithms". [https://music-ir.org/mirex/results/2009/beat/techreport_beateval.pdf ''Technical Report C4DM-TR-09-06'']. | ||
+ | |||
+ | For further details on the specifics of the methods please refer to the | ||
+ | paper. However, here is a brief summary with appropriate references: | ||
+ | |||
+ | --------------------------------------------------------------------- | ||
+ | '''F-measure''' - the standard calculation as used in onset evaluation but | ||
+ | with a ±70ms window. | ||
+ | |||
+ | S. Dixon, "Onset detection revisited," in ''Proceedings of 9th | ||
+ | International Conference on Digital Audio Effects (DAFx)'', Montreal, | ||
+ | Canada, pp. 133ΓÇô137, 2006. | ||
+ | |||
+ | S. Dixon, "Evaluation of audio beat tracking system beatroot," ''Journal | ||
+ | of New Music Research'', vol. 36, no. 1, pp. 39ΓÇô51, 2007. | ||
+ | |||
+ | --------------------------------------------------------------------- | ||
+ | '''Cemgil''' - beat accuracy is calculated using a Gaussian error function | ||
+ | with 40ms standard deviation. | ||
+ | |||
+ | A. T. Cemgil, B. Kappen, P. Desain, and H. Honing, "On tempo tracking: | ||
+ | Tempogram representation and Kalman filtering," ''Journal Of New Music | ||
+ | Research'', vol. 28, no. 4, pp. 259ΓÇô273, 2001 | ||
+ | |||
+ | --------------------------------------------------------------------- | ||
+ | '''Goto''' - binary decision of correct or incorrect tracking based on | ||
+ | statistical properties of a beat error sequence. | ||
+ | |||
+ | M. Goto and Y. Muraoka, "Issues in evaluating beat tracking systems," in | ||
+ | ''Working Notes of the IJCAI-97 Workshop on Issues in AI and Music - | ||
+ | Evaluation and Assessment'', 1997, pp. 9ΓÇô16. | ||
+ | |||
+ | --------------------------------------------------------------------- | ||
+ | '''PScore''' - McKinney's impulse train cross-correlation method as used in | ||
+ | 2006. | ||
+ | |||
+ | M. F. McKinney, D. Moelants, M. E. P. Davies, and A. Klapuri, | ||
+ | "Evaluation of audio beat tracking and music tempo extraction | ||
+ | algorithms," ''Journal of New Music Research'', vol. 36, no. 1, pp. 1ΓÇô16, | ||
+ | 2007. | ||
+ | |||
+ | --------------------------------------------------------------------- | ||
+ | '''CMLc''', '''CMLt''', '''AMLc''', '''AMLt''' - continuity-based evaluation methods based on | ||
+ | the longest continuously correctly tracked section. | ||
+ | |||
+ | S. Hainsworth, "Techniques for the automated analysis of musical audio," | ||
+ | Ph.D. dissertation, Department of Engineering, Cambridge University, | ||
+ | 2004. | ||
+ | |||
+ | A. P. Klapuri, A. Eronen, and J. Astola, "Analysis of the meter of | ||
+ | acoustic musical signals," IEEE Transactions on Audio, Speech and | ||
+ | Language Processing, vol. 14, no. 1, pp. 342ΓÇô355, 2006. | ||
+ | |||
+ | --------------------------------------------------------------------- | ||
+ | '''D''', '''Dg''' - information based criteria based on analysis of a beat error | ||
+ | histogram (note the results are measured in 'bits' and not percentages), | ||
+ | see the technical report for a description. | ||
+ | |||
== General Legend == | == General Legend == | ||
====Team ID==== | ====Team ID==== | ||
− | '''DRP1''' = [ | + | '''DRP1''' = [https://www.music-ir.org/mirex/abstracts/2009/drp.pdf Matthew Davies, Andrew Robertson, Mark Plumbley (Deterministic)] <br /> |
− | '''DRP2''' = [ | + | '''DRP2''' = [https://www.music-ir.org/mirex/abstracts/2009/drp.pdf Matthew Davies, Andrew Robertson, Mark Plumbley (Dumb)] <br /> |
− | '''DRP3''' = [ | + | '''DRP3''' = [https://www.music-ir.org/mirex/abstracts/2009/drp.pdf Matthew Davies, Andrew Robertson, Mark Plumbley (Flexible)] <br /> |
− | '''DRP4''' = [ | + | '''DRP4''' = [https://www.music-ir.org/mirex/abstracts/2009/drp.pdf Matthew Davies, Andrew Robertson, Mark Plumbley (Standard)] <br /> |
− | '''GP1''' = [ | + | '''GP1''' = [https://www.music-ir.org/mirex/abstracts/2009/Peeters_2009_MIREX_beat.pdf Geoffroy Peeters (VF)] <br /> |
− | '''GP2''' = [ | + | '''GP2''' = [https://www.music-ir.org/mirex/abstracts/2009/Peeters_2009_MIREX_beat.pdf Geoffroy Peeters (VE)] <br /> |
− | '''GP3''' = [ | + | '''GP3''' = [https://www.music-ir.org/mirex/abstracts/2009/Peeters_2009_MIREX_beat.pdf Geoffroy Peeters (CF)] <br /> |
− | '''GP4''' = [ | + | '''GP4''' = [https://www.music-ir.org/mirex/abstracts/2009/Peeters_2009_MIREX_beat.pdf Geoffroy Peeters (CE)] <br /> |
− | ''' | + | '''OGM1''' = [https://www.music-ir.org/mirex/abstracts/2009/GTfinal.pdf Joao Lobato Oliveira, Fabien Gouyon, Luis Gustavo Martins (SC)] <br /> |
− | ''' | + | '''OGM2''' = [https://www.music-ir.org/mirex/abstracts/2009/GTfinal.pdf Joao Lobato Oliveira, Fabien Gouyon, Luis Gustavo Martins(R)] <br /> |
− | '''TL''' = [ | + | '''TL''' = [https://www.music-ir.org/mirex/abstracts/2009/TL.pdf Tsung-Chi Lee] <br /> |
+ | ==Results== | ||
===McKinney Collection=== | ===McKinney Collection=== | ||
− | <csv>beat/ | + | <csv>2009/beat/mckinney_fixed_c.csv</csv> |
+ | |||
+ | ===Sapp's Mazurka Collection=== | ||
+ | <csv>2009/beat/mazurka_fixed_b.csv</csv> | ||
− | === | + | ===MIREX 2009 Audio Beat Tracking Runtime Data=== |
− | <csv>beat/ | + | <csv>2009/beat/beat.runtime.csv</csv> |
Latest revision as of 16:12, 23 July 2010
Contents
Introduction
This year was the first year since MIREX 2006 that we have run the Audio Beat Track (ABT) task. ABT 2009 saw two collection used, the original McKinney Collection, and the new Sapp's Mazurka Collection. A number of new scoring metrics were also introduced this year beyond McKinney's P-Score.
McKinney Collection Information
The McKinney Collection comprises 140 files of mixed genres. There are 40 ground-truth sets per music file across which the beats were averaged.
Sapp's Mazurka Collection Information
New this year, we used 322 files drawn from the Mazurka.org dataset put together by Craig Sapp. Craig Sapp was also responsible for creating the high-quality ground-truth files.
Understanding the Metrics
The evaluation methods were taken from the beat evaluation toolbox and are described in the following technical report:
M. E. P. Davies, N. Degara and M. D. Plumbley. "Evaluation methods for musical audio beat tracking algorithms". Technical Report C4DM-TR-09-06.
For further details on the specifics of the methods please refer to the paper. However, here is a brief summary with appropriate references:
F-measure - the standard calculation as used in onset evaluation but with a ±70ms window.
S. Dixon, "Onset detection revisited," in Proceedings of 9th International Conference on Digital Audio Effects (DAFx), Montreal, Canada, pp. 133ΓÇô137, 2006.
S. Dixon, "Evaluation of audio beat tracking system beatroot," Journal of New Music Research, vol. 36, no. 1, pp. 39ΓÇô51, 2007.
Cemgil - beat accuracy is calculated using a Gaussian error function with 40ms standard deviation.
A. T. Cemgil, B. Kappen, P. Desain, and H. Honing, "On tempo tracking: Tempogram representation and Kalman filtering," Journal Of New Music Research, vol. 28, no. 4, pp. 259ΓÇô273, 2001
Goto - binary decision of correct or incorrect tracking based on statistical properties of a beat error sequence.
M. Goto and Y. Muraoka, "Issues in evaluating beat tracking systems," in Working Notes of the IJCAI-97 Workshop on Issues in AI and Music - Evaluation and Assessment, 1997, pp. 9ΓÇô16.
PScore - McKinney's impulse train cross-correlation method as used in 2006.
M. F. McKinney, D. Moelants, M. E. P. Davies, and A. Klapuri, "Evaluation of audio beat tracking and music tempo extraction algorithms," Journal of New Music Research, vol. 36, no. 1, pp. 1ΓÇô16, 2007.
CMLc, CMLt, AMLc, AMLt - continuity-based evaluation methods based on the longest continuously correctly tracked section.
S. Hainsworth, "Techniques for the automated analysis of musical audio," Ph.D. dissertation, Department of Engineering, Cambridge University, 2004.
A. P. Klapuri, A. Eronen, and J. Astola, "Analysis of the meter of acoustic musical signals," IEEE Transactions on Audio, Speech and Language Processing, vol. 14, no. 1, pp. 342ΓÇô355, 2006.
D, Dg - information based criteria based on analysis of a beat error histogram (note the results are measured in 'bits' and not percentages), see the technical report for a description.
General Legend
Team ID
DRP1 = Matthew Davies, Andrew Robertson, Mark Plumbley (Deterministic)
DRP2 = Matthew Davies, Andrew Robertson, Mark Plumbley (Dumb)
DRP3 = Matthew Davies, Andrew Robertson, Mark Plumbley (Flexible)
DRP4 = Matthew Davies, Andrew Robertson, Mark Plumbley (Standard)
GP1 = Geoffroy Peeters (VF)
GP2 = Geoffroy Peeters (VE)
GP3 = Geoffroy Peeters (CF)
GP4 = Geoffroy Peeters (CE)
OGM1 = Joao Lobato Oliveira, Fabien Gouyon, Luis Gustavo Martins (SC)
OGM2 = Joao Lobato Oliveira, Fabien Gouyon, Luis Gustavo Martins(R)
TL = Tsung-Chi Lee
Results
McKinney Collection
TeamID | F-Measure | Cemgil | Goto | P-Score | CMLC | CMLT | AMLC | AMLT | D | Dg |
---|---|---|---|---|---|---|---|---|---|---|
units -> | (%) | (%) | (%) | (%) | (%) | (%) | (%) | (%) | (bits) | (bits) |
DRP1 | 24.6 | 17.5 | 0.4 | 37.4 | 3.8 | 9.6 | 7.8 | 19.2 | 0.625 | 0.008 |
DRP2 | 35.4 | 26.1 | 3.1 | 46.5 | 7.5 | 16.9 | 13.8 | 31.7 | 1.073 | 0.062 |
DRP3 | 47.0 | 35.6 | 16.2 | 52.8 | 19.4 | 27.6 | 45.3 | 61.3 | 1.671 | 0.245 |
DRP4 | 48.0 | 36.3 | 19.5 | 53.9 | 22.2 | 29.1 | 50.8 | 64.0 | 1.739 | 0.255 |
GP1 | 54.8 | 41.0 | 22.2 | 59.0 | 26.0 | 35.5 | 49.1 | 66.6 | 1.680 | 0.294 |
GP2 | 53.7 | 40.1 | 20.9 | 57.9 | 25.1 | 33.7 | 47.6 | 63.8 | 1.649 | 0.281 |
GP3 | 54.6 | 40.9 | 22.0 | 59.0 | 26.1 | 35.6 | 48.8 | 66.3 | 1.695 | 0.287 |
GP4 | 54.5 | 40.8 | 21.6 | 59.2 | 26.4 | 35.5 | 48.2 | 64.7 | 1.685 | 0.289 |
OGM1 | 39.7 | 29.1 | 6.5 | 47.8 | 11.9 | 19.9 | 29.5 | 47.4 | 1.283 | 0.100 |
OGM2 | 41.5 | 30.4 | 5.7 | 49.2 | 11.7 | 19.5 | 28.2 | 46.4 | 1.340 | 0.092 |
TL | 50.0 | 37.0 | 14.1 | 53.6 | 17.9 | 24.3 | 35.9 | 50.8 | 1.480 | 0.150 |
Sapp's Mazurka Collection
TeamID | F-Measure | Cemgil | Goto | P-Score | CMLC | CMLT | AMLC | AMLT | D | Dg |
---|---|---|---|---|---|---|---|---|---|---|
units -> | (%) | (%) | (%) | (%) | (%) | (%) | (%) | (%) | (bits) | (bits) |
DRP1 | 27.9 | 19.9 | 0.0 | 37.3 | 1.5 | 10.3 | 2.3 | 13.8 | 0.126 | 0.008 |
DRP2 | 48.4 | 32.5 | 0.0 | 55.5 | 2.6 | 26.0 | 3.1 | 28.1 | 0.683 | 0.183 |
DRP3 | 67.8 | 61.5 | 2.5 | 65.3 | 7.8 | 42.1 | 9.2 | 45.6 | 1.424 | 0.993 |
DRP4 | 45.3 | 37.6 | 0.6 | 50.1 | 3.9 | 24.9 | 5.3 | 28.5 | 0.393 | 0.198 |
GP1 | 54.2 | 42.5 | 0.0 | 59.5 | 4.4 | 33.2 | 5.6 | 36.6 | 0.447 | 0.252 |
GP2 | 54.7 | 43.1 | 0.0 | 59.9 | 4.5 | 33.7 | 5.7 | 37.0 | 0.467 | 0.269 |
GP3 | 44.7 | 34.8 | 0.0 | 51.2 | 3.2 | 24.7 | 4.0 | 27.9 | 0.262 | 0.111 |
GP4 | 45.4 | 35.6 | 0.0 | 51.9 | 3.3 | 25.2 | 4.0 | 28.4 | 0.284 | 0.124 |
OGM1 | 30.7 | 21.9 | 0.0 | 38.6 | 1.7 | 11.0 | 2.8 | 15.4 | 0.134 | 0.014 |
OGM2 | 32.1 | 23.1 | 0.0 | 39.7 | 1.7 | 11.5 | 2.7 | 15.6 | 0.134 | 0.016 |
TL | 44.9 | 35.7 | 0.3 | 51.0 | 4.4 | 20.5 | 4.7 | 22.5 | 0.440 | 0.095 |
MIREX 2009 Audio Beat Tracking Runtime Data
Participant | Machine | Runtime (hh:mm) |
---|---|---|
DRP1 | ALE | 00:35 |
DRP2 | ALE | 00:42 |
DRP3 | ALE | 01:19 |
DRP4 | ALE | 01:16 |
GP1 | ALE | 00:33 |
GP2 | ALE | 00:34 |
GP3 | ALE | 00:34 |
GP4 | ALE | 00:32 |
OGM1 | ALE | 01:52 |
OGM2 | ALE | 01:51 |
TL | ALE | 00:51 |