Difference between revisions of "2006:2006 Plenary Notes"

From MIREX Wiki
Line 8: Line 8:
 
*Appreciation to IMIRSEL team members.
 
*Appreciation to IMIRSEL team members.
  
=Overview=
+
==Overview==
  
 
*This year MIREX is highly successful. We got everything done on time!
 
*This year MIREX is highly successful. We got everything done on time!
Line 29: Line 29:
 
By tuning the parameters, we can get an optimal setting which is a tradeoff between precision and recall.  
 
By tuning the parameters, we can get an optimal setting which is a tradeoff between precision and recall.  
 
We need new dataset to see if the tuned parameters are good for onseen data.
 
We need new dataset to see if the tuned parameters are good for onseen data.
 +
Question: comparison to last year results?
 +
Answer:
 +
 +
==Evalutron 6000==
 +
Two judgments:
 +
*category judgment: Not similar; Similar; Very similar
 +
*continurous score: from 0 to 10, allowing one decimal after the decimal point.
 +
*the system: using CMS open source software
 +
*still have data that we haven't fully processed (other user/evaluator behaviors)
 +
*new evaluation on other facets? e.g. mood
 +
*suggestions?
 +
*appreciate evaluators' volunteer work. Your work makes life beautiful!

Revision as of 15:41, 12 October 2006

Oct. 12th @ Empress Crystal Hall, Victoria

Openning

Professor Stephen Downie gave the openning remarks:

  • We will present certificates for participants. Feel free to grab yours if you are leaving.
  • Appreciation to IMIRSEL team members.

Overview

  • This year MIREX is highly successful. We got everything done on time!
  • Matlab is widely used (universal retrieval language!)
  • All the evaluation result data files are available on the wiki.

Tasks

  • We had sub-tasks as tasks are getting matured.
  • New tasks:
    • Audio cover song: 13 different songs, each of which has 11 different versions
    • Score following: have ground work done for future years
    • QBSH: 48 ground truth melodies. Different versions of queries on the 48 melodies. About 2000 noise songs were selected from Essen dataset. Both audio input and MIDI input are supported.
  • Please think about new tasks next year.
  • New evaluations:
    • Evalutron 6000 got real-world human judgment.
    • Audio onset detection supported multiple parameters.
    • Friedman test: It is valuable experience from TREC conferences, the annual contests in Text Retrieval area.

Onset Detection

By tuning the parameters, we can get an optimal setting which is a tradeoff between precision and recall. We need new dataset to see if the tuned parameters are good for onseen data. Question: comparison to last year results? Answer:

Evalutron 6000

Two judgments:

  • category judgment: Not similar; Similar; Very similar
  • continurous score: from 0 to 10, allowing one decimal after the decimal point.
  • the system: using CMS open source software
  • still have data that we haven't fully processed (other user/evaluator behaviors)
  • new evaluation on other facets? e.g. mood
  • suggestions?
  • appreciate evaluators' volunteer work. Your work makes life beautiful!