Difference between revisions of "2011:Symbolic Melodic Similarity"

From MIREX Wiki
(Submission closing date)
(Evaluation)
 
(3 intermediate revisions by the same user not shown)
Line 19: Line 19:
 
==Evaluation ==
 
==Evaluation ==
  
The 2010 SMS task replicates the 2007 task. After the algorithms have been submitted, their results will be pooled for every query, and human evaluators, using the Evalutron 6000 system, will asked to judge the relevance of the matches to the queries.[[2010:Evalutron6000_Walkthrough |2010 Evalutron6000 Walkthrough Page]].
+
The 2011 SMS task replicates the 2007 task. After the algorithms have been submitted, their results will be pooled for every query, and human evaluators, using the Evalutron 6000 system, will asked to judge the relevance of the matches to the queries.[[2011:Evalutron6000_Walkthrough |2011 Evalutron6000 Walkthrough Page]].
  
 
For each query (and its four mutations), the returned results (candidates) from all systems will be anonymously grouped together (query set) for evaluation by the human graders. The graders will be provided with only the "base" perfect version against which to evaluate the candidates and thus did not know whether the candidates came from a perfect or mutated query. We expect that each query/candidate set will be evaluated by one individual grader. Using the Evalutron 6000 system, the graders will give each query/candidate pair two types of scores. Graders will be asked to provide one "BROAD" categorical score with three categories: NS,SS,VS as explained below, and one "FINE" score (in the range from 0 to 10).
 
For each query (and its four mutations), the returned results (candidates) from all systems will be anonymously grouped together (query set) for evaluation by the human graders. The graders will be provided with only the "base" perfect version against which to evaluate the candidates and thus did not know whether the candidates came from a perfect or mutated query. We expect that each query/candidate set will be evaluated by one individual grader. Using the Evalutron 6000 system, the graders will give each query/candidate pair two types of scores. Graders will be asked to provide one "BROAD" categorical score with three categories: NS,SS,VS as explained below, and one "FINE" score (in the range from 0 to 10).
Line 76: Line 76:
 
* All submissions should be statically linked to all libraries (the presence of dynamically linked libraries cannot be guaranteed). [mailto:mirproject@lists.lis.uiuc.edu IMIRSEL] should be notified of any dependencies that you cannot include with your submission at the earliest opportunity (in order to give them time to satisfy the dependency).
 
* All submissions should be statically linked to all libraries (the presence of dynamically linked libraries cannot be guaranteed). [mailto:mirproject@lists.lis.uiuc.edu IMIRSEL] should be notified of any dependencies that you cannot include with your submission at the earliest opportunity (in order to give them time to satisfy the dependency).
 
* Be sure to follow the [[2009:Best Coding Practices for MIREX | Best Coding Practices for MIREX]]
 
* Be sure to follow the [[2009:Best Coding Practices for MIREX | Best Coding Practices for MIREX]]
* Be sure to follow the [[MIREX 2010 Submission Instructions]]
+
* Be sure to follow the [[MIREX 2011 Submission Instructions]]
  
 
All submissions should include a README file including the following the information:
 
All submissions should include a README file including the following the information:

Latest revision as of 14:10, 8 July 2011

Description

The goal of SMS is to retrieve the most similar items from a collection of symbolic pieces, given a symbolic query, and rank them by melodic similarity. There will be only 1 task this year which comprises a set of six "base" monophonic MIDI queries to be matched against a monophonic MIDI collection.

Each system will be given a query and is asked to return the 10 most melodically similar songs from those taken from the Essen Collection (5274 pieces in the MIDI format; see ESAC Data Homepage for more information). For each of the six "base" queries, we have created four classes of error-mutations, thus the query set comprises the following query classes:

  1. No errors (i.e., "base")
  2. One note deleted
  3. One note inserted
  4. One interval enlarged
  5. One interval compressed

Each system will be asked to return the top ten items for each of the 30 total queries. That is to say, 6(base queries) X 5(versions) = 30 query/candidate lists to be returned.


Data

  • 5,274 tunes belonging to the Essen folksong collection. The tunes are in standard MIDI file format. Download (< 1 MB)


Evaluation

The 2011 SMS task replicates the 2007 task. After the algorithms have been submitted, their results will be pooled for every query, and human evaluators, using the Evalutron 6000 system, will asked to judge the relevance of the matches to the queries.2011 Evalutron6000 Walkthrough Page.

For each query (and its four mutations), the returned results (candidates) from all systems will be anonymously grouped together (query set) for evaluation by the human graders. The graders will be provided with only the "base" perfect version against which to evaluate the candidates and thus did not know whether the candidates came from a perfect or mutated query. We expect that each query/candidate set will be evaluated by one individual grader. Using the Evalutron 6000 system, the graders will give each query/candidate pair two types of scores. Graders will be asked to provide one "BROAD" categorical score with three categories: NS,SS,VS as explained below, and one "FINE" score (in the range from 0 to 10).

For more information, do take a look at the 2007 SMS Results Page.

Submission Format

Input

A path to the folder containing midi files.

Output

A list of the names of the 10 most similar matching MIDI files, ordered by melodic similarity output to the stdout. Please use stderr for the log and error messages.

E.g.

song242.mid song213.mid song1242.mid ...
song5454.mid song423.mid song454.mid ...
...

Commandline Calling Format

There are 2 different calling formats depending on if the program does indexing or not. We encourage the indexing step since programs are called for each query.

2 Step Calling Format

First Indexing step: E.g.

doIndex.sh -numThreads 4(optional) /path/to/scratch/folder(optional) /path/to/folder/withMIDIfiles/ /path/indexFileOrDirectory/

Where

/path/to/folder/withMIDIfiles/ - the path to the directory containing about 5,000 MIDI files containing monophonic folk songs

/path/indexFileOrDirectory/ - The path to the index file or folder that will be used in the query step.

Query in the second step: E.g.

doQuery.sh -numThreads 4(optional) /path/to/scratch/folder(optional) /path/to/indexFileOrFolder/ /path/to/query.mid

/path/to/indexFileOrFolder/ - Path to the index file or folder created in the first step.

/path/to/query.mid - path to the query midi file.

The program will be called once for each query.

1 Step Calling Format

If your program does not perform indexing.

E.g.

doSMS.sh -numThreads 4(optional) /path/to/scratch/folder /path/to/folder/withMIDIfiles/ /path/to/query.mid

The program will be called once for each query.


Packaging submissions

  • All submissions should be statically linked to all libraries (the presence of dynamically linked libraries cannot be guaranteed). IMIRSEL should be notified of any dependencies that you cannot include with your submission at the earliest opportunity (in order to give them time to satisfy the dependency).
  • Be sure to follow the Best Coding Practices for MIREX
  • Be sure to follow the MIREX 2011 Submission Instructions

All submissions should include a README file including the following the information:

  • Command line calling format for all executables including examples
  • Number of threads/cores used or whether this should be specified on the command line
  • Expected memory footprint
  • Expected runtime
  • Approximately how much scratch disk space will the submission need to store any feature/cache files?
  • Any required environments/architectures (and versions) such as Matlab, Java, Python, Bash, Ruby etc.
  • Any special notice regarding to running your algorithm

Note that the information that you place in the README file is extremely important in ensuring that your submission is evaluated properly.

Time and hardware limits

Due to the potentially high number of participants in this and other audio tasks, hard limits on the runtime of submissions will be imposed.

A hard limit of 48 hours total runtime will be imposed on each submission. Submissions exceeding this runtime may not receive a result.


Submission opening date

Friday 5th August 2011

Submission closing date

Friday 26th August 2011

Potential Participant List

name / email