Expected outputs for the Competition
For the competition, it was mandatory to deliver the executable files.
Participants will provide
- an executable file
- a description of the required environment, resources and the expected processing time for 1000 images.
The executable file should:
- read the images from a folder
- a “ symmetrical distance matrix” of the images present in the folder (for task 1 as well as for task 2). The sum of all entries of this matrix will be equal to 1.0.
- a CSV file with 2 columns: “FILENAME, SCRIPT_TYPE ” for task 1 (SCRIPT_TYPE, an integer between 1 and 12, indicating the belonging class).
- a CSV file with 13 columns: FILENAME,SCRIPT_TYPE1, …., SCRIPT_TYPE12” for task 2 (SCRIPT_TYPE1, …., SCRIPT_TYPE12 are real values, and each row has a sum of its values equal to 1.0). This information will be named Belonging Matrix in the following sections.
Nota: Participants are allowed to submit several independent proposals
Based on the test data-set, the evaluation will be given as follow:
- Accuracy per script type
- Global accuracy for fuzzy results
- Normalized distance matrix analysis
- Processing time
For each task, two rankings will be done. The first one will be based on the average global accuracy, the second based on the symmetrical distance matrix.
The “accuracy per script type” is given according to the ground truth, which has one label for each script image in the evaluation data set. The ranking will be based on the average accuracy.
As far as the distance matrix is concerned, the evaluation will be done on the average intraclasse distance that will yield to the second ranking.
The “global accuracy for fuzzy results” will be evaluated as follow:
- The ground truth indicates one or two labels for each script image, from the Belonging Matrix, only the two highest membership degrees will define SCRIPT_TYPE1 and SCRIPT_TYPE2.
- Scores will be attributed as follow:
- 4 points if SCRIPT_TYPE1 AND SCRIPT_TYPE2 match the labels given to the image in the ground truth.
- 2 points if SCRIPT_TYPE1 matches one of the labels given to the image in the ground truth, but not SCRIPT_TYPE2.
- 1 point if SCRIPT_TYPE2 matches one label given to the image in the ground truth, but not SCRIPT_TYPE1.
- (-2) points if SCRIPT_TYPE1 AND SCRIPT_TYPE2 do not match any of the labels given to the image in the ground truth.
The ranking is based on the average accuracy.
As far as the distance matrix is concerned, for each image, according to the ground truth, we consider the one or two classes with highest membership degree. Then the sum of distances to the images of the same ground truth classes will be computed and averaged all over the test base. The ranking will be done with respect to this global value.