Difference between revisions of "Evaltex"

From LRDE

Line 19: Line 19:
   
 
==== Performance measurements ====
 
==== Performance measurements ====
  +
===== Recall =====
  +
The Recall computes the amount of detected text
 
* RG = Global Recall
 
* RG = Global Recall
* PG = Global Precision
 
 
 
* R_quant = Quantitative Recall
 
* R_quant = Quantitative Recall
 
* R_qual = Qualitative Recall
 
* R_qual = Qualitative Recall
  +
  +
===== Precision =====
  +
The Precision computes the amount of detections with a match in the GT
 
* P_G = Global Precision
  +
* P_quant = Quantitative Precision
  +
* P_qual = Qualitative Precision

Revision as of 14:27, 8 January 2016

EvaLTex (Evaluating Text Localization) is an evaluation tool used to measure the performance of text detection algorithms. It takes as input text detection results that can be represented either by coordinates or by masks and outputs performance scores.

XML format

Input

The framework takes as input XML files containing the coordinates of the bounding boxes surrounding the text objects and compares it to another XML file representing the ground truth (GT). The GT XML format differs slightly from the result format. Its attributes are:

  • name : the image name
  • size : image size
  • region : 1st level components

Mask representation

In addition to bounding boxes, text objects can also be represented using masks. EvaLTex takes as input binary images (white corresponds to text). TODO add images

Output

The output consists in a local valuation (for each image), as well as a global evaluation (one XML file for a whole database).

TODO add local XML file and global XML file

Performance measurements

Recall

The Recall computes the amount of detected text

  • RG = Global Recall
  • R_quant = Quantitative Recall
  • R_qual = Qualitative Recall
Precision

The Precision computes the amount of detections with a match in the GT

  • P_G = Global Precision
  • P_quant = Quantitative Precision
  • P_qual = Qualitative Precision