This folder contain various tool for performing evaluation on various benchmarking Dataset for Table detection and structure recognition.
Official site : http://www.tamirhassan.com/html/competition.html
Download the Complete Data Here
Get our Result XML Here
This tool is tested on Windows 7 only
- Java Oracle 8
- Python 3
- Provided JAR file need to be placed in you java jre directory, inside the lib/ext
The format for PDF, Ground Truth and Result should be in the same format as specified in ICDAR 13 official site
Also, the naming convention as specified by ICDAR 13 are
file.pdf, file-reg.xml, file-reg-result.xml
Where, file.pdf is the PDF on which the detection is performed,
file-reg.xml is the Groundthruth XML file used for table Detection
file-reg-result.xml Contains the output Detected coordinate of algorithm.
Evalutation using orignal tool on single PDF can be found Here
Put all the PDF, their groundtruth and result XML in same directory as the eval.py and tool.java
simply run
python eval.py
to get result
Train
eu-003
eu-015
eu-020
eu-027
us-018
us-019
us-034
us-009
us-015
us-017
us-036
us-033
us-032
us-028
us-022
Test
Rest of PDFs are used for testing
ICDAR 19 evalution Tool can be found Here and Dataset Here
Evaluation is carried out in following steps:
- Model Loading
- For each image in test set
i) Extract ground truth for the corresponding image from annotations
ii) Pass the image to model to get its predictions
iii) Correct the predictions using line correction technique
iv) Map the ground truth bounding box to its appropriate predicted box. This is done using the euclidean distance.
v) Calculate the IoU for each pair of boxes (Ground-truth vs predicted)
vi) Using the IoU value calculate the precision and recall per object
vii) perform step iii) to v) for each table in image - Calculate the average precision and recall across all images to get the final result.
Above steps are implemented in evaluation.py file.
Download train and test json files for all of the three subsets (Latex, Word, Both) from Here
Set appropriate paths to of Words.json and Latex.json in script.
After setting paths run
python evaluation.py
Evaluation on tablebank is done using line correction on the model predictions.