Difference between revisions of "Publications/chazalon.21.icdar.1"
From LRDE
(Created page with "{{Publication | published = true | date = 2020-05-17 | title = Revisiting the Coco Panoptic Metric to Enable Visual and Qualitative Analysis of Historical Map Instance Segment...") |
|||
(4 intermediate revisions by the same user not shown) | |||
Line 1: | Line 1: | ||
{{Publication |
{{Publication |
||
| published = true |
| published = true |
||
− | | date = |
+ | | date = 2021-05-17 |
| title = Revisiting the Coco Panoptic Metric to Enable Visual and Qualitative Analysis of Historical Map Instance Segmentation |
| title = Revisiting the Coco Panoptic Metric to Enable Visual and Qualitative Analysis of Historical Map Instance Segmentation |
||
| authors = Joseph Chazalon, Edwin Carlinet |
| authors = Joseph Chazalon, Edwin Carlinet |
||
| booktitle = Proceedings of the 16th International Conference on Document Analysis and Recognition (ICDAR'21) |
| booktitle = Proceedings of the 16th International Conference on Document Analysis and Recognition (ICDAR'21) |
||
+ | | series = Lecture Notes in Computer Science |
||
⚫ | |||
+ | | publisher = Springer, Cham |
||
+ | | volume = 12824 |
||
⚫ | |||
| address = Lausanne, Switzerland |
| address = Lausanne, Switzerland |
||
− | | abstract = Segmentation is an important task. It is so important that there exist tens of metrics trying to score and rank segmentation systems. It is so important that each topic has its own metric because their problem is too specific. Does it? What are the fundamental differences with the ZoneMap metric used for page segmentation, the COCO Panoptic metric used in computer vision and metrics used to rank hierarchical segmentations? In this |
+ | | abstract = Segmentation is an important task. It is so important that there exist tens of metrics trying to score and rank segmentation systems. It is so important that each topic has its own metric because their problem is too specific. Does it? What are the fundamental differences with the ZoneMap metric used for page segmentation, the COCO Panoptic metric used in computer vision and metrics used to rank hierarchical segmentations? In this paper, while assessing segmentation accuracy for historical maps, we explain, compare and demystify some the most used segmentation evaluation protocols. In particular, we focus on an alternative view of the COCO Panoptic metric as a classification evaluation; we show its soundness and propose extensions with more “shape-oriented” metrics. Beyond a quantitative metric, this paper aims also at providing qualitative measures through precision-recall maps that enable visualizing the success and the failures of a segmentation method. |
+ | | lrdepaper = https://www.lrde.epita.fr/dload/papers/chazalon.21.icdar.1.pdf |
||
+ | | lrdeposter = https://www.lrde.epita.fr/dload/papers/chazalon.21.icdar.1.poster.pdf |
||
| lrdeprojects = Olena |
| lrdeprojects = Olena |
||
| lrdekeywords = Image |
| lrdekeywords = Image |
||
− | | lrdenewsdate = |
+ | | lrdenewsdate = 2021-05-17 |
| type = inproceedings |
| type = inproceedings |
||
| id = chazalon.21.icdar.1 |
| id = chazalon.21.icdar.1 |
||
+ | | identifier = doi:10.1007/978-3-030-86337-1_25 |
||
| bibtex = |
| bibtex = |
||
@InProceedings<nowiki>{</nowiki> chazalon.21.icdar.1, |
@InProceedings<nowiki>{</nowiki> chazalon.21.icdar.1, |
||
− | title = <nowiki>{</nowiki>Revisiting the |
+ | title = <nowiki>{</nowiki>Revisiting the <nowiki>{</nowiki>C<nowiki>}</nowiki>oco Panoptic Metric to Enable Visual and |
Qualitative Analysis of Historical Map Instance |
Qualitative Analysis of Historical Map Instance |
||
Segmentation<nowiki>}</nowiki>, |
Segmentation<nowiki>}</nowiki>, |
||
Line 23: | Line 29: | ||
year = <nowiki>{</nowiki>2021<nowiki>}</nowiki>, |
year = <nowiki>{</nowiki>2021<nowiki>}</nowiki>, |
||
month = sep, |
month = sep, |
||
− | + | series = <nowiki>{</nowiki>Lecture Notes in Computer Science<nowiki>}</nowiki>, |
|
+ | publisher = <nowiki>{</nowiki>Springer, Cham<nowiki>}</nowiki>, |
||
+ | volume = <nowiki>{</nowiki>12824<nowiki>}</nowiki>, |
||
+ | pages = <nowiki>{</nowiki>367--382<nowiki>}</nowiki>, |
||
address = <nowiki>{</nowiki>Lausanne, Switzerland<nowiki>}</nowiki>, |
address = <nowiki>{</nowiki>Lausanne, Switzerland<nowiki>}</nowiki>, |
||
abstract = <nowiki>{</nowiki>Segmentation is an important task. It is so important that |
abstract = <nowiki>{</nowiki>Segmentation is an important task. It is so important that |
||
Line 42: | Line 51: | ||
providing qualitative measures through |
providing qualitative measures through |
||
\emph<nowiki>{</nowiki>precision-recall maps<nowiki>}</nowiki> that enable visualizing the |
\emph<nowiki>{</nowiki>precision-recall maps<nowiki>}</nowiki> that enable visualizing the |
||
− | success and the failures of a segmentation method.<nowiki>}</nowiki> |
+ | success and the failures of a segmentation method.<nowiki>}</nowiki>, |
+ | doi = <nowiki>{</nowiki>10.1007/978-3-030-86337-1_25<nowiki>}</nowiki> |
||
<nowiki>}</nowiki> |
<nowiki>}</nowiki> |
||
Latest revision as of 10:54, 8 September 2021
- Authors
- Joseph Chazalon, Edwin Carlinet
- Where
- Proceedings of the 16th International Conference on Document Analysis and Recognition (ICDAR'21)
- Place
- Lausanne, Switzerland
- Type
- inproceedings
- Publisher
- Springer, Cham
- Projects
- Olena
- Keywords
- Image
- Date
- 2021-05-17
Abstract
Segmentation is an important task. It is so important that there exist tens of metrics trying to score and rank segmentation systems. It is so important that each topic has its own metric because their problem is too specific. Does it? What are the fundamental differences with the ZoneMap metric used for page segmentation, the COCO Panoptic metric used in computer vision and metrics used to rank hierarchical segmentations? In this paper, while assessing segmentation accuracy for historical maps, we explain, compare and demystify some the most used segmentation evaluation protocols. In particular, we focus on an alternative view of the COCO Panoptic metric as a classification evaluation; we show its soundness and propose extensions with more “shape-oriented” metrics. Beyond a quantitative metric, this paper aims also at providing qualitative measures through precision-recall maps that enable visualizing the success and the failures of a segmentation method.
Documents
Bibtex (lrde.bib)
@InProceedings{ chazalon.21.icdar.1, title = {Revisiting the {C}oco Panoptic Metric to Enable Visual and Qualitative Analysis of Historical Map Instance Segmentation}, author = {Joseph Chazalon and Edwin Carlinet}, booktitle = {Proceedings of the 16th International Conference on Document Analysis and Recognition (ICDAR'21)}, year = {2021}, month = sep, series = {Lecture Notes in Computer Science}, publisher = {Springer, Cham}, volume = {12824}, pages = {367--382}, address = {Lausanne, Switzerland}, abstract = {Segmentation is an important task. It is so important that there exist tens of metrics trying to score and rank segmentation systems. It is so important that each topic has its own metric because their problem is too specific. Does it? What are the fundamental differences with the ZoneMap metric used for page segmentation, the COCO Panoptic metric used in computer vision and metrics used to rank hierarchical segmentations? In this paper, while assessing segmentation accuracy for historical maps, we explain, compare and demystify some the most used segmentation evaluation protocols. In particular, we focus on an alternative view of the COCO Panoptic metric as a classification evaluation; we show its soundness and propose extensions with more ``shape-oriented'' metrics. Beyond a quantitative metric, this paper aims also at providing qualitative measures through \emph{precision-recall maps} that enable visualizing the success and the failures of a segmentation method.}, doi = {10.1007/978-3-030-86337-1_25} }