Skip to content

Commit

Permalink
massaging the docu
Browse files Browse the repository at this point in the history
  • Loading branch information
lfoppiano committed Dec 15, 2023
1 parent 9d16314 commit 2220802
Show file tree
Hide file tree
Showing 2 changed files with 37 additions and 30 deletions.
56 changes: 29 additions & 27 deletions doc/evaluation.rst
Original file line number Diff line number Diff line change
@@ -1,3 +1,5 @@
.. topic:: Evaluation

**********
Evaluation
**********
Expand Down Expand Up @@ -76,19 +78,21 @@ Quantities
Units
-----

+------------------+-----------+--------+---------+------------+--------+---------+
| | CRF | BERT_CRF |
+==================+===========+========+=========+============+========+=========+
| Labels | P | R | F1 | P | R | F1 |
+------------------+-----------+--------+---------+------------+--------+---------+
| `<base>` | 80.57 | 82.34 | 81.45 | 61.41 | 57.08 | 59.16 |
+------------------+-----------+--------+---------+------------+--------+---------+
| `<pow>` | 72.65 | 74.45 | 73.54 | 91.24 | 64.60 | 75.60 |
+------------------+-----------+--------+---------+------------+--------+---------+
| `<prefix>` | 93.80 | 84.69 | 89.02 | 82.10 | 85.30 | 83.62 |
+------------------+-----------+--------+---------+------------+--------+---------+
| All (micro avg) | 80.73 | 80.60 | 80.66 | 73.02 | 64.97 | 68.76 |
+------------------+-----------+--------+---------+------------+--------+---------+
Units were evaluated using UNISCOR dataset. For more information check the section :ref:`uniscor`.

+------------------+-----------+---------+---------+------------+--------+---------+
| | CRF | BERT_CRF |
+==================+===========+=========+=========+============+========+=========+
| Labels | P | R | F1 | P | R | F1 |
+------------------+-----------+---------+---------+------------+--------+---------+
| `<base>` | 80.57 | 82.34 | 81.45 | 61.41 | 57.08 | 59.16 |
+------------------+-----------+---------+---------+------------+--------+---------+
| `<pow>` | 72.65 | 74.45 | 73.54 | 91.24 | 64.60 | 75.60 |
+------------------+-----------+---------+---------+------------+--------+---------+
| `<prefix>` | 93.80 | 84.69 | 89.02 | 82.10 | 85.30 | 83.62 |
+------------------+-----------+---------+---------+------------+--------+---------+
| All (micro avg) | **80.73** |**80.60**|**80.66**| 73.02 | 64.97 | 68.76 |
+------------------+-----------+---------+---------+------------+--------+---------+

+------------------+--------------+--------+---------+-------------------------+--------+---------+
| | BidLSTM_CRF | BidLSTM_CRF_FEATURES |
Expand Down Expand Up @@ -141,15 +145,18 @@ Values
| All (micro avg) | 98.90 | 99.17 | 99.03 | 98.86 | 99.25 | 99.05 |
+-----------------+------------+--------+----------+-------------------------+---------+----------+

====================
Previous evaluations
====================
================
Previous results
================

The scores of this evaluation were obtained using n-fold cross-validation. The metrics are the micro average of n=10 folds.


Previous evaluation was provided using 10-fold cross-validation (with average metrics over the 10 folds).
.. note:: The `paper` "Automatic Identification and Normalisation of Physical Measurements in Scientific Literature," published in September 2019, reported macro averaged evaluation scores.

The `CRF` model was evaluated on the 30/04/2020.
The `BidLSTM_CRF_FEATURES` model was evaluated on the 28/11/2021
Evaluation notes:
- The `CRF` model was evaluated on the 30/04/2020.
- The `BidLSTM_CRF_FEATURES` model was evaluated on the 28/11/2021

----------
Quantities
Expand Down Expand Up @@ -200,9 +207,9 @@ CRF was updated on the 10/02/2021
+------------------+------------+--------+----------+-----------+-------+-----------+



------
Values
^^^^^^
------

+------------------+------------+--------+----------+-----------+--------+----------+
| | CRF | BidLSTM_CRF_FEATURES |
Expand All @@ -220,9 +227,4 @@ Values
| `<time>` | 73.07 | 86.82 | 79.26 | 80.84 | 100.00| 89.28 |
+------------------+------------+--------+----------+-----------+-------+-----------+
| All (micro avg) | 96.15 | 97.95 | 97.40 | 98.49 | 98.66 | 98.57 |
+------------------+------------+--------+----------+-----------+-------+-----------+


The current average results have been calculated using micro average, which provides more realistic results by giving
different weights to labels based on their frequency.
The `paper` "Automatic Identification and Normalisation of Physical Measurements in Scientific Literature," published in September 2019, reported average evaluation based on macro average.
+------------------+------------+--------+----------+-----------+-------+-----------+
11 changes: 8 additions & 3 deletions doc/references.rst
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,9 @@ How to cite

If you want to cite this work, please simply refer to the github project, with optionally the `Software Heritage <https://www.softwareheritage.org/>`_ project-level permanent identifier:
::
grobid-quantities (2015-2022) <https://github.com/kermitt2/grobid-quantities>, swh:1:dir:dbf9ee55889563779a09b16f9c451165ba62b6d7
grobid-quantities (2015-2022)
<https://github.com/kermitt2/grobid-quantities>,
swh:1:dir:dbf9ee55889563779a09b16f9c451165ba62b6d7

Here's a BibTeX entry using the `Software Heritage <https://www.softwareheritage.org/>`_ project-level permanent identifier:
::
Expand Down Expand Up @@ -38,8 +40,11 @@ Main papers about grobid-quantities
| https://arxiv.org/pdf/1710.04312.pdf

Other
~~~~~

.. _uniscor:

Datasets
~~~~~~~~

UNISCOR (Units Segmentation Corpus) is a corpus of "unit segmentation" and is available `here`_.
It was created with the support of `NIMS (National Institute for Material Science) <https://www.nims.go.jp>`_, in Japan.
Expand Down

0 comments on commit 2220802

Please sign in to comment.