Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Inter-annotator agreements #27

Open
twinkarma opened this issue May 12, 2021 · 8 comments
Open

Inter-annotator agreements #27

twinkarma opened this issue May 12, 2021 · 8 comments
Labels
enhancement New feature or request

Comments

@twinkarma
Copy link
Collaborator

twinkarma commented May 12, 2021

  • Support calculator IAA for

    • Feature where you can apply a nominal, ordinal or continuous scale i.e. only single selectable option
      • Radio buttons, Selector, Checkboxes where you can only select a single option
    • Multiple choice to be supported later, we'd require set arithmatic
  • Extract for each pair of annotators all the annotations they have done

  • Calculate simple matches if doc classification

  • For sequence labelling, we need to do this with begin/end offsets and per annotation type

@twinkarma twinkarma added the enhancement New feature or request label Oct 21, 2021
@davidwilby
Copy link
Contributor

discuss with Xingyi (@deansong), he has scripts for doing this

@davidwilby
Copy link
Contributor

rules in old teamware are same as rules in GATE developer

@twinkarma
Copy link
Collaborator Author

  • Check with team for existing inter-annotator agreement code.
  • Also ask the team for testing data to double check this

@twinkarma
Copy link
Collaborator Author

twinkarma commented Mar 6, 2023

Common remedial action

  • Remove all annotator's annotation from project, remove from project
  • Revise guideline for all annotators (will start new project in this case)

@ianroberts
Copy link
Member

ianroberts commented Apr 24, 2023

Things to consider - different types of IAA metrics for different types of task

  • binary / "nominal" classification - exact match is "correct", anything else is "wrong"
  • Likert-type "ordinal" scales with a finite number of options which are ordered, so e.g. 4 vs 5 is "closer" than 4 vs 1
  • continuous numerical scales

Maybe this needs to be part of the project configuration for each widget, so we know what sort of metrics to use for that particular question.

@ianroberts
Copy link
Member

Variables may not be independent, e.g. a classification plus associated confidence score - do we want to use the confidence score as a weighting of some kind or as a threshold (e.g. only compute IAA on pairs where both annotators confidence is > 4)

@twinkarma
Copy link
Collaborator Author

We need to check what the standard procedure is for comparing multiple choice agreement

@davidwilby
Copy link
Contributor

Depends upon defined scale for a particular project #349

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

3 participants