Biomedical Interpretable Entity Representations
Diego Garcia-Olano, Yasumasa Onoe, Ioana Baldini, Joydeep Ghosh, Byron Wallace, Kush Varshney
Findings of ACL 2021
- Paper: [ https://aclanthology.org/2021.findings-acl.311/ ]
- ACL slides: [ pdf ]
@inproceedings{garcia-olano-etal-2021-biomedical,
title = "Biomedical Interpretable Entity Representations",
author = "Garcia-Olano, Diego and
Onoe, Yasumasa and
Baldini, Ioana and
Ghosh, Joydeep and
Wallace, Byron and
Varshney, Kush",
booktitle = "Findings of the 59th Annual Meeting of the Association for Computational Linguistics",
year = "2021",
publisher = "Association for Computational Linguistics",
}
To use pre-trained models without re-training BIERS, see colab notebooks in "Replicating downstream tasks" section at bottom.
$ git clone https://github.com/diegoolano/biomedical_interpretable_entity_representations.git
$ virtualenv --python=~/envs/py37/bin/python biomed_env
$ source biomed_env/bin/activate
$ pip install -r requirements.txt
See ier_model/train.sh
Make sure to: - set goal to "medwiki", - set training and dev sets, - set paths in transformers\_constants.py appropriately, - make sure to use a GPU with a lot of memory ( ie v100 has 32GB) or lower the batch size. - set the intervals on which you'd like to get training acc, eval acc on dev, etc - set log location
BIER triples can be found [ here ]
Model files:
- BIER-PubMedBERT: [ model ckpt ]
- BIER-SciBERT: [ model ckpt ]
- BIER-BioBERT: [ model ckpt ]
See prior section for how to train BIER models using training data
See Colabs below for how to load and use models on downstream tasks
See experiments/README.md for baselines
- after generating (mention, context, categories) triples we then learn BIERs as follows:
- after learning BIERs we can test their efficacy in a Zeroshot capacity for different biomed tasks