Skip to Main content Skip to Navigation
Conference papers

Does constituency analysis enhance domain-specific pre-trained BERT models for relation extraction?

Abstract : Recently many studies have been conducted on the topic of relation extraction. The DrugProt track at BioCreative VII provides a manually-annotated corpus for the purpose of the development and evaluation of relation extraction systems, in which interactions between chemicals and genes are studied. We describe the ensemble system that we used for our submission, which combines predictions of fine-tuned bioBERT, sciBERT and const-bioBERT models by majority voting. We specifically tested the contribution of syntactic information to relation extraction with BERT. We observed that adding constituentbased syntactic information to BERT improved precision, but decreased recall, since relations rarely seen in the train set were less likely to be predicted by BERT models in which the syntactic information is infused. Our code is available online [https://github.com/Maple177/drugprot-relation-extraction].
Complete list of metadata

https://hal.archives-ouvertes.fr/hal-03447774
Contributor : Claire Nédellec Connect in order to contact the contributor
Submitted on : Wednesday, November 24, 2021 - 7:01:29 PM
Last modification on : Friday, January 21, 2022 - 4:13:40 AM

File

Track1_pos_6_BC7_submission_14...
Files produced by the author(s)

Licence


Distributed under a Creative Commons Attribution 4.0 International License

Identifiers

  • HAL Id : hal-03447774, version 1
  • ARXIV : 2112.02955

Citation

Anfu Tang, Louise Deléger, Robert Bossy, Pierre Zweigenbaum, Claire Nédellec. Does constituency analysis enhance domain-specific pre-trained BERT models for relation extraction?. BioCreative VII Challenge Evaluation Workshop, Nov 2021, on-line, Spain. ⟨hal-03447774⟩

Share

Metrics

Les métriques sont temporairement indisponibles