Inproceedings3942: Unterschied zwischen den Versionen
Jg2894 (Diskussion | Beiträge) |
Jg2894 (Diskussion | Beiträge) |
||
Zeile 15: | Zeile 15: | ||
|Title=AIFB-WebScience at SemEval-2022 Task 12: Relation Extraction First - Using Relation Extraction to Identify Entities | |Title=AIFB-WebScience at SemEval-2022 Task 12: Relation Extraction First - Using Relation Extraction to Identify Entities | ||
|Year=2022 | |Year=2022 | ||
− | |||
|Booktitle=Proceedings of the 16th International Workshop on Semantic Evaluation (SemEval-2022) | |Booktitle=Proceedings of the 16th International Workshop on Semantic Evaluation (SemEval-2022) | ||
|Publisher=Association for Computational Linguistics | |Publisher=Association for Computational Linguistics |
Aktuelle Version vom 4. Mai 2022, 14:33 Uhr
AIFB-WebScience at SemEval-2022 Task 12: Relation Extraction First - Using Relation Extraction to Identify Entities
AIFB-WebScience at SemEval-2022 Task 12: Relation Extraction First - Using Relation Extraction to Identify Entities
Published: 2022
Buchtitel: Proceedings of the 16th International Workshop on Semantic Evaluation (SemEval-2022)
Verlag: Association for Computational Linguistics
Referierte Veröffentlichung
BibTeX
Kurzfassung
In this paper, we present an end-to-end joint entity and relation extraction approach based on transformer-based language models. We apply the model to the task of linking mathematical symbols to their descriptions in LaTeX documents. In contrast to existing approaches, which perform entity and relation extraction in sequence, our system incorporates information from relation extraction into entity extraction. This means that the system can be trained even on data sets where only a subset of all valid entity spans is annotated. We provide an extensive evaluation of the proposed system and its strengths and weaknesses. Our approach, which can be scaled dynamically in computational complexity at inference time, produces predictions with high precision and reaches 3rd place in the leaderboard of SemEval-2022 Task 12. For inputs in the domain of physics and math, it achieves high relation extraction macro F1 scores of 95.43% and 79.17%, respectively. The code used for training and evaluating our models is available at: https://github.com/nicpopovic/RE1st
Download: Media:Semeval2022_RE1st.pdf
Text Mining, Informationsextraktion, Natürliche Sprachverarbeitung, Deep Learning