The Combination of BERT and Data Oversampling for Relation Set Prediction

Thang Ta Hoang, Sabur Butt, Jason Angel, Grigori Sidorov, Alexander Gelbukh

Research output: Contribution to journalConference articlepeer-review

1 Scopus citations

Abstract

In this paper, we engage the Task 2 of the SMART Task 2021 challenge in predicting relations used to identify the correct answer of a given question. This is a subtask of Knowledge Base Question Answering (KBQA) and offers valuable insights for the development of KBQA systems. We introduce our method, combining BERT and data oversampling with text replacements of linked terms to Wikidata and dependent noun phrases, in predicting answer relations in two datasets. For the DBpedia dataset, we obtain F1 of 83.15%, precision of 83.68%, and recall of 82.95%. Meanwhile, for the Wikidata dataset we achieved F1 of 60.70%, precision of 61.63%, and recall of 61.10%.

Original languageEnglish
JournalCEUR Workshop Proceedings
Volume3119
StatePublished - 2022
Event2nd SeMantic Answer Type and Relation Prediction Task at ISWC Semantic Web Challenge, SMART 2021 - Virtual, Online
Duration: 26 Oct 2021 → …

Keywords

  • ISWC
  • Knowledge Base Question Answering
  • Relation Linking
  • Relation Prediction
  • Semantic Web Challenge

Fingerprint

Dive into the research topics of 'The Combination of BERT and Data Oversampling for Relation Set Prediction'. Together they form a unique fingerprint.

Cite this