[3Rin4-76] Fine-tuning of BERT Based on Triplet Network with Relations between Wikipedia Articles
Keywords:Natural Language Processing, Sentence Embeddings, Fine-tuning
Recently, sentence embedding approaches by fine-tuning a pre-trained language model have achieved significant success in several NLP tasks. In order to improve the semantic representation of sentences, Ein Dor et al. have introduced a triplet network to fine-tune a pre-trained model, where each triplet consists of an anchor text and its related and unrelated sentences extracted from Wikipedia articles. In this work, we propose to improve the triplet network to learn the relatedness of articles by using the hyperlink network in Wikipedia. We first quantify the relatedness of two articles with the number of hops in the hyperlink network. The triplets are rebuilt according to the relatedness, and moreover, the loss function is modified to incorporate such relatedness to the model. Our result shows that our proposed model improves the performance in tasks of thematic similarity prediction.
Authentication for paper PDF access
A password is required to view paper PDFs. If you are a registered participant, please log on the site from Participant Log In.
You could view the PDF with entering the PDF viewing password bellow.