1:00 PM - 1:20 PM
[1J3-OS-10a-01] Serendipity Assessment in Recommender Systems Using Large Language Models
Keywords:information recommendation, serendipity, large language model, value judgement
Serendipity-oriented recommender systems have been proposed to prevent over-specialization in user preferences. However, evaluating serendipity is challenging due to its reliance on subjective user emotions. We try to address this issue by leveraging the rich knowledge of large language models (LLMs), which can perform a variety of tasks. As a first step, this study investigates the alignment between serendipity assessments made by LLMs and those by humans. Specifically, using GPT-3.5 we assess the serendipity of recommended items based on users' evaluation history. We evaluate the accuracy of assessment made by the LLM using an annotated benchmark dataset. Experimental results indicate that our method outperforms the baseline method, showing improvements of up to 0.6, 4.9 and 1.5 points in Accuracy, Precision and Macro-F1-score.
Authentication for paper PDF access
A password is required to view paper PDFs. If you are a registered participant, please log on the site from Participant Log In.
You could view the PDF with entering the PDF viewing password bellow.