[4Xin2-66] Hallucination Detection in Japanese LLMs under Zero-Resource Black-Box Fixed-Low-Temperature Constraint Through Data-Augmented Sampling
Keywords:Hallucination, LLM, SelfCheckGPT, Data Augmentation, Japanese
Inaccurate responses, termed hallucinations, pose challenges in various Large Language Model (LLM) applications. Although a sampling-based method called SelfCheckGPT has been devised to detect hallucinations by using the model's input-output interface without external knowledge, the method requires an increase in the temperature parameter, which cannot be controlled in some LLM services, including ChatGPT. In LLM services designed for accurate responses, the temperature parameter is fixed at a low level, which can degrade the performance of SelfCheckGPT. We therefore propose a novel methodology that utilizes data augmentation (adding random strings or back-translation) during sampling to detect hallucinations in Japanese LLMs under the fixed-low-temperature constraint. Our experimental results reveal that the proposed methodology outperforms SelfCheckGPT under the fixed-low-temperature constraint.
Authentication for paper PDF access
A password is required to view paper PDFs. If you are a registered participant, please log on the site from Participant Log In.
You could view the PDF with entering the PDF viewing password bellow.