〇Asei Ito1, Kota Takaguchi2
(1. The University of Tokyo, 2. Chiba University)
Keywords:Large Language Model, China, Censorship, Fine tuning
Large language models (LLMs) developed in China are required to "adhere to the core socialist values." Previous studies have constructed sensitive questions to examine this issue. This study aims to further elucidate the details of censorship by first introducing the Basic Requirements for the Security of Generative Artificial Intelligence Services, published in February 2024. Next, we evaluated LLMs using a benchmark questions created by the China Electronics Standardization Institute and Fudan University. The models analyzed included major Chinese open-source LLMs, derivative models fine-tuned for the Japanese market, and Western LLMs.The analysis revealed evidence of censorship in Chinese models and their derivative versions. The findings suggest that users of these LLMs should be aware of the censorship-based fine-tuning applied to Chinese models and conduct thorough checks to ensure their suitability for specific applications.
Authentication for paper PDF access
A password is required to view paper PDFs. If you are a registered participant, please log on the site from Participant Log In.
You could view the PDF with entering the PDF viewing password bellow.