JSAI2025

Presentation information

Organized Session

Organized Session » OS-42

[3F4-OS-42a] OS-42

Thu. May 29, 2025 1:40 PM - 3:20 PM Room F (Room 1001)

オーガナイザ:金子 正弘(MBZUAI),小島 武(東京大学),磯沼 大(The University of Edinburgh/東京大学),丹羽 彩奈(MBZUAI),大葉 大輔(ELYZA/東京科学大学),村上 明子(AIセーフティーインスティチュート),関根 聡(情報学研究所),内山 将夫(情報通信研究機構),Danushka Bollegala(The University of Liverpool/Amazon)

1:40 PM - 2:00 PM

[3F4-OS-42a-01] Censorship-based Fine-tuning in Chinese Large Language Models

〇Asei Ito1, Kota Takaguchi2 (1. The University of Tokyo, 2. Chiba University)

Keywords:Large Language Model, China, Censorship, Fine tuning

Large language models (LLMs) developed in China are required to "adhere to the core socialist values." Previous studies have constructed sensitive questions to examine this issue. This study aims to further elucidate the details of censorship by first introducing the Basic Requirements for the Security of Generative Artificial Intelligence Services, published in February 2024. Next, we evaluated LLMs using a benchmark questions created by the China Electronics Standardization Institute and Fudan University. The models analyzed included major Chinese open-source LLMs, derivative models fine-tuned for the Japanese market, and Western LLMs.The analysis revealed evidence of censorship in Chinese models and their derivative versions. The findings suggest that users of these LLMs should be aware of the censorship-based fine-tuning applied to Chinese models and conduct thorough checks to ensure their suitability for specific applications.

Authentication for paper PDF access
A password is required to view paper PDFs. If you are a registered participant, please log on the site from Participant Log In.
You could view the PDF with entering the PDF viewing password bellow.

Password