JSAI2022

Presentation information

General Session

General Session » GS-5 Language media processing

[3C3-GS-6] Language media processing: generation

Thu. Jun 16, 2022 1:30 PM - 2:50 PM Room C (Room C-2)

座長:赤間 怜奈(東北大学)[現地]

2:10 PM - 2:30 PM

[3C3-GS-6-03] Selective training with generated samples from generative language models for generalized zero-shot text classification

〇Sohta Kannan1, Taro Yano2, Kunihiro Takeoka2, Masafumi Oyamada2, Takeshi Okadome1 (1. Kwansei Gakuin University, 2. NEC Corporation)

Keywords:text classification, zero-shot classification, data augmentation, text generation

Generalized zero-shot text classification (GZSTC) is a task to classify text into a set of classes including unseen classes, which has no teacher data.
GZSTC is widely applied to such as news and product classification.
One of the existing approaches to zero-shot text classification is to use a language model to generate pseudo samples of the unseen class and use them as teacher data.
However, these existing approaches use language models that have been pre-trained by data from a wide range of domains, and thus cannot generate only samples that correspond to the target domain, which adversely affects the training of the classifier.
In this paper, we propose a GZSTC method that improves classification performance.
This method removes the out-of-domain samples generated by the language model and creates a data set with only the samples corresponding to the target domain.
Experiments on real data show the improvement of the classification performance of the proposed method against the baseline.

Authentication for paper PDF access

A password is required to view paper PDFs. If you are a registered participant, please log on the site from Participant Log In.
You could view the PDF with entering the PDF viewing password bellow.

Password