JSAI2025

Presentation information

Poster Session

Poster session » Poster Session

[3Win5] Poster session 3

Thu. May 29, 2025 3:30 PM - 5:30 PM Room W (Event hall D-E)

[3Win5-02] Improving Data-to-Text Generation with Large Language Models through Numerical Data Back-Translation

〇Masahiro Ebe1, Atsushi Aoyama2 (1.Keio Research Institute at SFC, 2.Faculty of Environment and Information Studies, Keio University)

Keywords:Large Language Model, Reinforcement Learning, Numerical Data-to-Text Generation, Back-Translation, Automatic Evaluation

We introduce a reinforcement learning approach that utilizes back-translation to numerical data for Data-to-Text generation with large language models (LLMs). Numerical data can have multiple possible interpretations, making it difficult to predefine their meaning and the key points to be explained before conducting an analysis. In this study, we focus on information recoverability in explaining numerical data and propose a reinforcement learning approach based on Proximal Policy Optimization (PPO). This approach does not require prior reference definitions and uses the error in back-translation to numerical data as a reward signal. Our experiments demonstrate that the proposed method significantly improves explanatory performance after training. Furthermore, the explanatory performance achieved with our method is significantly higher than that obtained using Direct Policy Optimization (DPO), a training method that does not require the design of a reward function. These results highlight the effectiveness of using back-translation error as a reward for enhancing explanatory performance.

Authentication for paper PDF access
A password is required to view paper PDFs. If you are a registered participant, please log on the site from Participant Log In.
You could view the PDF with entering the PDF viewing password bellow.

Password