9:00 AM - 9:20 AM
[4P1-OS-17b-01] Theoretical Analysis for Generalization Performance of Partial Concept Bottleneck Model in Three-layered and Linear Neural Networks
Keywords:Explainable AI, Concept Bottleneck Model, Neural Network, Singular Learning Theory, Bayesian Inference
For human centered future society, artificial intelligence (AI) has been desired in many fields.Since we need responsibility of information systems using AI, in particular, car driving or medical systems,explanation methods have been proposed for machine learning models.Concept Bottleneck Model (CBM) is one of methods to make neural networks explainable.In CBM, concepts which correspond to reasons of outputs are inserted in the last intermediate layer as observed values.It is expected that we can interpret the relation between the output and the concept like linear regression.However, it requires observation for all concepts and decrease generalization performance of neural networks.Partial CBM (PCBM) has been devised for resolving these difficulty.It uses partially-observed concepts.While some numerical experiments demonstrate that the generalization performance of PCBMs is almost as high as that of original neural networks,theoretical behavior of its generalization error has not been yet clarified since PCBM is singular statistical model.In this paper, we reveal the Bayesian generalization error in PCBM whose architecture is three-layered and linear.The result shows that the structure of partially-observed concepts makes decrease the Bayesian generalization error compared with CBM (full-obsereved concepts).
Authentication for paper PDF access
A password is required to view paper PDFs. If you are a registered participant, please log on the site from Participant Log In.
You could view the PDF with entering the PDF viewing password bellow.