4:30 PM - 4:50 PM
[3P5-OS-17a-04] Improvement of Mini-Batch Size Dependency in Deep Learning for Reduction of Required Machine Resources
Keywords:Neural Network, Normalization, Minibatch Learning
In deep learning, batch normalization, which is commonly used to improve training performance, is recommendedto be used in conjunction with large mini-batch sizes during training on large datasets. However, increasing mini-batch size leads to an increase in required machine resources. Therefore, by reducing this mini-batch size dependencywhen adopting batch normalization and thereby reducing the required machine resources, we aim to alleviate thebarriers to exploring deep learning and promote diversification in its application scenarios. To this end, we proposea method that combines modified batch normalization with weight standardization to achieve training resultssimilar to those obtained with large mini-batch sizes, even when small mini-batch sizes are used. We demonstratethat our proposed method improves the problem of mini-batch size dependency compared to existing methods.
Authentication for paper PDF access
A password is required to view paper PDFs. If you are a registered participant, please log on the site from Participant Log In.
You could view the PDF with entering the PDF viewing password bellow.