Whitened gradient descent, a new updating method for optimizers in deep neural networks

Message:
Article Type:
Research/Original Article (دارای رتبه معتبر)
Abstract:

Optimizers are vital components of deep neural networks that perform weight updates. This paper introduces a new updating method for optimizers based on gradient descent, called whitened gradient descent (WGD). This method is easy to implement and can be used in every optimizer based on the gradient descent algorithm. It does not increase the training time of the network significantly. This method smooths the training curve and improves classification metrics. To evaluate the proposed algorithm, we performed 48 different tests on two datasets, Cifar100 and Animals-10, using three network structures, including densenet121, resnet18, and resnet50. The experiments show that using the WGD method in gradient descent based optimizers, improves the classification results significantly. For example, integrating WGD in RAdam optimizer increased the accuracy of DenseNet from 87.69% to 90.02% on the Animals-10 dataset.

Language:
English
Published:
Journal of Artificial Intelligence and Data Mining, Volume:10 Issue: 4, Autumn 2022
Pages:
467 to 477
https://www.magiran.com/p2519428  
سامانه نویسندگان
  • Gholamalinejad، Hossein
    Author (1)
    Gholamalinejad, Hossein
    Assistant Professor Computer Engineering,
  • Khosravi، Hossein
    Corresponding Author (2)
    Khosravi, Hossein
    Associate Professor Electrical Engineering and Robotics, Shahrood University of Technology, Shahrud, Iran
اطلاعات نویسنده(گان) توسط ایشان ثبت و تکمیل شده‌است. برای مشاهده مشخصات و فهرست همه مطالب، صفحه رزومه را ببینید.
مقالات دیگری از این نویسنده (گان)