To Analysis and Compare 21 Weight Constraints in Stochastic Gradient Descent Algorithm Using Kernel Method
Imposing a constraint on the gradient descend algorithm for the purpose of neural network training with limited weights has several applications such as network transparency, reducing network volume in terms of storage, increasing the level of generalizability. In addition, it speeds up convergence and finds a more accurate answer. In this paper, using the kernel trick as a method for imposing various constraints on the training algorithm, 21 different constraints have been compared those 16 constraints of which are novel and inspired by the existing uncertainty in biological neural networks. There have been no data augmentation or regularization techniques are used to clearly show the effect of each constraint functions. To solve the classification problem of MNIST, CIFAR-10 and CIFAR-100 datasets for each constraint function, 63 experiments have been done. The results show constraint functions have different impact on solving each dataset and specially our biologically inspired constraints may lead to train a more accurate neural network than the other constraints.
- حق عضویت دریافتی صرف حمایت از نشریات عضو و نگهداری، تکمیل و توسعه مگیران میشود.
- پرداخت حق اشتراک و دانلود مقالات اجازه بازنشر آن در سایر رسانههای چاپی و دیجیتال را به کاربر نمیدهد.