Отрывок: A simple model of neural network consists of one hidden layer with different number of neurons nneuron = 4 (see Fig. 1 a)) and nneuron = 1 (see Fig. 1 b)). Fig. 1 a) shows the advantage of Adagrad with the adaptive learning rate over other methods. Fig. 1 b), in turn, reveals the Adagrad’s...
Полная запись метаданных
Поле DC Значение Язык
dc.contributor.authorKulikovskikh, I.-
dc.contributor.authorProkhorov, S.-
dc.contributor.authorLegović, T.-
dc.contributor.authorŠmuc, T.-
dc.date.accessioned2019-04-22 11:10:15-
dc.date.available2019-04-22 11:10:15-
dc.date.issued2019-05-
dc.identifierDspace\SGAU\20190421\75664ru
dc.identifier.citationKulikovskikh I. Growing descent of stochastic gradient with the generalized logistic map / Kulikovskikh I., Prokhorov S., Legović T., Šmuc T. // Сборник трудов ИТНТ-2019 [Текст]: V междунар. конф. и молодеж. шк. "Информ. технологии и нанотехнологии": 21-24 мая: в 4 т. / Самар. нац.-исслед. ун-т им. С. П. Королева (Самар. ун-т), Ин-т систем. обраб. изобр. РАН-фил. ФНИЦ "Кристаллография и фотоника" РАН; [под ред. В.А. Фурсова]. - Самара: Новая техника, 2019. – Т. 4: Науки о данных. - 2019. - С. 338-344.ru
dc.identifier.urihttp://repo.ssau.ru/handle/Informacionnye-tehnologii-i-nanotehnologii/Growing-descent-of-stochastic-gradient-with-the-generalized-logistic-map-75664-
dc.description.abstractThe paper considers the problem of accelerating the convergence of stochastic gradient descent (SGD) in an automatic way. Previous research puts forward such algorithms as Adagrad, Adadelta, RMSprop, Adam and etc. to adapt both the updates and learning rates to the slope of a loss function. However, these adaptive methods do not share the same regret bound as the gradient descent method. Adagrad provably achieves the optimal regret bound on the assumption of convexity but accumulates the squared gradients in the denominator that dramatically shrinks the learning rate. This research is aimed at introducing a generalized logistic map directly into the SGD method in order to automatically set its parameters to the slope of the logistic loss function. The optimizer based on the logistic map may be considered as a meta-learner that learns how to tune both the learning rate and gradient updates with respect to the rate of population growth. The present study yields the “growing” descent method and a series of computational experiments to point out the benefits of injecting the logistic map.ru
dc.description.sponsorshipThis work was supported by the Russian Federation President grant No. MK-6218.2018.9 and the Ministry of Education and Science of the Russian Federation grant No. 074-U01. The research in Section 3 was partly supported by RFBR (project No. 18-37-00219). The authors acknowledge the support by the Centre of Excellence project “DATACROSS”, co-financed by the Croatian Government and the European Union through the European Regional Development Fund – the Competitiveness and Cohesion Operational Programme (KK.01.1.1.01.0009).ru
dc.language.isoen_USru
dc.publisherНовая техникаru
dc.titleGrowing descent of stochastic gradient with the generalized logistic mapru
dc.typeArticleru
dc.textpartA simple model of neural network consists of one hidden layer with different number of neurons nneuron = 4 (see Fig. 1 a)) and nneuron = 1 (see Fig. 1 b)). Fig. 1 a) shows the advantage of Adagrad with the adaptive learning rate over other methods. Fig. 1 b), in turn, reveals the Adagrad’s...-
Располагается в коллекциях: Информационные технологии и нанотехнологии

Файлы этого ресурса:
Файл Описание Размер Формат  
paper42.pdfОсновная статья819.57 kBAdobe PDFПросмотреть/Открыть



Все ресурсы в архиве электронных ресурсов защищены авторским правом, все права сохранены.