AUB ScholarWorks

Second Order Trust Region Optimization Methods for Training Neural Networks: Beyond Inexact Newton

Show simple item record

dc.contributor.advisor Turkiyyah, George
dc.contributor.author Lomer, Kyle
dc.date.accessioned 2020-09-22T14:26:59Z
dc.date.available 2020-09-22T14:26:59Z
dc.date.issued 9/22/2020
dc.identifier.uri http://hdl.handle.net/10938/21953
dc.description Dr Shady Elbassuoni Dr Izzat El Hajj
dc.description.abstract Second order optimization methods have always been less widely used for training neural networks than first order methods such as Stochastic Gradient Descent. This is mainly due to the complexity and high costs in terms of both processor and memory resources of second order methods. In recent years more work has been done to adapt these methods to make them more suitable for training neural networks. In this paper we demonstrate how trust region methods can be used to improve the convergence and cost-effectiveness of second order optimization. This is achieved by only using cheap first order information when it is an appropriate approximation for the expensive second order information, based on the relative size of the trust region. We also present techniques to automatically tune the hyperparameters these methods introduce; including a novel approach to adaptive regularization. These methods are demonstrated on autoencoders and image classifiers in comparison to first order methods.
dc.language.iso en_US
dc.subject Computer Science
dc.subject Optimization
dc.subject Numerical Methods
dc.subject Machine Learning
dc.subject Neural Networks
dc.title Second Order Trust Region Optimization Methods for Training Neural Networks: Beyond Inexact Newton
dc.type Thesis
dc.contributor.department Department of Computer Science
dc.contributor.faculty Faculty of Arts and Sciences
dc.contributor.institution American University of Beirut


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search AUB ScholarWorks


Browse

My Account