Show simple item record

dc.contributor.advisorOwren, Brynjulf
dc.contributor.authorHellan, Ottar Passano
dc.date.accessioned2022-11-25T18:21:46Z
dc.date.available2022-11-25T18:21:46Z
dc.date.issued2022
dc.identifierno.ntnu:inspera:104646180:36318324
dc.identifier.urihttps://hdl.handle.net/11250/3034256
dc.description.abstractSentrale konsepter og strukturer i riemannsk optimering presenteres og diskuteres for å gi en uavhengig behandling av Riemannian gradient descent-metoden (RGD). Egenskaper ved RGD diskuteres og sammenliknes med de av euklidsk gradient descent, som RGD er en generalisering av. Mulige anvendelser av riemannsk optimering og RGD i feltet dyp læring diskuteres, samt nødvendige hensyn man må ta i implementering av disse. Enkle beregningseksperimenter for demonstrasjon gjøres med RGD for lav rang matrise-mangfoldigheten og den ortogonale gruppen til et CIFAR-10 bildeklassifiseringsproblem og et recurrent neural network problem med lange tidsavhengigheter.
dc.description.abstractCentral concepts and structures of Riemannian optimization are presented and discussed to give a self-contained treatment of the Riemannian gradient descent method (RGD). Properties of RGD are discussed and compared with those of euclidean gradient descent, which RGD is a generalization of. Possible applications of Riemannian optimization and RGD in the field of deep learning are discussed along with considerations one must make in implementations of such methods. Proof-of-concept computational experiments are made using RGD for the fixed-rank matrix manifold and the orthogonal group on CIFAR-10 image classification and a long time-dependence recurrent neural network problem.
dc.languageeng
dc.publisherNTNU
dc.titleRiemannian Optimization for Deep Learning
dc.typeMaster thesis


Files in this item

Thumbnail
Thumbnail

This item appears in the following Collection(s)

Show simple item record