Vis enkel innførsel

dc.contributor.advisorLangseth, Helge
dc.contributor.advisorTidemann, Axel
dc.contributor.advisorBanino, Cyril
dc.contributor.authorSve, Thomas
dc.contributor.authorRemmen, Bjørnar Moe
dc.date.accessioned2017-10-12T14:00:24Z
dc.date.available2017-10-12T14:00:24Z
dc.date.created2017-06-08
dc.date.issued2017
dc.identifierntnudaim:15964
dc.identifier.urihttp://hdl.handle.net/11250/2459946
dc.description.abstractVisual recognition systems are often limited to the object categories previously trained on and thus suffer in their ability to scale. This is in part due to the difficulty of acquiring sufficient labeled images as the number of object categories grows. To solve this, earlier research have presented models that uses other sources, such as text data, to help classify object categories unseen during training. However, most of these models are limited on images with a single label and most images can contain more than one object category, and therefore more than one label. This master's thesis implements a model capable of classifying unseen categories for both single- and multi-labeled images. The architecture consist of several modules: A pre-trained neural network that generates image features for each image, a model trained on text that represents words as vectors, and a neural network that projects the image features to the dimension native to the vector representation of words. On this architecture, we compared two approaches to generate word vectors using GloVe and Word2vec, with different vector dimensions and on spaces containing different numbers of word vectors. The model was adapted to multi-label predictions comparing three approaches for image box generation: YOLOv2, Faster R-CNN and randomly generated boxes. Here each box represents a section of the image cut out and this approach was chosen to fit each label to a one of these boxes. The results showed that increasing the word vector dimension increased the accuracy, with Word2vec outperforming GloVe, and when adding more words to the word vector space the accuracy dropped. In the single-label scenario the model achieves similar results to existing models with similar architecture. While in the multi-label scenario, the model trained on boxes generated by Faster R-CNN and predicted on random generated boxes had highest accuracy, but was not able to outperform comparative alternatives. The architecture gives promising results, but more investigation is needed to answer if the results can be improved further.
dc.languageeng
dc.publisherNTNU
dc.subjectInformatikk, Kunstig intelligens
dc.titleInvestigating Zero-Shot Learning techniques in multi-label scenarios
dc.typeMaster thesis


Tilhørende fil(er)

Thumbnail
Thumbnail
Thumbnail

Denne innførselen finnes i følgende samling(er)

Vis enkel innførsel