NUST Institutional Repository

Natural Image Synthesis Using Scene Graphs

Show simple item record

dc.contributor.author Malik, Mahgul
dc.date.accessioned 2022-07-29T10:36:40Z
dc.date.available 2022-07-29T10:36:40Z
dc.date.issued 2022
dc.identifier.uri http://10.250.8.41:8080/xmlui/handle/123456789/30017
dc.description.abstract Complex scene graphs are still battling to create fewer artefacts and distinct instances of an object when employed for image generation tasks. A closer look at the generated images found in the literature reveals that the appearance of each object is often distorted and devoid of the essential distinguishing characteristics connected with the object class. We believe; it is due to the loss of sensitive information during the process where each node and edge of the graph transform into their respective encodings. We aim to address this issue by employing an autoencoder-based graph convolutional network to reduce information loss. Our model encodes topological structure presenting relationship and object content of the scene graph into meaningful compact representation. Our approach significantly denoise the learned embeddings and promises to carry more relevant and discriminative information than embeddings obtained directly from a scene graph via graph convolutional networks. These newly learnt embeddings are then transmitted to the appropriate networks to generate bounding boxes and segmentation masks, which are subsequently utilised in the scene layout generating process. The cascaded refinement network uses scene layout to synthesize an image. Using COCO-Stuff, we evaluate the ability of our system in terms of the inception score and human subjective judgement. en_US
dc.description.sponsorship Dr. Muhammad Moazam Fraz en_US
dc.language.iso en en_US
dc.publisher SEECS, National University of Science and Technology, Islamabad. en_US
dc.subject Complex scene graphs are still battling to create fewer artefacts and distinct instances of an object when employed for image generation tasks. A closer look at the generated images found in the literature reveals that the appearance of each object is often distorted and devoid of the essential distinguishing characteristics connected with the object class. We believe; it is due to the loss of sensitive information during the process where each node and edge of the graph transform into their respective encodings. We aim to address this issue by employing an autoencoder-based graph convolutional network to reduce information loss. Our model encodes topological structure presenting relationship and object content of the scene graph into meaningful compact representation. Our approach significantly denoise the learned embeddings and promises to carry more relevant and discriminative information than embeddings obtained directly from a scene graph via graph convolutional networks. These newly learnt embeddings are then transmitted to the appropriate networks to generate bounding boxes and segmentation masks, which are subsequently utilised in the scene layout generating process. The cascaded refinement network uses scene layout to synthesize an image. Using COCO-Stuff, we evaluate the ability of our system in terms of the inception score and human subjective judgement. en_US
dc.title Natural Image Synthesis Using Scene Graphs en_US
dc.type Thesis en_US


Files in this item

This item appears in the following Collection(s)

  • MS [375]

Show simple item record

Search DSpace


Advanced Search

Browse

My Account