dc.contributor.author |
Mohsan, Mashood Mohammad |
|
dc.date.accessioned |
2023-07-24T07:29:38Z |
|
dc.date.available |
2023-07-24T07:29:38Z |
|
dc.date.issued |
2022 |
|
dc.identifier.other |
328023 |
|
dc.identifier.uri |
http://10.250.8.41:8080/xmlui/handle/123456789/34955 |
|
dc.description |
Supervisor: Prof. Dr. Muhammad Usman Akram Co-Supervisor: Dr. Sajid Gul Khawaja |
en_US |
dc.description.abstract |
Recent advancements in transformers exploited computer vision problems which
results in state-of-the-art models. Transformer-based models in various sequence
prediction tasks such as language translation, sentiment classification, and caption
generation have shown remarkable performance. Auto report generation scenarios in
medical imaging through caption generation models is one of the applied scenarios for
language models and have strong social impact. In these models, convolution neural
networks have been used as encoder to gain spatial information and recurrent neural
networks are used as decoder to generate caption or medical report. However, using
transformer architecture as encoder and decoder in caption or report writing task
is still unexplored. In this research, we explored the effect of losing spatial biasness
information in encoder by using pre-trained vanilla image transformer architecture
and combine it with different pre-trained language transformers as decoder. In
order to evaluate the proposed methodology, the Indiana University Chest X-Rays
dataset is used where ablation study is also conducted with respect to different
evaluations. The comparative analysis shows that the proposed methodology has
represented remarkable performance when compared with existing techniques in
terms of different performance parameters. |
en_US |
dc.language.iso |
en |
en_US |
dc.publisher |
College of Electrical & Mechanical Engineering (CEME), NUST |
en_US |
dc.title |
Radiology Report Generation for Chest X-ray Images using Transformers |
en_US |
dc.type |
Thesis |
en_US |