dc.description.abstract |
Text summarization involves distilling essential sentences from a large collection of
articles or documents and condensing them into a shorter rendition. This process
utilizes a variety of techniques, including statistical, graphical, and deep learningbased methods. Nowadays, these techniques are employed in the field to create
summaries. However, existing approaches face several challenges, such as accurately
identifying crucial information, handling diverse document types (like news articles,
research papers, and online reviews), and crafting coherent, grammatically accurate summaries. This thesis presents an automated extractive text summarization
framework that aims to address the existing challenges in text summarization. This
framework seeks to reduce time, costs, and effort while also converting the summarized content into Braille language. Multiple summarization models are trained to
evaluate the performance of BERT and its variants in the summarization task. Additionally, a BERT-based architecture for Braille conversion is proposed to translate
machine-generated English summaries into Braille. The proposed model generates
summarized text and evaluates its performance using metrics like precision, recall,
and F-score. Among the different BERT variants, the SqueezeBERT model, which
is designed specifically for text summarization, maintains 98% of the original BERT
model’s performance while utilizing 49% fewer trainable parameters. SqueezeBERT
emerges as a promising choice for training a summarizer that is nearly half the
size of the original model, with only minimal reductions in summarization performance. Upon assessment at the 30,000-step mark, the models RoBERTa Small, and
SqueezeBERT exhibit better R1, R2, and RL values compared to other models. |
en_US |