Show simple item record

dc.contributor.authorAnand, Mayank
dc.date.accessioned2023-08-18T18:16:20Z
dc.date.available2023-08-18T18:16:20Z
dc.date.issued2023-08-18
dc.identifier.urihttp://hdl.handle.net/10222/82804
dc.description.abstractIncorporating hierarchical structures for various Natural Language Processing (NLP) tasks, which involves training the model with syntactic information of constituency trees, has been shown to be very effective. Constituency trees in the simplest form are graph representations of sentences that capture and illustrate syntactic hierarchical structure of a sentence by showing how words are grouped into constituents. However, the majority of research in NLP using Deep Learning to incorporate structural information has been conducted on recurrent models, which are effective but operate sequentially. To the best of our knowledge, no research has been done on attention-based models for the reading comprehension task. In this work, we aim to include syntactic information of constituency trees in the model QAnet which is based on self-attention and specifically designed for Machine Reading Comprehension task. The proposed solution involves the use of “Hierarchical Accumulation” to encode constituency trees in self-attention in parallel time complexity. Our model, QATnet, achieved competitive results compared to the baseline QAnet model. Furthermore, we demonstrated by analyzing context-question pair examples that using a hierarchical structure model exhibited a remarkable ability to retain contextual information over longer distances and enhanced attention towards punctuation and other grammatical intricacies.en_US
dc.language.isoenen_US
dc.subjectnatural language processingen_US
dc.subjectlanguage modelingen_US
dc.subjectlarge language models(LLMs)en_US
dc.subjectDeep Learningen_US
dc.subjectconstituency treesen_US
dc.subjectQuestion Answeringen_US
dc.subjectSQuADen_US
dc.subjectSQuAD2.0en_US
dc.subjecttransformersen_US
dc.subjectmulti-head attentionen_US
dc.subjectself-attentionen_US
dc.subjectMachine Reading Comprehensionen_US
dc.subjectcomputer scienceen_US
dc.titleStructural Embedding of Constituency Trees in the Attention-based Model for Machine Comprehensionen_US
dc.date.defence2023-08-01
dc.contributor.departmentFaculty of Computer Scienceen_US
dc.contributor.degreeMaster of Computer Scienceen_US
dc.contributor.external-examinerN/Aen_US
dc.contributor.graduate-coordinatorDr. Michael McAllisteren_US
dc.contributor.thesis-readerDr. Evangelos Miliosen_US
dc.contributor.thesis-readerDr. Ga Wuen_US
dc.contributor.thesis-supervisorDr. Vlado Keseljen_US
dc.contributor.ethics-approvalNot Applicableen_US
dc.contributor.manuscriptsNot Applicableen_US
dc.contributor.copyright-releaseNot Applicableen_US
 Find Full text

Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record