Addressing encoder-only transformer limitations with graph neural networks for text classification

buir.advisorKoç, Aykut
dc.contributor.authorAras, Arda Can
dc.date.accessioned2025-01-24T13:56:37Z
dc.date.available2025-01-24T13:56:37Z
dc.date.copyright2025-01
dc.date.issued2025-01
dc.date.submitted2025-01-23
dc.descriptionCataloged from PDF version of article.
dc.descriptionIncludes bibliographical references (leaves 70-86).
dc.description.abstractRecent advancements in NLP have been primarily driven by transformer-based models, which capture contextual information within sequences, revolutionizing tasks such as text classification and natural language understanding. In parallel, GNNs have emerged as powerful tools for modeling structured data, leveraging graph representations to capture global relationships across entities. However, significant challenges persist at the intersection of these fields, limiting the efficacy and scalability of existing models. These challenges include the inability to seamlessly integrate contextual and structural information, computational inefficiencies associated with static graph construction and transductive learning, and the underperformance of models in low-labeled data scenarios. This thesis explores and addresses these challenges by developing novel methodologies that unify transformers and GNNs, leveraging their complementary strengths. The first contribution, GRTE, introduces an architecture that combines pre-trained transformer models with heterogeneous and homogeneous graph representations to enhance text classification in both inductive and transductive settings. Compared to state-of-the-art models, GRTE achieves significant computational efficiency, reducing training overhead by up to 100 times. The second contribution, Text-RGNN, proposes a relational modeling framework for heterogeneous text graphs, enabling the nuanced representation of diverse interactions between nodes and demonstrating substantial accuracy improvements of up to 10.61% over existing models, particularly in low-labeled data settings. Finally, the third contribution, VISPool, introduces a scalable architecture that dynamically constructs vector visibility graphs from transformer outputs, enabling seamless integration of graph-based reasoning into transformer pipelines while improving performance on NLP benchmarks such as GLUE, with performance improvements of up to 13% in specific tasks. Through comprehensive experimentation and benchmarking against state-of-the-art models, this thesis establishes the efficacy of these proposed methodologies. The results demonstrate the potential for improved performance, scalability, and the ability to address long-standing challenges in NLP and GNN integration. These contributions lay a robust foundation for future research and applications at the intersection of graph-based and transformer-based approaches, advancing the state of the art in text representation and classification.
dc.description.provenanceSubmitted by Betül Özen (ozen@bilkent.edu.tr) on 2025-01-24T13:56:37Z No. of bitstreams: 1 B149012.pdf: 3221799 bytes, checksum: f33596e4820c082807b325329e057c6c (MD5)en
dc.description.provenanceMade available in DSpace on 2025-01-24T13:56:37Z (GMT). No. of bitstreams: 1 B149012.pdf: 3221799 bytes, checksum: f33596e4820c082807b325329e057c6c (MD5) Previous issue date: 2025-01en
dc.description.statementofresponsibilityby Arda Can Aras
dc.format.extentxvi, 87 leaves : illustrations, charts ; 30 cm.
dc.identifier.itemidB149012
dc.identifier.urihttps://hdl.handle.net/11693/115964
dc.language.isoEnglish
dc.rightsinfo:eu-repo/semantics/openAccess
dc.subjectNatural language processing (NLP)
dc.subjectTransformer
dc.subjectGraph neural networks (GNNs)
dc.subjectText classification
dc.titleAddressing encoder-only transformer limitations with graph neural networks for text classification
dc.title.alternativeYalnızca kodlayıcı kullanan dönüştürücülerin metin sınıflandırmasındaki sınırlamalarının çizge sınır ağları ile aşılması
dc.typeThesis
thesis.degree.disciplineElectrical and Electronic Engineering
thesis.degree.grantorBilkent University
thesis.degree.levelMaster's
thesis.degree.nameMS (Master of Science)

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
B149012.pdf
Size:
3.07 MB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.71 KB
Format:
Item-specific license agreed upon to submission
Description: