VISPool: enhancing transformer encoders with vector visibility graph neural networks

buir.contributor.authorAlikaşifoğlu, Tuna
buir.contributor.authorAras, Arda Can
buir.contributor.authorKoç, Aykut
buir.contributor.orcidAlikaşifoğlu, Tuna|0000-0001-8030-8088
buir.contributor.orcidKoç, Aykut|0000-0002-6348-2663
dc.citation.epage2556
dc.citation.spage2547
dc.contributor.authorAlikaşifoğlu, Tuna
dc.contributor.authorAras, Arda Can
dc.contributor.authorKoç, Aykut
dc.coverage.spatialHybrid, Bangkok, Thailand
dc.date.accessioned2025-02-28T07:01:28Z
dc.date.available2025-02-28T07:01:28Z
dc.date.issued2024-08-16
dc.departmentDepartment of Electrical and Electronics Engineering
dc.departmentNational Magnetic Resonance Research Center (UMRAM)
dc.descriptionConference Name: Findings of the 62nd Annual Meeting of the Association for Computational Linguistics, ACL 2024
dc.descriptionDate of Conference: 11 August 2024 - 16 August 2024
dc.description.abstractThe emergence of transformers has revolutionized natural language processing (NLP), as evidenced in various NLP tasks. While graph neural networks (GNNs) show recent promise in NLP, they are not standalone replacements for transformers. Rather, recent research explores combining transformers and GNNs. Existing GNN-based approaches rely on static graph construction methods requiring excessive text processing, and most of them are not scalable with the increasing document and word counts. We address these limitations by proposing a novel dynamic graph construction method for text documents based on vector visibility graphs (VVGs) generated from transformer output. Then, we introduce visibility pooler (VISPool), a scalable model architecture that seamlessly integrates VVG convolutional networks into transformer pipelines. We evaluate the proposed model on the General Language Understanding Evaluation (GLUE) benchmark datasets. VISPool outperforms the baselines with less trainable parameters, demonstrating the viability of the visibility-based graph construction method for enhancing transformers with GNNs. © 2024 Association for Computational Linguistics.
dc.identifier.doi10.18653/v1/2024.findings-acl.149
dc.identifier.isbn979-889176099-8
dc.identifier.issn0736587X
dc.identifier.urihttps://hdl.handle.net/11693/116965
dc.language.isoEnglish
dc.publisherAssociation for Computational Linguistics
dc.relation.isversionofhttps://dx.doi.org/10.18653/v1/2024.findings-acl.149
dc.source.titleAssociation for Computational Linguistics. Annual Meeting. Conference Proceedings
dc.subjectHUMANITIES and RELIGION::Languages and linguistics::Linguistic subjects::Computational linguistics
dc.subjectConvolutional neural networks
dc.subjectDistribution transformers
dc.subjectGluing
dc.subjectModeling languages
dc.subjectNatural language processing systems
dc.subjectWord processing
dc.titleVISPool: enhancing transformer encoders with vector visibility graph neural networks
dc.typeConference Paper

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
VISPool_Enhancing_Transformer_Encoders_with_Vector_Visibility_Graph_Neural_Networks.pdf
Size:
614.27 KB
Format:
Adobe Portable Document Format