Show simple item record

dc.contributor.authorKong, Lingping
dc.contributor.authorOjha, Varun
dc.contributor.authorGao, Ruobin
dc.contributor.authorSuganthan, Ponnuthurai Nagaratnam
dc.contributor.authorSnášel, Václav
dc.date.accessioned2024-01-29T11:15:04Z
dc.date.available2024-01-29T11:15:04Z
dc.date.issued2023
dc.identifier.citationInformation Sciences. 2023, vol. 642, art. no. 119108.cs
dc.identifier.issn0020-0255
dc.identifier.issn1872-6291
dc.identifier.urihttp://hdl.handle.net/10084/151976
dc.description.abstractTransformer architectures have been applied to graph-specific data such as protein structure and shopper lists, and they perform accurately on graph/node classification and prediction tasks. Researchers have proved that the attention matrix in Transformers has low-rank properties, and the self-attention plays a scoring role in the aggregation function of the Transformers. However, it can not solve the issues such as heterophily and over-smoothing. The low-rank properties and the limitations of Transformers inspire this work to propose a Global Representation (GR) based attention mechanism to alleviate the two heterophily and over-smoothing issues. First, this GR based model integrates geometric information of the nodes of interest that conveys the structural properties of the graph. Unlike a typical Transformer where a node feature forms a Key, we propose to use GR to construct the Key, which discovers the relation between the nodes and the structural representation of the graph. Next, we present various compositions of GR emanating from nodes of interest and 𝛼-hop neighbors. Then, we explore this attention property with an extensive experimental test to assess the performance and the possible direction of improvements for future works. Additionally, we provide mathematical proof showing the efficient feature update in our proposed method. Finally, we verify and validate the performance of the model on eight benchmark datasets that show the effectiveness of the proposed method.cs
dc.language.isoencs
dc.publisherElseviercs
dc.relation.ispartofseriesInformation Sciencescs
dc.relation.urihttps://doi.org/10.1016/j.ins.2023.119108cs
dc.rights© 2023 The Authors. Published by Elsevier Inc.cs
dc.rights.urihttp://creativecommons.org/licenses/by/4.0/cs
dc.subjectgraph transformercs
dc.subjectgraph representationcs
dc.subjectlow-rank attentioncs
dc.subjectglobal representation vectorcs
dc.titleLow-rank and global-representation-key-based attention for graph transformercs
dc.typearticlecs
dc.identifier.doi10.1016/j.ins.2023.119108
dc.rights.accessopenAccesscs
dc.type.versionpublishedVersioncs
dc.type.statusPeer-reviewedcs
dc.description.sourceWeb of Sciencecs
dc.description.volume642cs
dc.description.firstpageart. no. 119108cs
dc.identifier.wos000998393300001


Files in this item

This item appears in the following Collection(s)

Show simple item record

© 2023 The Authors. Published by Elsevier Inc.
Except where otherwise noted, this item's license is described as © 2023 The Authors. Published by Elsevier Inc.