Show simple item record

dc.contributor.authorYu, Tong
dc.contributor.authorKhalitov, Ruslan
dc.contributor.authorCheng, Lei
dc.contributor.authorYang, Zhirong
dc.date.accessioned2023-02-07T10:05:26Z
dc.date.available2023-02-07T10:05:26Z
dc.date.created2022-11-16T16:38:17Z
dc.date.issued2022
dc.identifier.issn2575-7075
dc.identifier.urihttps://hdl.handle.net/11250/3048820
dc.description.abstractSelf-Attention is a widely used building block in neural modeling to mix long-range data elements. Most self-attention neural networks employ pairwise dot-products to specify the attention coefficients. However, these methods require O(N 2 ) computing cost for sequence length N. Even though some approximation methods have been introduced to relieve the quadratic cost, the performance of the dot-product approach is still bottlenecked by the lowrank constraint in the attention matrix factorization. In this paper, we propose a novel scalable and effective mixing building block called Paramixer. Our method factorizes the interaction matrix into several sparse matrices, where we parameterize the non-zero entries by MLPs with the data elements as input. The overall computing cost of the new building block is as low as O(N log N). Moreover, all factorizing matrices in Paramixer are full-rank, so it does not suffer from the low-rank bottleneck. We have tested the new method on both synthetic and various real-world long sequential data sets and compared it with several state-of-the-art attention networks. The experimental results show that Paramixer has better performance in most learning tasks.en_US
dc.language.isoengen_US
dc.publisherIEEEen_US
dc.titleParamixer: Parameterizing Mixing Links in Sparse Factors Works Better than Dot-Product Self-Attentionen_US
dc.title.alternativeParamixer: Parameterizing Mixing Links in Sparse Factors Works Better than Dot-Product Self-Attentionen_US
dc.typeJournal articleen_US
dc.description.versionacceptedVersionen_US
dc.rights.holder© IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.en_US
dc.source.journal2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)en_US
dc.identifier.doi10.1109/CVPR52688.2022.00077
dc.identifier.cristin2075105
cristin.ispublishedtrue
cristin.fulltextpostprint


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record