Blar i Øvrige samlinger på tidsskrift "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)"
Viser treff 1-1 av 1
-
Paramixer: Parameterizing Mixing Links in Sparse Factors Works Better than Dot-Product Self-Attention
(Journal article, 2022)Self-Attention is a widely used building block in neural modeling to mix long-range data elements. Most self-attention neural networks employ pairwise dot-products to specify the attention coefficients. However, these ...