摘要: Do You Even Need Attention? A Stack of Feed-Forward Layers Does Surprisingly Well on ImageNet Abstract 视觉transformers在图像分类和其他视觉任务上的强大性能通常归因于其multi-hea 阅读全文
posted @ 2021-06-17 10:39 慢行厚积 阅读(296) 评论(0) 推荐(0) 编辑
摘要: Beyond Self-attention: External Attention using Two Linear Layers for Visual Tasks Abstract 注意力机制,尤其是自注意力(self-attention),在视觉任务的深度特征表征中起着越来越重要的作用。自注意力 阅读全文
posted @ 2021-06-17 10:37 慢行厚积 阅读(1377) 评论(0) 推荐(0) 编辑