Axial Attention 轴向attention

self attention的计算量是二阶的,用axial-attention可以较少计算量,计算效率高一些

axial-attention做法就是先在竖直方向进行self-attention,然后再在水平方向进行self-attention,以这种形式降低计算复杂度

Guess you like

Origin blog.csdn.net/hxxjxw/article/details/121445561