Transformer framework has been showing superior performances in visual object tracking for its great strength information aggregation across the template and search image with well-known attention mechanism. Most recent advances focus on exploring mechanism variants better aggregation. We find these schemes are equivalent to or even just a subset of basic self-attention In this paper, we prove ...