earticle

논문검색

Poster Session I : Next Generation Computing Applications I

Reversing Attention Mechanisms in Transformers to Improve Object Detection Performance

초록

영어

Recent advancements in object detection increasingly leverage end-to-end transformer architectures. However, many studies in this domain have applied transformer structures, originally designed for natural language processing, directly to object detection models. This direct application can lead to issues such as skipping self-attention in first decoder layer and the prediction of duplicate objects during training. In this study, we propose a novel approach to address these challenges by reversing the attention order in the transformer decoder from the self-cross to a cross-self structure. This modification structurally prevents the initial attention skip and mitigates the issue of predicting the same object multiple times by delaying the implementation of self-attention. Experimental results demonstrate that reversing the attention order in the decoder improves both the training loss and test performance across all stages of the learning process.

목차

Abstract
I. INTRODUCTION
II. METHODOLOGY
A. RT-DETR
B. Reversing Attention Mechanisms
C. Experimental Results
III. CONCLUSION
ACKNOWLEDGMENT
REFERENCES

저자정보

  • Chan-Young Choi School of Computing Gachon University
  • Sung-Yoon Ahn School of Computing Gachon University
  • Abrar Alabdulwahab School of Computing Gachon University
  • Joo-Hee Oh School of Computing Gachon Gachon University
  • Sang-Woong Lee School of Computing Gachon University

참고문헌

자료제공 : 네이버학술정보

    함께 이용한 논문

      0개의 논문이 장바구니에 담겼습니다.