Would you like to see your presentation here, made available to a global audience of researchers?
Add your own presentation or have us affordably record your next conference.
Scene recognition (SR) is a fundamental task in computer vision (CV). In recent years, transformer-based methods have achieved remarkable success in scene recognition tasks. Most existing approaches primarily rely on visual features and fail to effectively model the structural relationships within scenes, which are crucial for accurate scene recognition. To this end, we propose TANSR, an innovative method that leverages topological relationships from graphs to guide scene recognition. Specifically, GAMGN generates topology-aware masks from graph representations constructed by GGM and integrates them with patch embeddings by TAG, enabling the transformer's attention mechanism to be aware of topological information. Furthermore, we introduce an innovative attention-driven multimodal fusion strategy that integrates graph-derived topological cues with visual patch embeddings, substantially enhancing the transformer’s capability to capture topological information and improving performance in complex scene recognition tasks. We evaluate the model on the benchmarks MIT-67, Scene-15 and SUN397, where it achieves consistent state-of-the-art (SOTA) performance, including 98.58% accuracy on MIT-67.
