[口頭報告]Domain adaptive semantic segmentation based on prototype-guided and adaptive feature fusion
Domain adaptive semantic segmentation based on prototype-guided and adaptive feature fusion
編號:23
稿件編號:224 訪問權限:僅限參會人
更新:2024-05-15 17:47:13
瀏覽:145次
口頭報告

報告開始:2024年05月30日 18:00 (Asia/Shanghai)
報告時間:20min
所在會議:[S4] Intelligent Equipment Technology ? [S4-4] Afternoon of May 30th-4
暫無文件
摘要
Unsupervised domain adaptation technology is key to reducing the need for data labeling in computer vision tasks and implementing intelligent perception in equipment. Faced with the dispersion of feature distribution and class imbalance in real scenes (i.e., the target domain), such as blurry class boundaries and scarce samples, this paper proposes a Prototypes-Guided Adaptive Feature Fusion Model. It incorporates a Prototype-Guided Dual Attention Network that blends spatial and channel attention features to enhance class compactness. Moreover, an adaptive feature fusion module is introduced to flexibly adjust the importance of each feature, enabling the model to capture more class-discriminative features across different spatial locations and channels, thereby further improving semantic segmentation performance. Experiments on two challenging synthetic-to-real benchmarks, GTA5-to-Cityscape and SYNTHIA-to-Cityscape, validate the effectiveness of our method, demonstrating its advantages in dealing with complex scenes and data imbalance issues, and providing robust support for the visual perception technology of intelligent equipment.
關鍵字
domain adaptation, semantic segmentation, intelligent sensing, attention mechanism, self-training learning
報告人

發表評論