Skip to content
/ UAT Public

[2025 TIP] Official implement of Uncertainty-Aware Transformer for Referring Camouflaged Object Detection

Notifications You must be signed in to change notification settings

CVL-hub/UAT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

62 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Uncertainty-Aware Transformer for Referring Camouflaged Object Detection

Authors: Ranwan Wu, Tian-Zhu Xiang, Guo-Sen Xie, Rongrong Gao, Xiangbo Shu, Fang Zhao, Ling Shao

Welcome to the official PyTorch implementation repository of our paper Uncertainty-Aware Transformer for Referring Camouflaged Object Detection, accepted to IEEE TIP 2025.

Framework

image
Figure.1 Architecture of uncertainty-aware transformer (UAT) for Ref-COD. UAT takes a camouflaged image and several referring images as input, respectively. Next, basic feature extraction on these images is performed. Then, the extracted features are fed into referring feature aggregation (RFA), cross-attention encoder (CAE), and transformer probabilistic decoder (TPD) to integrate visual reference into camouflage feature, aggregate multi-layer camouflage features, and model the dependencies between patches/tokens via Bayesian uncertainty learning, respectively. Finally, the predictions from all four stages are supervised by $L_{total}$ collaboratively.

Requirements

Python v3.6, Pytorch 0.4.0+, Cuda 10.0, TensorboardX 2.0, opencv-python

Get Start

1. Data Preparation

  • Please visiting RefCOD for training and testing data. Thanks for their contributions.

2. Training

  • Download the training and testing dataset, and place them in the ./dataset floder.
  • Download the pre-trained weights of pvtv2[code:2025] on Baidu Netdisk, and place them in the ./pvt_weights floder.
  • Run python train.py to train the model.
  • You can also download the our pre-trained UAT.pth with access code 2025 on Baidu Netdisk directly.

3. Inference

  • After training, run python infer.py to generate the prediction maps of UAT.
  • You can also download our prediction maps UAT-Maps[code:2025] on Baidu Netdisk.

4. Testing

  • After training, run python test.py to evaluate the performance of UAT.

5. Results

  • Qualitative comparison

image
Table.1 Quantitative comparison with some SOTA models on referring camouflaged bbject detection benchmark datasets.

Acknowlegement

This repo is mainly built based on R2CNet. Thanks for the great work! If you have any technical questions, feel free to contact [email protected]. If our work inspires your research, please cite it and start this project. We appreciate your support!

About

[2025 TIP] Official implement of Uncertainty-Aware Transformer for Referring Camouflaged Object Detection

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages