Skip to content
/ DeMo Public

【AAAI2025】DeMo: Decoupled Feature-Based Mixture of Experts for Multi-Modal Object Re-Identification

Notifications You must be signed in to change notification settings

924973292/DeMo

Repository files navigation

DeMo: Decoupled Feature-Based Mixture of Experts for Multi-Modal Object Re-Identification

Description of the image

Yuhao Wang · Yang Liu · Aihua Zheng · Pingping Zhang*

AAAI 2025 Paper

RGBNT201 Results

DeMo is an advanced multi-modal object Re-Identification (ReID) framework designed to tackle dynamic imaging quality variations across modalities. By employing decoupled features and a novel Attention-Triggered Mixture of Experts (ATMoE), DeMo dynamically balances modality-specific and modality-shared information, enabling robust performance even under missing modality conditions. The framework sets new benchmarks for multi-modal and missing-modality object ReID.

News

  • Great news! Our paper has been accepted to AAAI 2025! 🎉 Paper

Table of Contents


Introduction

Multi-modal object ReID combines the strengths of different modalities (e.g., RGB, NIR, TIR) to achieve robust identification across challenging scenarios. DeMo introduces a decoupled approach using Mixture of Experts (MoE) to preserve modality uniqueness and enhance diversity. This is achieved through:

  1. Patch-Integrated Feature Extractor (PIFE): Captures multi-granular representations.
  2. Hierarchical Decoupling Module (HDM): Separates modality-specific and shared features.
  3. Attention-Triggered Mixture of Experts (ATMoE): Dynamically adjusts feature importance with adaptive attention-guided weights.

Contributions

  • Introduced a decoupled feature-based MoE framework, DeMo, addressing dynamic quality changes in multi-modal imaging.
  • Developed the Hierarchical Decoupling Module (HDM) for enhanced feature diversity and Attention-Triggered Mixture of Experts (ATMoE) for context-aware weighting.
  • Achieved state-of-the-art performance on RGBNT201, RGBNT100, and MSVR310 benchmarks under both full and missing-modality settings.

Results

Multi-Modal Object ReID

Multi-Modal Person ReID [RGBNT201]

RGBNT201 Results

Multi-Modal Vehicle ReID [RGBNT100 & MSVR310]

RGBNT100 Results

Missing-Modality Object ReID

Missing-Modality Performance [RGBNT201]

RGBNT201 Missing-Modality

Missing-Modality Performance [RGBNT100]

RGBNT100 Missing-Modality

Ablation Studies [RGBNT201]

RGBNT201 Ablation


Visualizations

Feature Distribution (t-SNE)

t-SNE

Decoupled Features

Decoupled Features

Rank-list Visualization

Rank-list


Reproduction

Datasets

Pretrained Models

Configuration

  • RGBNT201: configs/RGBNT201/DeMo.yml
  • RGBNT100: configs/RGBNT100/DeMo.yml
  • MSVR310: configs/MSVR310/DeMo.yml

Training

#!/bin/bash
source activate (your_env)
cd (your_path)
python train_net.py --config_file configs/RGBNT201/DeMo.yml

Notes

  • This repository is based on MambaPro. The prompt and adapter tuning on the CLIP backbone are reserved (the corresponding hyperparameters are set to False), allowing users to explore them independently.
  • This code provides multi-modal Grad-CAM visualization, multi-modal ranking list generation, and t-SNE visualization tools to facilitate further research.
  • The hyperparameter configuration is designed to ensure compatibility with devices equipped with less than 24GB of memory.
  • Thank you for your attention and interest!

Star History

Star History Chart


Citation

If you find DeMo helpful in your research, please consider citing:

@inproceedings{wang2025DeMo,
  title={DeMo: Decoupled Feature-Based Mixture of Experts for Multi-Modal Object Re-Identification},
  author={Wang, Yuhao and Liu, Yang and Zheng, Aihua and Zhang, Pingping},
  booktitle={Proceedings of the AAAI Conference on Artificial Intelligence},
  year={2025}
}

About

【AAAI2025】DeMo: Decoupled Feature-Based Mixture of Experts for Multi-Modal Object Re-Identification

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published