Skip to content

[CVPR 2024] Troika: Multi-Path Cross-Modal Traction for Compositional Zero-Shot Learning

Notifications You must be signed in to change notification settings

bighuang624/Troika

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 

Repository files navigation

Troika: Multi-Path Cross-Modal Traction for Compositional Zero-Shot Learning

🎉The paper has been accepted by CVPR 2024! The code has been released.

Overview

With a particular focus on the universality of the solution, in this work, we propose a novel Multi-Path paradigm for VLM-based CZSL models that establishes three identification branches to jointly model the state, object, and composition. The presented Troika is an outstanding implementation that aligns the branch-specific prompt representations with decomposed visual features. To calibrate the bias between semantically similar multi-modal representations, we further devise a Cross-Modal Traction module into Troika that shifts the prompt representation towards the current visual content. Experiments show that on the closed-world setting, Troika exceeds the current state-of-the-art methods by up to +7.4% HM and +5.7% AUC. And on the more challenging open-world setting, Troika still surpasses the best CLIP-based method by up to +3.8% HM and +2.7% AUC.

Results

Main Results

The following results are obtained with a pre-trained CLIP (ViT-L/14). More experimental results can be found in the paper.

Qualitative Results

Citation

If you find this work useful in your research, please cite our paper:

@inproceedings{Huang2024Troika,
    title={Troika: Multi-Path Cross-Modal Traction for Compositional Zero-Shot Learning},
    author={Siteng Huang and Biao Gong and Yutong Feng and Min Zhang and Yiliang Lv and Donglin Wang},
    booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
    year={2024}
}

Acknowledgement

Our code references the following projects:

About

[CVPR 2024] Troika: Multi-Path Cross-Modal Traction for Compositional Zero-Shot Learning

Resources

Stars

Watchers

Forks