The Chinese University of Hong Kong (CUHK) • National Taiwan University (NTU)
Few-shot learning (FSL) aims to classify unseen classes with limited labeled samples. However, real-world scenarios often involve distribution shifts between support and query sets, significantly degrading performance. We introduce DUAL, a novel framework that addresses both inter-set shifts (between support and query) and intra-set shifts (within individual sets) through a dual alignment mechanism.
DUAL employs adversarial training to learn robust feature representations and optimal transport to align distributions in embedding space. Our approach includes a pixel-level repairer network that mitigates corruption effects and a dual optimal transport mechanism that performs both intra-set and inter-set alignment. Extensive experiments demonstrate that DUAL achieves state-of-the-art performance across multiple benchmarks under realistic distribution shifts.
Dual Support-Query Shift (DSQS): We formalize the challenging scenario where both support and query sets experience distribution shifts, creating a more realistic but difficult few-shot learning setting.
Robust Feature Learning: Adversarial training with a generator network creates challenging examples, forcing the encoder to learn more robust and transferable representations.
Two-Stage Optimal Transport: First, intra-set alignment reduces within-set variance. Then, inter-set alignment bridges the gap between support and query distributions using optimal transport.
DUAL achieves state-of-the-art performance across multiple datasets under distribution shift scenarios:
| Method | 1-shot | 5-shot |
|---|---|---|
| ProtoNet | 42.3±0.8 | 58.1±0.9 |
| PGADA | 45.7±0.9 | 62.4±0.8 |
| DUAL (Ours) | 48.9±0.8 | 65.7±0.7 |
| Method | 1-shot | 5-shot |
|---|---|---|
| ProtoNet | 49.4±0.8 | 68.2±0.7 |
| PGADA | 52.3±0.9 | 71.5±0.6 |
| DUAL (Ours) | 55.8±0.7 | 74.2±0.6 |
DUAL maintains superior performance even under severe distribution shifts, demonstrating the effectiveness of our dual alignment approach.
Despite the sophisticated alignment mechanism, DUAL maintains reasonable computational overhead with significant accuracy improvements.
Each component (repairer, adversarial training, dual OT) contributes meaningfully to the overall performance, validating our design choices.
If you use DUAL in your research, please cite our paper:
@inproceedings{jiang2025dual,
title = {Dual Alignment Framework for Few-shot Learning with Inter-Set and Intra-Set Shifts},
author = {Jiang, Siyang and Fang, Rui and Chen, Hsi-Wen and Ding, Wei and Xing, Guoliang and Chen, Ming-Syan},
booktitle = {Proceedings of the 39th Annual Conference on Neural Information Processing Systems (NeurIPS 2025)},
year = {2025}
}
For questions or collaborations: