Exploring Universal Domain Adaptation with CLIP Models: A Calibration Method
Abstract
1. Introduction
- To the best of our knowledge, we are the first to tackle the UniDA problem and conduct a comprehensive study of existing methods when applied to CLIP models. Our findings underscore the urgent need for further research in UniDA using these powerful foundation models.
- We propose a straightforward calibration method for UniDA, establishing a new baseline for adaptation from CLIP models. Our approach incorporates a self-calibration technique based on automatic temperature scaling, making it parameter-free and robust across various task scenarios.
- We introduce a novel evaluation metric for UniDA, the Universal Classification Rate (UCR), which is insensitive to thresholds and class ratios. Additionally, to facilitate rigorous and replicable experimentation in UniDA, we have developed and made publicly available the UniOOD framework. UniOOD simplifies the incorporation of new datasets and algorithms with only a few lines of code, thereby ensuring fairer comparisons between different methods.
2. Related Works
3. Problem Formulation
4. Proposed Method
4.1. Motivations
4.2. Learning Temperature Scaling by Source Confidence Calibration
5. Evaluation and Discussion
5.1. Hard Out-of-Class Detection Criteria
5.2. Soft Out-of-Class Detection Criteria
6. Empirical Results
6.1. UniDA Methods Review
6.2. Datasets and Experimental Setup
6.3. Comparison with SOTA UniDA Methods
6.4. Comparison with SOTA CLIP-Adaptation Methods
6.5. Analysis and Ablation Study
7. Conclusions, Limitations and Future Work
Funding
Institutional Review Board Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
| DA | Domain Adaptation |
| UniDA | Universal Domain Adaptation |
| IID | In Identical Distribution |
| OOD | Out Of Distribution |
| OSR | Open-Set Recognition |
| UCR | Universal Classification Rate |
| SO | Source Only |
| NLL | Negative Log Likelihood |
| SOTA | State-Of-The-Art |
Appendix A. Experimental Setup Details
Appendix A.1. Dataset
| Office (31 categories) | Domains | Amazon (A) | DSLR (D) | Webcam (W) | - |
| Number of Samples | 2817 | 498 | 795 | - | |
| OfficeHome (65 categories) | Domains | Art (A) | Clipart (C) | Product (P) | RealWorld (R) |
| Number of Samples | 2427 | 4365 | 4439 | 4357 | |
| VisDA (12 categories) | Domains | Syn (S) | Real (R) | - | - |
| Number of Samples | 152,397 | 55,388 | - | - | |
| DomainNet (345 categories) | Domains | Painting (P) | Real (R) | Sketch (S) | - |
| Number of Samples | 50,416 | 120,906 | 48,212 | - |
Appendix A.2. Classes Split Settings
| Datasets | Split Settings | |||
|---|---|---|---|---|
| Open-Partial | Open | Closed | Partial | |
| Office | (10/10) | (10/0) | (31/0) | (10/21) |
| OfficeHome | (10/5) | (15/0) | (65/0) | (25/40) |
| VisDA | (6/3) | (6/0) | (12/0) | (6/6) |
| DomainNet | (150/50) | (150/0) | (345/0) | (150/195) |
Appendix A.3. Number of Training Iterations
| Datasets | Split Settings | |||
|---|---|---|---|---|
| Open-Partial | Open | Closed | Partial | |
| Office | 5000 | 5000 | 10,000 | 10,000 |
| OfficeHome | 5000 | 5000 | 10,000 | 10,000 |
| VisDA | 10,000 | 10,000 | 20,000 | 20,000 |
| DomainNet | 10,000 | 10,000 | 20,000 | 20,000 |
Appendix A.4. Other Settings
Appendix B. Justification for Frozen Foundation Model Backbone
| Methods | ImageNet-Pretrained | Random Initialization | DINOv2-Pretrained | CLIP-Pretrained | ||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| H-Score | H3-Score | UCR | H-Score | H3-Score | UCR | H-Score | H3-Score | UCR | H-Score | H3-Score | UCR | |
| Full Fine-tuning backbone | ||||||||||||
| SO | 58.16 | 63.73 | 52.31 | 8.01 | 1.45 | 6.78 | 0.54 | 0.46 | 7.16 | 2.15 | 1.95 | 8.46 |
| DANCE [22] | 42.36 | 49.17 | 28.36 | 0.82 | 1.02 | 5.81 | 0.42 | 0.44 | 5.94 | 0.46 | 0.58 | 5.98 |
| OVANet [25] | 38.27 | 45.60 | 53.63 | 1.55 | 1.23 | 7.36 | 6.65 | 0.82 | 5.51 | 1.06 | 1.08 | 4.40 |
| UniOT [26] | 71.23 | 70.93 | 65.52 | 11.95 | 2.17 | 9.44 | 7.56 | 2.13 | 5.98 | 15.02 | 2.67 | 8.42 |
| Freeze backbone | ||||||||||||
| SO | 56.69 | 62.19 | 52.90 | 14.16 | 1.73 | 5.09 | 57.63 | 65.16 | 53.12 | 70.30 | 73.58 | 67.28 |
| DANCE [22] | 42.59 | 50.07 | 28.12 | 10.08 | 1.67 | 6.40 | 44.79 | 53.58 | 34.53 | 67.79 | 71.73 | 54.17 |
| OVANet [25] | 56.36 | 62.75 | 67.77 | 6.55 | 1.58 | 4.30 | 57.91 | 65.40 | 48.86 | 56.36 | 62.75 | 67.77 |
| UniOT [26] | 68.26 | 67.16 | 62.25 | 11.24 | 1.39 | 7.59 | 62.73 | 67.12 | 52.27 | 75.87 | 74.81 | 67.58 |
Appendix C. Detail Experimental Results
| Methods | A2D | A2W | D2A | D2W | W2A | W2D | Avg |
|---|---|---|---|---|---|---|---|
| H-score | |||||||
| SO | 54.48 ± 3.1 | 53.47 ± 1.48 | 79.11 ± 0.25 | 76.59 ± 1.0 | 73.56 ± 1.42 | 60.82 ± 1.84 | 66.34 |
| DANCE | 77.87 ± 0.94 | 75.97 ± 0.92 | 77.0 ± 1.74 | 90.92 ± 2.59 | 72.37 ± 2.69 | 87.79 ± 2.16 | 80.32 |
| OVANet | 82.95 ± 0.56 | 77.61 ± 1.19 | 67.91 ± 2.24 | 94.99 ± 0.3 | 81.25 ± 0.37 | 95.26 ± 0.27 | 83.33 |
| UniOT | 77.09 ± 1.11 | 76.73 ± 0.57 | 86.44 ± 0.42 | 91.22 ± 0.79 | 85.45 ± 0.53 | 89.3 ± 0.52 | 84.37 |
| H3-score | |||||||
| SO | 60.18 ± 2.76 | 59.35 ± 1.0 | 66.71 ± 0.44 | 75.07 ± 0.61 | 65.44 ± 0.49 | 66.54 ± 1.46 | 65.55 |
| DANCE | 73.08 ± 1.55 | 70.89 ± 0.66 | 64.03 ± 1.56 | 79.95 ± 1.89 | 62.11 ± 1.63 | 79.35 ± 1.81 | 71.57 |
| OVANet | 80.45 ± 0.67 | 76.76 ± 1.04 | 59.72 ± 1.45 | 87.29 ± 1.24 | 67.26 ± 0.37 | 89.29 ± 0.93 | 76.8 |
| UniOT | 76.88 ± 2.23 | 75.02 ± 0.64 | 69.41 ± 1.86 | 86.17 ± 1.65 | 68.86 ± 1.4 | 87.66 ± 0.74 | 77.33 |
| UCR | |||||||
| SO | 69.72 ± 1.45 | 62.83 ± 2.31 | 79.54 ± 0.62 | 94.19 ± 1.27 | 82.78 ± 0.9 | 98.2 ± 0.22 | 81.21 |
| DANCE | 79.04 ± 2.5 | 79.86 ± 0.87 | 82.61 ± 0.48 | 93.21 ± 1.29 | 81.68 ± 0.72 | 90.42 ± 2.18 | 84.47 |
| OVANet | 71.79 ± 0.64 | 65.18 ± 1.03 | 73.98 ± 2.03 | 97.3 ± 0.67 | 81.46 ± 0.44 | 98.54 ± 0.14 | 81.38 |
| UniOT | 72.66 ± 2.24 | 72.81 ± 2.18 | 87.12 ± 0.98 | 94.74 ± 0.75 | 87.51 ± 0.5 | 93.57 ± 2.33 | 84.73 |
| Methods | A2D | A2W | D2A | D2W | W2A | W2D | Avg |
|---|---|---|---|---|---|---|---|
| H-score | |||||||
| SO | 92.71 ± 0.13 | 89.3 ± 0.19 | 90.08 ± 0.21 | 93.95 ± 0.17 | 88.16 ± 0.22 | 97.69 ± 0.2 | 91.98 |
| DANCE | 96.02 ± 0.2 | 90.18 ± 0.83 | 93.68 ± 2.56 | 98.64 ± 0.21 | 90.23 ± 2.67 | 99.42 ± 0.24 | 94.69 |
| OVANet | 93.82 ± 0.58 | 88.88 ± 0.63 | 92.3 ± 2.09 | 97.63 ± 0.0 | 89.35 ± 0.21 | 98.16 ± 0.0 | 93.36 |
| UniOT | 84.59 ± 1.64 | 92.24 ± 1.23 | 94.5 ± 1.58 | 94.84 ± 1.79 | 94.76 ± 0.95 | 92.99 ± 0.61 | 92.32 |
| WiSE-FT | 77.89 ± 0.42 | 70.19 ± 0.1 | 80.83 ± 0.22 | 92.68 ± 0.17 | 76.75 ± 0.33 | 95.7 ± 0.37 | 82.34 |
| CLIP cross-model | 94.24 ± 0.39 | 89.82 ± 0.32 | 92.34 ± 0.11 | 93.65 ± 0.3 | 92.23 ± 0.05 | 95.99 ± 0.24 | 93.05 |
| Our calibration | 91.47 ± 0.0 | 85.13 ± 0.0 | 85.3 ± 0.0 | 82.04 ± 0.0 | 86.46 ± 0.0 | 90.06 ± 0.0 | 86.74 |
| H3-score | |||||||
| SO | 92.41 ± 0.09 | 91.54 ± 0.13 | 83.22 ± 0.12 | 94.74 ± 0.12 | 82.12 ± 0.13 | 95.65 ± 0.13 | 89.95 |
| DANCE | 94.58 ± 0.13 | 92.15 ± 0.58 | 85.22 ± 1.4 | 97.87 ± 0.14 | 83.29 ± 1.53 | 96.75 ± 0.15 | 91.64 |
| OVANet | 93.14 ± 0.38 | 91.24 ± 0.44 | 84.47 ± 1.17 | 97.2 ± 0.0 | 82.81 ± 0.12 | 95.95 ± 0.0 | 90.8 |
| UniOT | 87.14 ± 0.7 | 91.12 ± 0.48 | 84.31 ± 1.23 | 94.03 ± 2.03 | 84.27 ± 1.17 | 93.56 ± 0.34 | 89.07 |
| WiSE-FT | 82.04 ± 0.31 | 77.18 ± 0.08 | 77.74 ± 0.14 | 93.88 ± 0.11 | 75.18 ± 0.21 | 94.37 ± 0.24 | 83.4 |
| CLIP cross-model | 93.41 ± 0.25 | 91.9 ± 0.22 | 84.5 ± 0.06 | 94.54 ± 0.2 | 84.43 ± 0.03 | 94.55 ± 0.16 | 90.56 |
| Our calibration | 91.59 ± 0.0 | 88.57 ± 0.0 | 80.45 ± 0.0 | 86.32 ± 0.0 | 81.13 ± 0.0 | 90.64 ± 0.0 | 86.45 |
| UCR | |||||||
| SO | 88.41 ± 0.08 | 90.78 ± 0.01 | 93.09 ± 0.24 | 98.55 ± 0.09 | 93.27 ± 0.15 | 99.8 ± 0.01 | 93.98 |
| DANCE | 95.96 ± 0.3 | 93.03 ± 1.31 | 92.92 ± 3.15 | 99.61 ± 0.03 | 89.5 ± 2.19 | 99.99 ± 0.0 | 95.17 |
| OVANet | 91.39 ± 0.11 | 92.25 ± 0.45 | 94.92 ± 0.2 | 99.18 ± 0.02 | 94.46 ± 0.26 | 99.94 ± 0.01 | 95.36 |
| UniOT | 76.31 ± 2.68 | 90.53 ± 0.4 | 93.58 ± 1.57 | 95.05 ± 0.95 | 94.31 ± 1.24 | 93.94 ± 0.64 | 90.62 |
| WiSE-FT | 91.82 ± 0.07 | 92.29 ± 0.05 | 94.53 ± 0.1 | 98.52 ± 0.04 | 95.0 ± 0.08 | 99.48 ± 0.3 | 95.27 |
| CLIP cross-model | 90.86 ± 0.17 | 92.18 ± 0.07 | 95.49 ± 0.01 | 98.55 ± 0.01 | 95.42 ± 0.01 | 99.76 ± 0.0 | 95.38 |
| CLIP zero-shot | 91.26 ± 0.0 | 89.87 ± 0.0 | 89.17 ± 0.0 | 89.87 ± 0.0 | 89.17 ± 0.0 | 91.26 ± 0.0 | 90.1 |
| Our calibration | 93.64 ± 0.0 | 92.6 ± 0.0 | 94.0 ± 0.0 | 92.54 ± 0.0 | 94.04 ± 0.0 | 93.54 ± 0.0 | 93.39 |
| Methods | A2C | A2P | A2R | C2A | C2P | C2R | P2A | P2C | P2R | R2A | R2C | R2P | Avg |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| H-score | |||||||||||||
| SO | 50.35 ± 0.25 | 50.87 ± 0.31 | 55.44 ± 0.59 | 59.31 ± 1.19 | 49.15 ± 0.57 | 57.57 ± 1.09 | 62.01 ± 0.48 | 50.21 ± 0.68 | 56.83 ± 0.41 | 56.43 ± 0.12 | 52.31 ± 0.3 | 52.19 ± 0.41 | 54.39 |
| DANCE | 39.64 ± 3.7 | 38.23 ± 5.91 | 38.69 ± 3.66 | 38.55 ± 3.4 | 13.22 ± 0.27 | 37.21 ± 0.43 | 51.73 ± 2.72 | 43.89 ± 1.31 | 43.2 ± 1.78 | 29.33 ± 4.01 | 44.47 ± 3.19 | 50.62 ± 0.88 | 39.06 |
| OVANet | 58.01 ± 0.64 | 78.91 ± 0.18 | 82.15 ± 0.56 | 69.4 ± 0.63 | 68.1 ± 0.22 | 76.41 ± 0.08 | 71.98 ± 0.58 | 56.77 ± 0.23 | 81.72 ± 0.11 | 77.94 ± 0.55 | 58.91 ± 0.21 | 79.81 ± 0.46 | 71.68 |
| UniOT | 66.13 ± 0.97 | 80.42 ± 0.7 | 84.56 ± 0.54 | 72.79 ± 0.2 | 76.59 ± 1.66 | 82.42 ± 0.96 | 75.82 ± 0.96 | 65.87 ± 0.76 | 85.07 ± 1.14 | 76.61 ± 0.3 | 64.8 ± 1.23 | 80.6 ± 0.44 | 75.97 |
| H3-score | |||||||||||||
| SO | 51.29 ± 0.3 | 57.77 ± 0.29 | 60.11 ± 0.52 | 57.45 ± 0.76 | 55.1 ± 0.45 | 60.21 ± 0.78 | 60.58 ± 0.38 | 50.49 ± 0.28 | 61.05 ± 0.32 | 57.72 ± 0.08 | 52.1 ± 0.12 | 58.55 ± 0.33 | 56.87 |
| DANCE | 42.95 ± 3.02 | 45.61 ± 5.5 | 45.11 ± 3.32 | 43.56 ± 2.91 | 18.22 ± 0.34 | 43.86 ± 0.39 | 54.24 ± 2.11 | 46.3 ± 0.88 | 49.38 ± 1.61 | 35.1 ± 3.77 | 46.67 ± 2.44 | 56.98 ± 0.71 | 44.0 |
| OVANet | 53.85 ± 0.3 | 77.3 ± 0.23 | 77.29 ± 0.11 | 65.06 ± 0.44 | 68.91 ± 0.25 | 72.69 ± 0.18 | 66.75 ± 0.54 | 53.51 ± 0.29 | 76.83 ± 0.02 | 70.97 ± 0.53 | 55.14 ± 0.38 | 77.25 ± 0.29 | 67.96 |
| UniOT | 59.59 ± 1.0 | 78.39 ± 0.47 | 78.43 ± 0.36 | 65.85 ± 0.31 | 74.71 ± 1.07 | 75.91 ± 0.74 | 69.21 ± 1.03 | 58.94 ± 0.6 | 79.31 ± 1.22 | 68.95 ± 0.19 | 58.33 ± 0.81 | 78.23 ± 0.41 | 70.49 |
| UCR | |||||||||||||
| SO | 38.15 ± 0.75 | 74.76 ± 0.35 | 89.28 ± 0.45 | 61.78 ± 1.34 | 63.97 ± 0.77 | 78.44 ± 1.84 | 62.67 ± 0.39 | 37.08 ± 0.14 | 85.66 ± 0.41 | 69.25 ± 0.91 | 39.61 ± 0.22 | 81.54 ± 0.92 | 65.18 |
| DANCE | 53.53 ± 0.77 | 72.81 ± 0.82 | 80.43 ± 1.28 | 70.95 ± 0.91 | 69.87 ± 1.65 | 78.64 ± 1.24 | 73.94 ± 0.38 | 50.25 ± 0.52 | 82.14 ± 0.77 | 69.4 ± 2.15 | 52.06 ± 0.82 | 78.11 ± 0.51 | 69.34 |
| OVANet | 42.81 ± 0.9 | 77.99 ± 0.1 | 89.85 ± 0.29 | 65.89 ± 0.97 | 65.1 ± 0.36 | 79.33 ± 0.74 | 65.7 ± 1.32 | 39.84 ± 0.03 | 87.6 ± 0.43 | 74.79 ± 1.27 | 42.57 ± 0.35 | 82.55 ± 0.44 | 67.83 |
| UniOT | 50.72 ± 1.15 | 83.24 ± 1.46 | 92.86 ± 0.45 | 72.98 ± 1.91 | 77.33 ± 1.93 | 89.54 ± 1.01 | 68.89 ± 2.22 | 47.06 ± 1.21 | 89.51 ± 0.93 | 75.29 ± 1.35 | 51.6 ± 1.26 | 84.75 ± 0.71 | 73.65 |
| Methods | A2C | A2P | A2R | C2A | C2P | C2R | P2A | P2C | P2R | R2A | R2C | R2P | Avg |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| H-score | |||||||||||||
| SO | 75.73 ± 0.08 | 84.45 ± 0.07 | 88.31 ± 0.02 | 86.28 ± 0.16 | 88.14 ± 0.23 | 91.1 ± 0.12 | 81.52 ± 0.14 | 78.42 ± 0.1 | 89.53 ± 0.13 | 85.3 ± 0.07 | 81.68 ± 0.04 | 83.82 ± 0.09 | 84.52 |
| DANCE | 82.2 ± 0.08 | 94.02 ± 0.01 | 90.2 ± 0.06 | 86.63 ± 0.16 | 93.72 ± 0.16 | 93.79 ± 0.02 | 85.35 ± 0.21 | 84.47 ± 0.09 | 91.7 ± 0.11 | 86.44 ± 0.11 | 83.97 ± 0.05 | 95.64 ± 0.39 | 89.01 |
| OVANet | 80.55 ± 0.4 | 91.84 ± 0.08 | 90.74 ± 0.27 | 86.07 ± 0.17 | 90.52 ± 0.11 | 91.78 ± 0.12 | 75.64 ± 0.67 | 68.54 ± 0.71 | 90.29 ± 0.12 | 85.74 ± 0.2 | 80.56 ± 0.31 | 92.75 ± 0.77 | 85.42 |
| UniOT | 88.28 ± 0.33 | 92.03 ± 0.56 | 93.39 ± 0.31 | 87.61 ± 0.56 | 90.19 ± 0.44 | 92.09 ± 0.82 | 85.99 ± 0.37 | 86.08 ± 0.36 | 92.04 ± 0.7 | 86.87 ± 0.13 | 86.47 ± 0.43 | 92.35 ± 0.57 | 89.45 |
| WiSE-FT | 68.39 ± 0.04 | 89.68 ± 0.01 | 90.9 ± 0.04 | 69.46 ± 0.27 | 87.43 ± 0.14 | 87.14 ± 0.2 | 68.7 ± 0.45 | 59.92 ± 0.34 | 85.8 ± 0.05 | 78.64 ± 0.26 | 73.97 ± 0.29 | 92.43 ± 0.11 | 79.37 |
| CLIP cross-model | 78.37 ± 0.06 | 85.06 ± 0.05 | 89.18 ± 0.12 | 88.98 ± 0.08 | 88.09 ± 0.24 | 91.52 ± 0.1 | 85.91 ± 0.05 | 82.6 ± 0.11 | 91.59 ± 0.07 | 86.8 ± 0.17 | 82.09 ± 0.05 | 84.24 ± 0.09 | 86.2 |
| Our calibration | 82.39 ± 0.0 | 79.09 ± 0.0 | 85.41 ± 0.0 | 89.43 ± 0.0 | 81.65 ± 0.0 | 87.21 ± 0.0 | 90.06 ± 0.0 | 84.62 ± 0.0 | 93.76 ± 0.0 | 90.4 ± 0.0 | 85.17 ± 0.0 | 87.64 ± 0.0 | 86.4 |
| H3-score | |||||||||||||
| SO | 74.95 ± 0.05 | 86.46 ± 0.05 | 87.22 ± 0.01 | 79.88 ± 0.09 | 89.0 ± 0.16 | 89.02 ± 0.08 | 77.1 ± 0.09 | 76.69 ± 0.06 | 88.01 ± 0.09 | 79.31 ± 0.04 | 78.74 ± 0.02 | 86.02 ± 0.06 | 82.7 |
| DANCE | 79.06 ± 0.05 | 92.91 ± 0.01 | 88.44 ± 0.04 | 80.08 ± 0.09 | 92.72 ± 0.11 | 90.71 ± 0.01 | 79.34 ± 0.12 | 80.45 ± 0.06 | 89.4 ± 0.07 | 79.97 ± 0.06 | 80.14 ± 0.03 | 93.96 ± 0.25 | 85.6 |
| OVANet | 78.03 ± 0.25 | 91.48 ± 0.05 | 88.79 ± 0.17 | 79.76 ± 0.1 | 90.61 ± 0.07 | 89.45 ± 0.08 | 73.5 ± 0.42 | 70.1 ± 0.5 | 88.5 ± 0.07 | 79.57 ± 0.11 | 78.04 ± 0.19 | 92.08 ± 0.51 | 83.33 |
| UniOT | 83.27 ± 0.32 | 91.93 ± 0.37 | 91.63 ± 0.23 | 83.12 ± 0.53 | 90.97 ± 0.37 | 90.32 ± 0.58 | 82.57 ± 0.81 | 82.11 ± 0.1 | 90.76 ± 0.35 | 84.06 ± 0.56 | 82.16 ± 0.2 | 92.22 ± 0.12 | 87.09 |
| WiSE-FT | 69.99 ± 0.03 | 90.04 ± 0.01 | 88.89 ± 0.02 | 69.49 ± 0.18 | 88.52 ± 0.09 | 86.46 ± 0.13 | 68.98 ± 0.3 | 63.84 ± 0.26 | 85.58 ± 0.03 | 75.36 ± 0.16 | 73.79 ± 0.19 | 91.87 ± 0.07 | 79.4 |
| CLIP cross-model | 76.66 ± 0.04 | 86.88 ± 0.03 | 87.79 ± 0.08 | 81.4 ± 0.05 | 88.97 ± 0.16 | 89.29 ± 0.06 | 79.67 ± 0.03 | 79.31 ± 0.07 | 89.33 ± 0.04 | 80.17 ± 0.09 | 78.99 ± 0.03 | 86.31 ± 0.06 | 83.73 |
| Our calibration | 79.18 ± 0.0 | 82.63 ± 0.0 | 85.32 ± 0.0 | 81.65 ± 0.0 | 84.48 ± 0.0 | 86.5 ± 0.0 | 82.0 ± 0.0 | 80.54 ± 0.0 | 90.69 ± 0.0 | 82.19 ± 0.0 | 80.87 ± 0.0 | 88.66 ± 0.0 | 83.73 |
| UCR | |||||||||||||
| SO | 71.07 ± 0.09 | 91.59 ± 0.09 | 95.69 ± 0.06 | 90.87 ± 0.12 | 93.28 ± 0.11 | 95.59 ± 0.06 | 76.68 ± 0.1 | 76.23 ± 0.13 | 94.01 ± 0.04 | 86.36 ± 0.06 | 76.35 ± 0.19 | 95.01 ± 0.08 | 86.89 |
| DANCE | 80.3 ± 0.07 | 96.38 ± 0.07 | 95.85 ± 0.04 | 90.33 ± 0.07 | 96.25 ± 0.08 | 96.87 ± 0.04 | 80.12 ± 0.12 | 85.02 ± 0.17 | 95.72 ± 0.06 | 85.97 ± 0.11 | 82.83 ± 0.12 | 98.33 ± 0.03 | 90.33 |
| OVANet | 74.67 ± 0.25 | 94.7 ± 0.11 | 96.66 ± 0.08 | 91.45 ± 0.12 | 95.04 ± 0.09 | 95.75 ± 0.11 | 74.86 ± 0.48 | 80.2 ± 0.12 | 95.05 ± 0.03 | 86.19 ± 0.25 | 77.77 ± 0.15 | 95.88 ± 0.4 | 88.18 |
| UniOT | 83.66 ± 1.53 | 95.93 ± 0.71 | 97.09 ± 0.12 | 85.3 ± 0.44 | 94.79 ± 1.04 | 95.52 ± 0.62 | 79.53 ± 3.62 | 81.36 ± 0.86 | 95.12 ± 0.1 | 79.36 ± 2.17 | 82.07 ± 1.19 | 96.52 ± 0.38 | 88.85 |
| WiSE-FT | 76.68 ± 0.05 | 94.9 ± 0.03 | 97.41 ± 0.03 | 93.97 ± 0.08 | 96.57 ± 0.11 | 97.54 ± 0.04 | 87.12 ± 0.31 | 80.76 ± 0.03 | 96.53 ± 0.02 | 92.18 ± 0.1 | 78.95 ± 0.13 | 96.6 ± 0.01 | 90.77 |
| CLIP cross-model | 74.37 ± 0.12 | 93.56 ± 0.2 | 96.71 ± 0.04 | 92.74 ± 0.18 | 95.33 ± 0.1 | 96.82 ± 0.04 | 86.83 ± 0.01 | 80.21 ± 0.07 | 96.41 ± 0.01 | 89.66 ± 0.05 | 77.93 ± 0.05 | 95.92 ± 0.08 | 89.71 |
| CLIP zero-shot | 80.3 ± 0.0 | 94.82 ± 0.0 | 94.29 ± 0.0 | 91.45 ± 0.0 | 94.82 ± 0.0 | 94.29 ± 0.0 | 91.45 ± 0.0 | 80.3 ± 0.0 | 94.29 ± 0.0 | 91.45 ± 0.0 | 80.3 ± 0.0 | 94.82 ± 0.0 | 90.21 |
| Our calibration | 83.11 ± 0.0 | 96.4 ± 0.0 | 97.78 ± 0.0 | 94.63 ± 0.0 | 96.45 ± 0.0 | 97.8 ± 0.0 | 94.75 ± 0.0 | 83.08 ± 0.0 | 97.87 ± 0.0 | 94.73 ± 0.0 | 83.09 ± 0.0 | 96.59 ± 0.0 | 93.02 |
| Methods | P2R | P2S | R2P | R2S | S2P | S2R | Avg |
|---|---|---|---|---|---|---|---|
| H-score | |||||||
| SO | 42.24 ± 0.13 | 40.84 ± 0.22 | 45.31 ± 0.43 | 41.1 ± 0.12 | 28.05 ± 0.24 | 37.67 ± 0.47 | 39.2 |
| DANCE | 20.09 ± 0.26 | 25.86 ± 0.92 | 34.6 ± 1.19 | 41.83 ± 0.42 | 18.68 ± 0.49 | 20.42 ± 0.29 | 26.91 |
| OVANet | 55.29 ± 0.08 | 45.35 ± 0.2 | 52.17 ± 0.17 | 44.69 ± 0.05 | 44.62 ± 0.33 | 55.32 ± 0.06 | 49.57 |
| UniOT | 56.75 ± 0.11 | 47.4 ± 0.21 | 51.72 ± 0.23 | 47.23 ± 0.24 | 46.17 ± 0.23 | 56.02 ± 0.28 | 50.88 |
| H3-score | |||||||
| SO | 47.97 ± 0.11 | 40.92 ± 0.13 | 47.43 ± 0.29 | 41.54 ± 0.08 | 31.99 ± 0.21 | 43.08 ± 0.39 | 42.16 |
| DANCE | 26.22 ± 0.31 | 29.49 ± 0.86 | 38.98 ± 1.04 | 42.53 ± 0.33 | 23.42 ± 0.54 | 26.47 ± 0.31 | 31.18 |
| OVANet | 58.24 ± 0.07 | 44.2 ± 0.15 | 52.12 ± 0.12 | 43.67 ± 0.08 | 46.09 ± 0.23 | 57.72 ± 0.1 | 50.34 |
| UniOT | 55.7 ± 0.12 | 42.39 ± 0.18 | 47.78 ± 0.28 | 42.51 ± 0.18 | 44.28 ± 0.17 | 54.78 ± 0.29 | 47.91 |
| UCR | |||||||
| SO | 43.03 ± 0.23 | 25.57 ± 0.21 | 36.19 ± 0.13 | 26.03 ± 0.15 | 21.6 ± 0.02 | 33.85 ± 0.31 | 31.04 |
| DANCE | 42.56 ± 0.13 | 23.88 ± 0.58 | 35.89 ± 0.58 | 29.33 ± 0.36 | 25.61 ± 0.12 | 38.99 ± 0.09 | 32.71 |
| OVANet | 43.03 ± 0.11 | 27.57 ± 0.38 | 38.18 ± 0.02 | 27.65 ± 0.26 | 30.79 ± 0.29 | 39.34 ± 0.19 | 34.43 |
| UniOT | 43.41 ± 0.15 | 27.72 ± 0.42 | 36.16 ± 0.06 | 28.57 ± 0.34 | 29.8 ± 0.09 | 41.48 ± 0.43 | 34.52 |
| Methods | P2R | P2S | R2P | R2S | S2P | S2R | Avg |
|---|---|---|---|---|---|---|---|
| H-score | |||||||
| SO | 67.35 ± 0.07 | 57.55 ± 0.04 | 58.17 ± 0.12 | 61.14 ± 0.19 | 53.01 ± 0.14 | 71.72 ± 0.06 | 61.49 |
| DANCE | 67.6 ± 0.07 | 57.33 ± 0.06 | 55.57 ± 0.25 | 59.65 ± 0.06 | 52.09 ± 0.08 | 70.94 ± 0.12 | 60.53 |
| OVANet | 74.87 ± 0.15 | 69.55 ± 0.11 | 67.97 ± 0.27 | 70.41 ± 0.17 | 65.73 ± 0.09 | 75.65 ± 0.11 | 70.7 |
| UniOT | 74.67 ± 0.39 | 69.56 ± 0.19 | 69.32 ± 0.47 | 71.38 ± 0.21 | 67.15 ± 0.52 | 76.42 ± 0.16 | 71.42 |
| WiSE-FT | 5.79 ± 0.04 | 1.51 ± 0.04 | 4.24 ± 0.14 | 3.71 ± 0.03 | 1.77 ± 0.04 | 5.45 ± 0.09 | 3.74 |
| CLIP cross-model | 69.19 ± 0.1 | 58.18 ± 0.14 | 57.11 ± 0.03 | 60.85 ± 0.18 | 53.13 ± 0.16 | 73.43 ± 0.05 | 61.98 |
| Our calibration | 69.2 ± 0.0 | 69.09 ± 0.0 | 74.78 ± 0.0 | 74.85 ± 0.0 | 72.68 ± 0.0 | 73.61 ± 0.0 | 72.37 |
| H3-score | |||||||
| SO | 72.52 ± 0.06 | 58.72 ± 0.03 | 61.83 ± 0.09 | 61.17 ± 0.13 | 57.84 ± 0.11 | 75.83 ± 0.04 | 64.65 |
| DANCE | 72.71 ± 0.05 | 58.57 ± 0.04 | 59.84 ± 0.19 | 60.16 ± 0.04 | 57.11 ± 0.06 | 75.25 ± 0.09 | 63.94 |
| OVANet | 78.15 ± 0.11 | 66.53 ± 0.07 | 68.86 ± 0.18 | 67.06 ± 0.1 | 67.31 ± 0.06 | 78.72 ± 0.08 | 71.11 |
| UniOT | 74.89 ± 0.11 | 65.78 ± 0.12 | 67.97 ± 0.34 | 67.65 ± 0.09 | 66.15 ± 0.36 | 76.95 ± 0.25 | 69.9 |
| WiSE-FT | 8.4 ± 0.05 | 2.24 ± 0.06 | 6.18 ± 0.2 | 5.41 ± 0.05 | 2.63 ± 0.06 | 7.92 ± 0.13 | 5.46 |
| CLIP cross-model | 73.93 ± 0.08 | 59.16 ± 0.1 | 61.03 ± 0.02 | 60.97 ± 0.12 | 57.94 ± 0.12 | 77.1 ± 0.04 | 65.02 |
| Our calibration | 73.94 ± 0.0 | 66.25 ± 0.0 | 73.38 ± 0.0 | 69.68 ± 0.0 | 72.02 ± 0.0 | 77.24 ± 0.0 | 72.08 |
| UCR | |||||||
| SO | 66.02 ± 0.08 | 58.9 ± 0.15 | 62.84 ± 0.14 | 62.92 ± 0.1 | 56.69 ± 0.15 | 71.78 ± 0.18 | 63.19 |
| DANCE | 67.79 ± 0.1 | 60.3 ± 0.19 | 62.77 ± 0.05 | 63.68 ± 0.09 | 62.86 ± 0.15 | 71.88 ± 0.16 | 64.88 |
| OVANet | 68.0 ± 0.2 | 59.14 ± 0.18 | 64.61 ± 0.13 | 62.38 ± 0.11 | 58.54 ± 0.19 | 73.13 ± 0.18 | 64.3 |
| UniOT | 68.37 ± 0.31 | 57.98 ± 0.38 | 59.43 ± 0.4 | 61.62 ± 0.05 | 57.71 ± 0.52 | 72.16 ± 0.22 | 62.88 |
| WiSE-FT | 73.53 ± 0.11 | 63.93 ± 0.18 | 66.78 ± 0.07 | 66.38 ± 0.1 | 64.47 ± 0.08 | 77.24 ± 0.12 | 68.72 |
| CLIP cross-model | 74.17 ± 0.11 | 63.87 ± 0.05 | 66.87 ± 0.1 | 66.39 ± 0.09 | 63.48 ± 0.19 | 78.06 ± 0.06 | 68.81 |
| CLIP zero-shot | 79.43 ± 0.0 | 65.78 ± 0.0 | 67.12 ± 0.0 | 65.78 ± 0.0 | 67.12 ± 0.0 | 79.43 ± 0.0 | 70.78 |
| Our calibration | 83.53 ± 0.0 | 69.39 ± 0.0 | 71.62 ± 0.0 | 69.49 ± 0.0 | 71.57 ± 0.0 | 83.58 ± 0.0 | 74.86 |
| Methods | A2D | A2W | D2A | D2W | W2A | W2D | Avg |
|---|---|---|---|---|---|---|---|
| H-score | |||||||
| SO | 92.04 ± 0.19 | 92.01 ± 0.13 | 91.71 ± 0.27 | 95.51 ± 0.09 | 89.77 ± 0.33 | 90.17 ± 0.43 | 91.87 |
| DANCE | 93.15 ± 0.21 | 93.08 ± 0.05 | 96.74 ± 0.06 | 98.74 ± 0.05 | 95.81 ± 1.5 | 99.0 ± 0.89 | 96.09 |
| OVANet | 93.3 ± 1.17 | 90.3 ± 0.84 | 82.93 ± 3.14 | 97.48 ± 0.11 | 85.37 ± 0.83 | 97.58 ± 1.61 | 91.16 |
| UniOT | 95.11 ± 1.4 | 95.92 ± 0.42 | 95.45 ± 0.71 | 98.36 ± 1.04 | 95.68 ± 0.94 | 98.35 ± 0.17 | 96.48 |
| WiSE-FT | 92.77 ± 0.0 | 90.99 ± 0.0 | 92.78 ± 0.37 | 97.87 ± 0.08 | 92.1 ± 0.23 | 97.93 ± 0.2 | 94.07 |
| CLIP cross-model | 93.51 ± 0.16 | 92.19 ± 0.05 | 93.85 ± 0.01 | 94.97 ± 0.0 | 92.47 ± 0.01 | 94.54 ± 0.08 | 93.59 |
| Our calibration | 89.75 ± 0.0 | 92.0 ± 0.0 | 92.67 ± 0.0 | 93.51 ± 0.0 | 92.14 ± 0.0 | 91.28 ± 0.0 | 91.89 |
| H3-score | |||||||
| SO | 91.55 ± 0.12 | 91.87 ± 0.08 | 86.0 ± 0.16 | 94.17 ± 0.06 | 84.85 ± 0.2 | 90.32 ± 0.29 | 89.79 |
| DANCE | 92.29 ± 0.14 | 92.58 ± 0.04 | 88.89 ± 0.03 | 96.24 ± 0.03 | 88.36 ± 0.85 | 96.04 ± 0.56 | 92.4 |
| OVANet | 92.38 ± 0.76 | 90.73 ± 0.57 | 80.64 ± 2.0 | 95.44 ± 0.07 | 82.18 ± 0.51 | 95.13 ± 1.03 | 89.42 |
| UniOT | 94.19 ± 0.85 | 95.24 ± 0.09 | 88.09 ± 0.89 | 96.52 ± 0.34 | 88.22 ± 0.27 | 97.18 ± 0.13 | 93.24 |
| WiSE-FT | 92.04 ± 0.0 | 91.19 ± 0.0 | 86.62 ± 0.22 | 95.69 ± 0.05 | 86.22 ± 0.14 | 95.36 ± 0.13 | 91.19 |
| CLIP cross-model | 92.52 ± 0.1 | 91.99 ± 0.03 | 87.24 ± 0.01 | 93.82 ± 0.0 | 86.44 ± 0.01 | 93.19 ± 0.05 | 90.87 |
| Our calibration | 90.04 ± 0.0 | 91.86 ± 0.0 | 86.56 ± 0.0 | 92.86 ± 0.0 | 86.25 ± 0.0 | 91.05 ± 0.0 | 89.77 |
| UCR | |||||||
| SO | 88.49 ± 0.1 | 91.81 ± 0.21 | 95.35 ± 0.18 | 99.32 ± 0.03 | 94.83 ± 0.18 | 99.88 ± 0.02 | 94.95 |
| DANCE | 94.74 ± 1.76 | 94.2 ± 0.04 | 97.14 ± 0.01 | 99.58 ± 0.0 | 96.94 ± 0.27 | 99.97 ± 0.02 | 97.09 |
| OVANet | 91.89 ± 0.35 | 94.05 ± 0.29 | 95.31 ± 0.41 | 99.42 ± 0.02 | 94.91 ± 0.22 | 99.24 ± 1.05 | 95.8 |
| UniOT | 93.27 ± 1.82 | 97.54 ± 1.0 | 96.7 ± 0.12 | 99.61 ± 0.23 | 96.65 ± 0.07 | 99.44 ± 0.4 | 97.2 |
| WiSE-FT | 91.21 ± 0.07 | 95.19 ± 0.07 | 96.26 ± 0.09 | 99.45 ± 0.02 | 96.13 ± 0.07 | 99.74 ± 0.3 | 96.33 |
| CLIP cross-model | 90.77 ± 0.27 | 94.46 ± 0.33 | 96.53 ± 0.0 | 99.35 ± 0.0 | 96.13 ± 0.05 | 99.85 ± 0.0 | 96.18 |
| CLIP zero-shot | 98.69 ± 0.0 | 98.98 ± 0.0 | 95.38 ± 0.0 | 98.98 ± 0.0 | 95.38 ± 0.0 | 98.69 ± 0.0 | 97.68 |
| Our calibration | 98.74 ± 0.0 | 99.0 ± 0.0 | 95.98 ± 0.0 | 99.03 ± 0.0 | 95.98 ± 0.0 | 98.73 ± 0.0 | 97.91 |
| Methods | A2C | A2P | A2R | C2A | C2P | C2R | P2A | P2C | P2R | R2A | R2C | R2P | Avg |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| H-score | |||||||||||||
| SO | 75.41 ± 0.09 | 83.2 ± 0.05 | 86.19 ± 0.14 | 81.69 ± 0.12 | 87.15 ± 0.21 | 86.87 ± 0.08 | 75.93 ± 0.28 | 76.58 ± 0.06 | 86.49 ± 0.21 | 83.02 ± 0.09 | 78.68 ± 0.02 | 83.39 ± 0.08 | 82.05 |
| DANCE | 80.03 ± 0.04 | 89.48 ± 0.04 | 87.88 ± 0.1 | 80.8 ± 0.1 | 89.69 ± 0.17 | 88.31 ± 0.09 | 73.42 ± 0.06 | 76.28 ± 0.14 | 86.18 ± 0.16 | 83.27 ± 0.04 | 80.29 ± 0.05 | 91.72 ± 0.25 | 83.95 |
| OVANet | 78.42 ± 0.21 | 84.47 ± 0.43 | 88.3 ± 0.27 | 80.71 ± 0.41 | 84.84 ± 0.45 | 86.53 ± 0.11 | 64.85 ± 0.7 | 60.25 ± 0.53 | 84.74 ± 0.22 | 82.66 ± 0.17 | 77.01 ± 0.2 | 90.66 ± 0.49 | 80.29 |
| UniOT | 84.19 ± 0.27 | 90.91 ± 0.95 | 90.48 ± 0.58 | 83.97 ± 0.52 | 91.71 ± 0.97 | 88.47 ± 0.82 | 80.06 ± 1.1 | 82.36 ± 0.62 | 87.92 ± 0.26 | 85.26 ± 0.45 | 82.58 ± 0.95 | 91.76 ± 0.63 | 86.64 |
| WiSE-FT | 64.63 ± 0.04 | 82.12 ± 0.16 | 84.11 ± 0.03 | 60.82 ± 0.31 | 82.93 ± 0.25 | 79.64 ± 0.26 | 59.95 ± 0.38 | 54.5 ± 0.31 | 80.59 ± 0.1 | 71.29 ± 0.23 | 69.77 ± 0.27 | 90.9 ± 0.22 | 73.44 |
| CLIP cross-model | 77.8 ± 0.06 | 84.46 ± 0.06 | 87.64 ± 0.08 | 84.68 ± 0.1 | 87.19 ± 0.23 | 88.59 ± 0.09 | 81.77 ± 0.11 | 81.85 ± 0.1 | 89.5 ± 0.09 | 84.61 ± 0.09 | 79.31 ± 0.15 | 83.66 ± 0.1 | 84.26 |
| Our calibration | 80.99 ± 0.0 | 78.58 ± 0.0 | 84.97 ± 0.0 | 87.66 ± 0.0 | 81.09 ± 0.0 | 86.76 ± 0.0 | 86.46 ± 0.0 | 81.5 ± 0.0 | 92.56 ± 0.0 | 87.44 ± 0.0 | 82.25 ± 0.0 | 86.96 ± 0.0 | 84.77 |
| H3-score | |||||||||||||
| SO | 74.74 ± 0.06 | 85.58 ± 0.04 | 85.83 ± 0.09 | 77.2 ± 0.07 | 88.32 ± 0.14 | 86.28 ± 0.05 | 73.68 ± 0.17 | 75.51 ± 0.04 | 86.03 ± 0.14 | 77.99 ± 0.05 | 76.86 ± 0.02 | 85.71 ± 0.06 | 81.14 |
| DANCE | 77.71 ± 0.02 | 89.91 ± 0.02 | 86.95 ± 0.06 | 76.67 ± 0.06 | 90.04 ± 0.12 | 87.22 ± 0.06 | 72.09 ± 0.04 | 75.31 ± 0.09 | 85.82 ± 0.11 | 78.13 ± 0.02 | 77.87 ± 0.03 | 91.4 ± 0.17 | 82.43 |
| OVANet | 76.69 ± 0.13 | 86.47 ± 0.3 | 87.22 ± 0.18 | 76.61 ± 0.24 | 86.73 ± 0.31 | 86.06 ± 0.07 | 66.34 ± 0.49 | 64.09 ± 0.4 | 84.87 ± 0.15 | 77.77 ± 0.1 | 75.78 ± 0.13 | 90.69 ± 0.33 | 79.94 |
| UniOT | 80.45 ± 0.38 | 91.01 ± 0.54 | 89.65 ± 0.24 | 80.66 ± 0.8 | 91.68 ± 0.64 | 87.73 ± 0.51 | 78.7 ± 0.74 | 79.83 ± 0.34 | 87.94 ± 0.34 | 82.5 ± 0.48 | 79.68 ± 0.48 | 92.04 ± 0.43 | 85.16 |
| WiSE-FT | 67.32 ± 0.03 | 84.82 ± 0.11 | 84.45 ± 0.02 | 63.48 ± 0.22 | 85.39 ± 0.18 | 81.39 ± 0.18 | 62.84 ± 0.28 | 59.62 ± 0.24 | 82.05 ± 0.07 | 70.7 ± 0.15 | 70.96 ± 0.19 | 90.86 ± 0.15 | 75.32 |
| CLIP cross-model | 76.29 ± 0.04 | 86.47 ± 0.04 | 86.79 ± 0.06 | 78.96 ± 0.06 | 88.35 ± 0.16 | 87.41 ± 0.06 | 77.25 ± 0.06 | 78.84 ± 0.06 | 88.0 ± 0.06 | 78.92 ± 0.05 | 77.26 ± 0.09 | 85.9 ± 0.07 | 82.54 |
| Our calibration | 78.31 ± 0.0 | 82.26 ± 0.0 | 85.02 ± 0.0 | 80.66 ± 0.0 | 84.08 ± 0.0 | 86.21 ± 0.0 | 79.98 ± 0.0 | 78.63 ± 0.0 | 89.94 ± 0.0 | 80.54 ± 0.0 | 79.09 ± 0.0 | 88.19 ± 0.0 | 82.74 |
| UCR | |||||||||||||
| SO | 73.39 ± 0.06 | 89.19 ± 0.07 | 92.23 ± 0.02 | 87.93 ± 0.05 | 92.11 ± 0.15 | 90.87 ± 0.18 | 72.67 ± 0.26 | 74.93 ± 0.13 | 89.21 ± 0.07 | 81.48 ± 0.04 | 76.85 ± 0.16 | 93.48 ± 0.07 | 84.53 |
| DANCE | 80.41 ± 0.04 | 91.01 ± 0.22 | 92.63 ± 0.08 | 88.15 ± 0.09 | 93.8 ± 0.05 | 91.57 ± 0.1 | 75.77 ± 0.02 | 78.93 ± 0.19 | 91.13 ± 0.28 | 83.35 ± 0.21 | 80.17 ± 0.1 | 94.22 ± 0.05 | 86.76 |
| OVANet | 76.65 ± 0.18 | 92.89 ± 0.06 | 93.55 ± 0.15 | 87.56 ± 0.1 | 93.41 ± 0.2 | 90.89 ± 0.26 | 71.32 ± 0.23 | 77.95 ± 0.06 | 90.35 ± 0.07 | 81.78 ± 0.22 | 78.15 ± 0.1 | 94.09 ± 0.34 | 85.72 |
| UniOT | 79.4 ± 0.54 | 90.87 ± 2.05 | 94.6 ± 0.51 | 83.25 ± 0.88 | 93.25 ± 0.34 | 90.9 ± 0.92 | 76.83 ± 4.88 | 75.45 ± 1.72 | 89.5 ± 0.32 | 80.12 ± 1.47 | 78.79 ± 0.85 | 94.17 ± 0.73 | 85.59 |
| WiSE-FT | 78.15 ± 0.02 | 93.32 ± 0.04 | 95.79 ± 0.03 | 92.46 ± 0.05 | 95.31 ± 0.15 | 95.16 ± 0.06 | 83.67 ± 0.19 | 80.21 ± 0.02 | 93.99 ± 0.06 | 88.57 ± 0.13 | 79.71 ± 0.09 | 95.0 ± 0.0 | 89.28 |
| CLIP cross-model | 75.95 ± 0.1 | 91.07 ± 0.13 | 93.99 ± 0.06 | 90.54 ± 0.18 | 93.73 ± 0.1 | 93.36 ± 0.09 | 83.61 ± 0.07 | 79.94 ± 0.09 | 93.65 ± 0.08 | 85.53 ± 0.08 | 78.29 ± 0.04 | 94.15 ± 0.03 | 87.82 |
| CLIP zero-shot | 80.42 ± 0.0 | 93.53 ± 0.0 | 93.71 ± 0.0 | 91.03 ± 0.0 | 93.53 ± 0.0 | 93.71 ± 0.0 | 91.03 ± 0.0 | 80.42 ± 0.0 | 93.71 ± 0.0 | 91.03 ± 0.0 | 80.42 ± 0.0 | 93.53 ± 0.0 | 89.67 |
| Our calibration | 82.27 ± 0.0 | 94.52 ± 0.0 | 96.52 ± 0.0 | 93.32 ± 0.0 | 94.59 ± 0.0 | 96.56 ± 0.0 | 93.48 ± 0.0 | 82.29 ± 0.0 | 96.71 ± 0.0 | 93.45 ± 0.0 | 82.29 ± 0.0 | 94.81 ± 0.0 | 91.73 |
| Methods | P2R | P2S | R2P | R2S | S2P | S2R | Avg |
|---|---|---|---|---|---|---|---|
| H-score | |||||||
| SO | 72.15 ± 0.06 | 62.77 ± 0.1 | 62.5 ± 0.19 | 65.55 ± 0.09 | 56.61 ± 0.05 | 74.19 ± 0.03 | 65.63 |
| DANCE | 72.2 ± 0.07 | 62.16 ± 0.2 | 60.81 ± 0.06 | 65.86 ± 0.16 | 56.69 ± 0.01 | 73.75 ± 0.02 | 65.25 |
| OVANet | 77.09 ± 0.17 | 71.64 ± 0.04 | 69.57 ± 0.13 | 72.51 ± 0.1 | 67.04 ± 0.11 | 76.58 ± 0.11 | 72.41 |
| UniOT | 76.78 ± 0.08 | 70.35 ± 0.31 | 71.84 ± 0.05 | 73.56 ± 0.15 | 69.32 ± 0.36 | 77.43 ± 0.19 | 73.21 |
| WiSE-FT | 11.76 ± 0.11 | 4.52 ± 0.09 | 8.84 ± 0.11 | 7.73 ± 0.07 | 3.77 ± 0.08 | 10.9 ± 0.13 | 7.92 |
| CLIP cross-model | 74.9 ± 0.1 | 64.16 ± 0.03 | 62.16 ± 0.06 | 65.96 ± 0.05 | 58.29 ± 0.04 | 77.11 ± 0.08 | 67.1 |
| Our calibration | 71.59 ± 0.0 | 70.67 ± 0.0 | 76.39 ± 0.0 | 76.81 ± 0.0 | 73.88 ± 0.0 | 76.01 ± 0.0 | 74.22 |
| H3-score | |||||||
| SO | 75.38 ± 0.04 | 63.2 ± 0.07 | 64.39 ± 0.14 | 65.05 ± 0.06 | 60.09 ± 0.04 | 76.86 ± 0.02 | 67.5 |
| DANCE | 75.42 ± 0.05 | 62.78 ± 0.14 | 63.18 ± 0.04 | 65.25 ± 0.11 | 60.15 ± 0.01 | 76.54 ± 0.02 | 67.22 |
| OVANet | 78.9 ± 0.12 | 68.92 ± 0.03 | 69.22 ± 0.09 | 69.46 ± 0.06 | 67.53 ± 0.07 | 78.55 ± 0.08 | 72.1 |
| UniOT | 76.08 ± 0.08 | 67.15 ± 0.42 | 68.52 ± 0.1 | 69.74 ± 0.09 | 66.52 ± 0.07 | 76.99 ± 0.19 | 70.83 |
| WiSE-FT | 16.48 ± 0.15 | 6.55 ± 0.13 | 12.46 ± 0.15 | 10.93 ± 0.1 | 5.51 ± 0.12 | 15.34 ± 0.17 | 11.21 |
| CLIP cross-model | 77.36 ± 0.07 | 64.13 ± 0.02 | 64.15 ± 0.04 | 65.32 ± 0.03 | 61.35 ± 0.03 | 78.92 ± 0.05 | 68.54 |
| Our calibration | 74.97 ± 0.0 | 68.32 ± 0.0 | 73.58 ± 0.0 | 72.04 ± 0.0 | 72.01 ± 0.0 | 78.15 ± 0.0 | 73.18 |
| UCR | |||||||
| SO | 70.57 ± 0.02 | 62.09 ± 0.09 | 64.72 ± 0.15 | 65.34 ± 0.17 | 59.2 ± 0.21 | 74.13 ± 0.04 | 66.01 |
| DANCE | 72.14 ± 0.06 | 63.86 ± 0.13 | 65.17 ± 0.13 | 67.27 ± 0.02 | 66.4 ± 0.05 | 75.36 ± 0.03 | 68.37 |
| OVANet | 72.32 ± 0.16 | 62.88 ± 0.06 | 65.72 ± 0.18 | 65.36 ± 0.02 | 60.15 ± 0.14 | 75.22 ± 0.1 | 66.94 |
| UniOT | 73.72 ± 0.17 | 61.89 ± 0.29 | 64.05 ± 0.18 | 65.77 ± 0.19 | 62.91 ± 0.19 | 75.66 ± 0.22 | 67.33 |
| WiSE-FT | 77.55 ± 0.03 | 67.24 ± 0.08 | 68.76 ± 0.1 | 68.97 ± 0.16 | 67.45 ± 0.11 | 79.97 ± 0.02 | 71.66 |
| CLIP cross-model | 78.0 ± 0.09 | 67.6 ± 0.15 | 68.62 ± 0.16 | 68.86 ± 0.05 | 65.94 ± 0.11 | 80.14 ± 0.08 | 71.53 |
| CLIP zero-shot | 82.0 ± 0.0 | 68.37 ± 0.0 | 69.64 ± 0.0 | 68.37 ± 0.0 | 69.64 ± 0.0 | 82.0 ± 0.0 | 73.34 |
| Our calibration | 85.72 ± 0.0 | 71.97 ± 0.0 | 73.89 ± 0.0 | 72.07 ± 0.0 | 73.84 ± 0.0 | 85.78 ± 0.0 | 77.21 |
| Methods | A2D | A2W | D2A | D2W | W2A | W2D | Avg |
|---|---|---|---|---|---|---|---|
| H-score/H3-score | |||||||
| SO | 76.99 ± 0.54 | 71.98 ± 0.1 | 68.54 ± 0.19 | 97.64 ± 0.15 | 66.94 ± 0.25 | 99.26 ± 0.0 | 80.22 |
| DANCE | 73.29 ± 1.1 | 63.14 ± 0.16 | 62.46 ± 1.66 | 98.4 ± 0.0 | 57.63 ± 3.83 | 99.64 ± 0.32 | 75.76 |
| OVANet | 72.5 ± 0.73 | 64.35 ± 0.3 | 58.02 ± 1.13 | 96.26 ± 0.05 | 58.07 ± 0.66 | 98.67 ± 0.0 | 74.64 |
| UniOT | 58.03 ± 2.06 | 56.8 ± 1.14 | 53.8 ± 0.89 | 69.88 ± 1.5 | 53.23 ± 0.88 | 67.96 ± 3.44 | 59.95 |
| WiSE-FT | 31.89 ± 0.31 | 25.87 ± 0.19 | 39.04 ± 0.14 | 74.62 ± 0.26 | 36.65 ± 0.29 | 79.13 ± 0.09 | 47.87 |
| CLIP cross-model | 79.87 ± 0.45 | 77.27 ± 0.38 | 73.99 ± 0.08 | 97.46 ± 0.0 | 72.03 ± 0.05 | 98.65 ± 0.12 | 83.21 |
| Our calibration | 82.52 ± 0.0 | 82.41 ± 0.0 | 82.91 ± 0.0 | 85.71 ± 0.0 | 83.25 ± 0.0 | 85.97 ± 0.0 | 83.8 |
| UCR | |||||||
| SO | 93.37 ± 0.28 | 95.14 ± 0.26 | 81.03 ± 0.08 | 99.66 ± 0.06 | 79.65 ± 0.09 | 99.8 ± 0.0 | 91.44 |
| DANCE | 89.56 ± 0.28 | 85.7 ± 0.57 | 76.3 ± 0.32 | 99.75 ± 0.0 | 74.81 ± 0.92 | 100.0 ± 0.0 | 87.69 |
| OVANet | 93.31 ± 0.25 | 94.93 ± 0.16 | 81.04 ± 0.09 | 99.66 ± 0.06 | 79.64 ± 0.15 | 99.8 ± 0.0 | 91.4 |
| UniOT | 92.44 ± 0.96 | 94.17 ± 0.99 | 84.17 ± 0.56 | 98.78 ± 0.62 | 84.32 ± 0.3 | 98.93 ± 0.34 | 92.14 |
| WiSE-FT | 93.24 ± 0.19 | 92.91 ± 0.42 | 84.76 ± 0.07 | 98.99 ± 0.0 | 84.2 ± 0.1 | 99.67 ± 0.09 | 92.3 |
| CLIP cross-model | 94.98 ± 0.0 | 94.8 ± 0.16 | 85.56 ± 0.04 | 99.75 ± 0.0 | 84.52 ± 0.09 | 99.8 ± 0.0 | 93.23 |
| CLIP zero-shot | 88.15 ± 0.0 | 89.18 ± 0.0 | 85.73 ± 0.0 | 89.18 ± 0.0 | 85.73 ± 0.0 | 88.15 ± 0.0 | 87.69 |
| Our calibration | 88.15 ± 0.0 | 89.18 ± 0.0 | 85.73 ± 0.0 | 89.18 ± 0.0 | 85.73 ± 0.0 | 88.15 ± 0.0 | 87.69 |
| Methods | A2C | A2P | A2R | C2A | C2P | C2R | P2A | P2C | P2R | R2A | R2C | R2P | Avg |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| H-score/H3-score | |||||||||||||
| SO | 42.18 ± 0.02 | 65.78 ± 0.1 | 67.64 ± 0.01 | 46.06 ± 0.18 | 68.16 ± 0.16 | 64.49 ± 0.08 | 42.68 ± 0.12 | 37.5 ± 0.07 | 71.0 ± 0.07 | 58.19 ± 0.19 | 48.05 ± 0.09 | 85.69 ± 0.05 | 58.12 |
| DANCE | 39.45 ± 0.12 | 59.51 ± 1.16 | 68.27 ± 0.05 | 45.45 ± 0.17 | 58.08 ± 0.31 | 63.0 ± 0.1 | 36.2 ± 0.56 | 35.31 ± 0.1 | 68.49 ± 0.29 | 58.1 ± 0.47 | 47.07 ± 0.06 | 86.12 ± 0.02 | 55.42 |
| OVANet | 53.14 ± 0.16 | 67.63 ± 0.26 | 77.34 ± 0.07 | 61.55 ± 0.37 | 67.28 ± 0.31 | 73.51 ± 0.21 | 50.26 ± 0.2 | 40.58 ± 0.43 | 75.23 ± 0.19 | 68.72 ± 0.14 | 53.28 ± 0.23 | 87.32 ± 0.2 | 64.65 |
| UniOT | 52.4 ± 0.23 | 54.54 ± 0.9 | 64.91 ± 0.76 | 61.16 ± 0.31 | 66.6 ± 1.57 | 72.71 ± 0.73 | 47.04 ± 0.66 | 47.5 ± 0.23 | 59.05 ± 0.61 | 60.14 ± 0.58 | 54.32 ± 0.22 | 70.81 ± 1.06 | 59.27 |
| WiSE-FT | 5.98 ± 0.11 | 14.86 ± 0.05 | 17.47 ± 0.04 | 3.65 ± 0.12 | 9.79 ± 0.11 | 9.0 ± 0.13 | 9.35 ± 0.07 | 5.21 ± 0.08 | 26.2 ± 0.02 | 14.31 ± 0.11 | 8.7 ± 0.18 | 39.13 ± 0.07 | 13.64 |
| CLIP cross-model | 44.97 ± 0.19 | 73.26 ± 0.06 | 72.75 ± 0.08 | 52.76 ± 0.25 | 77.72 ± 0.06 | 71.95 ± 0.09 | 47.85 ± 0.1 | 41.82 ± 0.04 | 73.98 ± 0.09 | 58.98 ± 0.03 | 49.1 ± 0.09 | 86.71 ± 0.03 | 62.65 |
| Our calibration | 69.65 ± 0.0 | 92.95 ± 0.0 | 92.37 ± 0.0 | 81.08 ± 0.0 | 92.65 ± 0.0 | 91.97 ± 0.0 | 70.84 ± 0.0 | 59.19 ± 0.0 | 86.44 ± 0.0 | 75.15 ± 0.0 | 63.08 ± 0.0 | 91.53 ± 0.0 | 80.58 |
| UCR | |||||||||||||
| SO | 72.07 ± 0.11 | 87.95 ± 0.04 | 90.96 ± 0.04 | 82.19 ± 0.19 | 88.87 ± 0.06 | 89.79 ± 0.08 | 77.21 ± 0.15 | 69.28 ± 0.27 | 89.85 ± 0.09 | 85.55 ± 0.02 | 75.01 ± 0.05 | 93.86 ± 0.07 | 83.55 |
| DANCE | 71.78 ± 0.19 | 82.62 ± 0.23 | 90.46 ± 0.01 | 81.33 ± 0.17 | 86.07 ± 0.11 | 88.24 ± 0.15 | 72.35 ± 0.34 | 69.52 ± 0.17 | 87.54 ± 0.08 | 84.74 ± 0.1 | 74.08 ± 0.08 | 92.16 ± 0.08 | 81.74 |
| OVANet | 72.07 ± 0.12 | 87.99 ± 0.08 | 90.98 ± 0.06 | 82.01 ± 0.14 | 88.78 ± 0.14 | 89.79 ± 0.05 | 77.27 ± 0.07 | 69.14 ± 0.09 | 89.75 ± 0.08 | 85.51 ± 0.16 | 75.22 ± 0.3 | 93.77 ± 0.2 | 83.52 |
| UniOT | 75.48 ± 0.86 | 91.03 ± 0.24 | 92.17 ± 0.23 | 84.97 ± 0.91 | 90.74 ± 0.41 | 90.18 ± 0.09 | 79.06 ± 1.0 | 73.81 ± 0.14 | 90.98 ± 0.13 | 85.87 ± 0.29 | 77.72 ± 0.3 | 93.97 ± 0.39 | 85.5 |
| WiSE-FT | 76.3 ± 0.08 | 92.93 ± 0.03 | 93.61 ± 0.03 | 87.76 ± 0.09 | 92.36 ± 0.07 | 93.02 ± 0.02 | 82.98 ± 0.12 | 74.78 ± 0.08 | 92.4 ± 0.04 | 88.74 ± 0.19 | 77.6 ± 0.08 | 94.86 ± 0.02 | 87.28 |
| CLIP cross-model | 75.67 ± 0.08 | 93.52 ± 0.07 | 93.33 ± 0.01 | 86.24 ± 0.09 | 92.14 ± 0.02 | 92.33 ± 0.1 | 82.75 ± 0.07 | 74.43 ± 0.12 | 92.39 ± 0.05 | 88.31 ± 0.13 | 77.15 ± 0.12 | 95.17 ± 0.01 | 86.95 |
| CLIP zero-shot | 77.69 ± 0.0 | 94.32 ± 0.0 | 94.51 ± 0.0 | 89.82 ± 0.0 | 94.32 ± 0.0 | 94.51 ± 0.0 | 89.82 ± 0.0 | 77.69 ± 0.0 | 94.51 ± 0.0 | 89.82 ± 0.0 | 77.69 ± 0.0 | 94.32 ± 0.0 | 89.08 |
| Our calibration | 77.69 ± 0.0 | 94.32 ± 0.0 | 94.51 ± 0.0 | 89.82 ± 0.0 | 94.32 ± 0.0 | 94.51 ± 0.0 | 89.82 ± 0.0 | 77.69 ± 0.0 | 94.51 ± 0.0 | 89.82 ± 0.0 | 77.69 ± 0.0 | 94.32 ± 0.0 | 89.08 |
| Methods | P2R | P2S | R2P | R2S | S2P | S2R | Avg |
|---|---|---|---|---|---|---|---|
| H-score/H3-score | |||||||
| SO | 47.32 ± 0.07 | 31.92 ± 0.03 | 33.24 ± 0.06 | 35.42 ± 0.06 | 30.24 ± 0.04 | 51.46 ± 0.04 | 38.27 |
| DANCE | 47.09 ± 0.03 | 32.1 ± 0.11 | 31.62 ± 0.24 | 33.88 ± 0.14 | 29.89 ± 0.11 | 50.78 ± 0.04 | 37.56 |
| OVANet | 65.92 ± 0.1 | 51.21 ± 0.11 | 49.47 ± 0.07 | 54.56 ± 0.05 | 51.87 ± 0.19 | 70.27 ± 0.09 | 57.22 |
| UniOT | 71.17 ± 0.32 | 60.36 ± 0.17 | 57.09 ± 0.35 | 63.79 ± 0.17 | 56.98 ± 0.15 | 72.93 ± 0.07 | 63.72 |
| WiSE-FT | 0.56 ± 0.01 | 0.09 ± 0.0 | 0.47 ± 0.03 | 0.25 ± 0.01 | 0.11 ± 0.01 | 0.3 ± 0.01 | 0.3 |
| CLIP cross-model | 46.33 ± 0.11 | 29.96 ± 0.09 | 29.69 ± 0.12 | 32.42 ± 0.1 | 28.35 ± 0.06 | 50.46 ± 0.06 | 36.2 |
| Our calibration | 85.82 ± 0.0 | 72.1 ± 0.0 | 66.93 ± 0.0 | 70.96 ± 0.0 | 67.97 ± 0.0 | 85.78 ± 0.0 | 74.93 |
| UCR | |||||||
| SO | 76.06 ± 0.08 | 64.06 ± 0.23 | 69.55 ± 0.18 | 68.6 ± 0.04 | 67.67 ± 0.13 | 81.64 ± 0.08 | 71.26 |
| DANCE | 75.6 ± 0.11 | 66.47 ± 0.13 | 70.01 ± 0.03 | 68.07 ± 0.03 | 69.35 ± 0.12 | 80.44 ± 0.06 | 71.66 |
| OVANet | 76.15 ± 0.06 | 64.36 ± 0.23 | 69.65 ± 0.09 | 68.78 ± 0.05 | 67.66 ± 0.14 | 81.63 ± 0.05 | 71.37 |
| UniOT | 79.76 ± 0.04 | 68.2 ± 0.13 | 71.57 ± 0.2 | 71.12 ± 0.1 | 69.99 ± 0.04 | 82.45 ± 0.08 | 73.85 |
| WiSE-FT | 82.2 ± 0.08 | 69.95 ± 0.11 | 72.67 ± 0.08 | 71.79 ± 0.02 | 72.33 ± 0.11 | 85.51 ± 0.02 | 75.74 |
| CLIP cross-model | 82.4 ± 0.01 | 69.93 ± 0.16 | 72.48 ± 0.1 | 71.33 ± 0.02 | 71.61 ± 0.06 | 85.66 ± 0.04 | 75.57 |
| CLIP zero-shot | 88.5 ± 0.0 | 74.54 ± 0.0 | 75.27 ± 0.0 | 74.65 ± 0.0 | 75.27 ± 0.0 | 88.68 ± 0.0 | 79.48 |
| Our calibration | 88.5 ± 0.0 | 74.54 ± 0.0 | 75.27 ± 0.0 | 74.66 ± 0.0 | 75.27 ± 0.0 | 88.68 ± 0.0 | 79.49 |
| Methods | A2D | A2W | D2A | D2W | W2A | W2D | Avg |
|---|---|---|---|---|---|---|---|
| H-score/H3-score | |||||||
| SO | 86.4 ± 0.5 | 83.72 ± 0.24 | 89.13 ± 0.52 | 96.43 ± 0.22 | 84.28 ± 0.55 | 98.75 ± 0.0 | 89.79 |
| DANCE | 65.23 ± 6.21 | 43.33 ± 4.71 | 57.13 ± 8.0 | 90.0 ± 0.0 | 47.77 ± 7.84 | 97.5 ± 0.0 | 66.83 |
| OVANet | 91.42 ± 0.2 | 77.48 ± 2.74 | 81.96 ± 4.4 | 95.79 ± 0.11 | 81.63 ± 4.31 | 96.92 ± 0.0 | 87.53 |
| UniOT | 43.36 ± 1.88 | 36.25 ± 2.27 | 39.29 ± 1.02 | 42.39 ± 0.83 | 37.23 ± 2.93 | 49.34 ± 1.09 | 41.31 |
| WiSE-FT | 39.89 ± 0.39 | 28.34 ± 0.15 | 50.65 ± 0.36 | 76.36 ± 0.18 | 43.01 ± 0.44 | 83.15 ± 0.39 | 53.57 |
| CLIP cross-model | 89.62 ± 0.2 | 87.18 ± 0.18 | 93.65 ± 0.13 | 96.64 ± 0.0 | 91.03 ± 0.13 | 97.19 ± 0.2 | 92.55 |
| Our calibration | 93.32 ± 0.0 | 91.33 ± 0.0 | 95.81 ± 0.0 | 94.81 ± 0.0 | 96.01 ± 0.0 | 95.05 ± 0.0 | 94.39 |
| UCR | |||||||
| SO | 95.12 ± 0.6 | 97.63 ± 0.28 | 95.27 ± 0.05 | 99.44 ± 0.16 | 94.5 ± 0.13 | 100.0 ± 0.0 | 96.99 |
| DANCE | 76.01 ± 5.34 | 71.41 ± 1.31 | 76.55 ± 2.69 | 91.53 ± 0.0 | 71.71 ± 3.41 | 99.36 ± 0.0 | 81.09 |
| OVANet | 94.69 ± 0.6 | 97.18 ± 0.16 | 95.27 ± 0.05 | 99.44 ± 0.16 | 94.43 ± 0.18 | 100.0 ± 0.0 | 96.83 |
| UniOT | 47.56 ± 2.1 | 48.14 ± 1.73 | 61.97 ± 0.63 | 63.73 ± 2.0 | 58.21 ± 1.8 | 56.05 ± 0.52 | 55.94 |
| WiSE-FT | 96.18 ± 0.0 | 97.74 ± 0.16 | 96.31 ± 0.05 | 99.66 ± 0.0 | 95.93 ± 0.09 | 99.58 ± 0.3 | 97.57 |
| CLIP cross-model | 96.18 ± 0.0 | 97.63 ± 0.28 | 96.35 ± 0.0 | 99.66 ± 0.0 | 95.65 ± 0.05 | 100.0 ± 0.0 | 97.58 |
| CLIP zero-shot | 95.54 ± 0.0 | 97.63 ± 0.0 | 96.66 ± 0.0 | 97.63 ± 0.0 | 96.66 ± 0.0 | 95.54 ± 0.0 | 96.61 |
| Our calibration | 95.54 ± 0.0 | 97.63 ± 0.0 | 96.66 ± 0.0 | 97.63 ± 0.0 | 96.66 ± 0.0 | 95.54 ± 0.0 | 96.61 |
| Methods | A2C | A2P | A2R | C2A | C2P | C2R | P2A | P2C | P2R | R2A | R2C | R2P | Avg |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| H-score/H3-score | |||||||||||||
| SO | 44.33 ± 0.19 | 65.34 ± 0.35 | 71.4 ± 0.35 | 44.78 ± 0.1 | 59.45 ± 0.47 | 63.91 ± 0.17 | 46.61 ± 0.26 | 41.37 ± 0.14 | 72.02 ± 0.12 | 59.08 ± 0.17 | 53.28 ± 0.1 | 78.2 ± 0.12 | 58.31 |
| DANCE | 34.87 ± 0.17 | 41.08 ± 1.78 | 74.03 ± 1.62 | 31.93 ± 1.16 | 34.31 ± 0.22 | 54.84 ± 2.92 | 25.85 ± 1.81 | 21.6 ± 0.81 | 63.45 ± 2.65 | 57.55 ± 0.35 | 45.7 ± 0.79 | 74.36 ± 0.06 | 46.63 |
| OVANet | 60.09 ± 0.43 | 70.91 ± 0.3 | 81.93 ± 0.39 | 60.59 ± 0.48 | 61.88 ± 1.41 | 72.19 ± 0.26 | 52.46 ± 0.48 | 45.8 ± 1.0 | 76.8 ± 0.16 | 68.38 ± 0.23 | 58.47 ± 0.2 | 81.48 ± 0.52 | 65.92 |
| UniOT | 44.17 ± 0.19 | 43.91 ± 0.93 | 39.98 ± 1.07 | 47.21 ± 0.58 | 41.66 ± 2.14 | 47.94 ± 3.9 | 37.16 ± 1.5 | 44.8 ± 0.48 | 39.38 ± 1.05 | 48.13 ± 0.6 | 43.92 ± 0.44 | 44.92 ± 2.23 | 43.6 |
| WiSE-FT | 9.8 ± 0.04 | 17.69 ± 0.16 | 22.75 ± 0.08 | 6.1 ± 0.04 | 11.27 ± 0.08 | 13.4 ± 0.28 | 11.26 ± 0.09 | 7.26 ± 0.14 | 29.79 ± 0.09 | 17.6 ± 0.12 | 13.3 ± 0.28 | 38.55 ± 0.19 | 16.56 |
| CLIP cross-model | 48.18 ± 0.04 | 70.76 ± 0.25 | 75.41 ± 0.24 | 51.32 ± 0.17 | 70.49 ± 0.36 | 70.68 ± 0.13 | 52.32 ± 0.19 | 48.38 ± 0.12 | 75.23 ± 0.09 | 59.69 ± 0.2 | 55.33 ± 0.29 | 79.92 ± 0.18 | 63.14 |
| Our calibration | 73.32 ± 0.0 | 90.08 ± 0.0 | 92.55 ± 0.0 | 81.06 ± 0.0 | 89.85 ± 0.0 | 92.22 ± 0.0 | 70.63 ± 0.0 | 63.18 ± 0.0 | 87.26 ± 0.0 | 74.48 ± 0.0 | 66.75 ± 0.0 | 88.36 ± 0.0 | 80.81 |
| UCR | |||||||||||||
| SO | 79.38 ± 0.32 | 87.73 ± 0.24 | 91.94 ± 0.14 | 81.14 ± 0.38 | 81.94 ± 0.21 | 87.7 ± 0.33 | 78.6 ± 0.3 | 76.66 ± 0.22 | 91.33 ± 0.09 | 84.82 ± 0.16 | 81.93 ± 0.16 | 90.05 ± 0.1 | 84.44 |
| DANCE | 69.39 ± 0.34 | 75.16 ± 1.63 | 90.8 ± 0.4 | 72.39 ± 0.6 | 70.55 ± 0.52 | 82.33 ± 0.83 | 62.63 ± 0.81 | 61.95 ± 0.7 | 83.45 ± 0.18 | 80.38 ± 0.11 | 74.07 ± 0.35 | 84.48 ± 0.09 | 75.63 |
| OVANet | 79.5 ± 0.12 | 87.45 ± 0.12 | 91.81 ± 0.09 | 80.96 ± 0.3 | 81.89 ± 0.03 | 87.61 ± 0.29 | 78.51 ± 0.0 | 76.42 ± 0.15 | 91.22 ± 0.08 | 84.73 ± 0.17 | 82.41 ± 0.2 | 89.9 ± 0.44 | 84.37 |
| UniOT | 57.03 ± 1.24 | 58.43 ± 0.76 | 75.43 ± 1.72 | 64.19 ± 3.19 | 68.33 ± 2.08 | 79.72 ± 2.56 | 44.14 ± 0.74 | 55.84 ± 1.23 | 69.87 ± 1.01 | 63.7 ± 0.87 | 59.86 ± 0.61 | 66.76 ± 0.11 | 63.61 |
| WiSE-FT | 82.93 ± 0.21 | 91.69 ± 0.19 | 93.82 ± 0.12 | 88.03 ± 0.23 | 87.58 ± 0.23 | 91.9 ± 0.13 | 84.21 ± 0.2 | 82.07 ± 0.16 | 93.34 ± 0.03 | 88.77 ± 0.35 | 84.82 ± 0.14 | 92.12 ± 0.03 | 88.44 |
| CLIP cross-model | 82.37 ± 0.15 | 92.57 ± 0.12 | 93.12 ± 0.16 | 85.8 ± 0.09 | 86.87 ± 0.1 | 90.34 ± 0.05 | 84.27 ± 0.04 | 81.83 ± 0.32 | 93.58 ± 0.03 | 87.82 ± 0.09 | 84.3 ± 0.05 | 92.75 ± 0.12 | 87.97 |
| CLIP zero-shot | 81.61 ± 0.0 | 91.2 ± 0.0 | 94.53 ± 0.0 | 90.36 ± 0.0 | 91.2 ± 0.0 | 94.53 ± 0.0 | 90.36 ± 0.0 | 81.61 ± 0.0 | 94.53 ± 0.0 | 90.36 ± 0.0 | 81.61 ± 0.0 | 91.2 ± 0.0 | 89.43 |
| Our calibration | 81.61 ± 0.0 | 91.2 ± 0.0 | 94.53 ± 0.0 | 90.36 ± 0.0 | 91.2 ± 0.0 | 94.53 ± 0.0 | 90.36 ± 0.0 | 81.61 ± 0.0 | 94.53 ± 0.0 | 90.36 ± 0.0 | 81.61 ± 0.0 | 91.2 ± 0.0 | 89.43 |
| Methods | P2R | P2S | R2P | R2S | S2P | S2R | Avg |
|---|---|---|---|---|---|---|---|
| H-score/H3-score | |||||||
| SO | 44.52 ± 0.09 | 31.86 ± 0.16 | 32.39 ± 0.02 | 34.58 ± 0.12 | 25.71 ± 0.11 | 46.24 ± 0.26 | 35.88 |
| DANCE | 42.13 ± 0.74 | 28.66 ± 0.23 | 25.54 ± 0.33 | 27.51 ± 0.41 | 21.75 ± 0.42 | 39.93 ± 0.71 | 30.92 |
| OVANet | 65.01 ± 0.07 | 51.31 ± 0.26 | 48.81 ± 0.22 | 54.76 ± 0.22 | 47.84 ± 0.25 | 67.4 ± 0.19 | 55.85 |
| UniOT | 60.68 ± 0.59 | 54.95 ± 0.29 | 50.63 ± 0.31 | 56.8 ± 0.05 | 47.35 ± 0.97 | 60.69 ± 0.56 | 55.18 |
| WiSE-FT | 0.57 ± 0.01 | 0.06 ± 0.01 | 0.52 ± 0.04 | 0.32 ± 0.02 | 0.07 ± 0.01 | 0.2 ± 0.01 | 0.29 |
| CLIP cross-model | 43.44 ± 0.12 | 29.73 ± 0.18 | 28.59 ± 0.05 | 31.68 ± 0.23 | 24.92 ± 0.05 | 46.01 ± 0.03 | 34.06 |
| Our calibration | 85.69 ± 0.0 | 73.22 ± 0.0 | 67.02 ± 0.0 | 71.65 ± 0.0 | 68.36 ± 0.0 | 85.4 ± 0.0 | 75.22 |
| UCR | |||||||
| SO | 74.92 ± 0.05 | 63.6 ± 0.4 | 68.76 ± 0.24 | 70.77 ± 0.11 | 67.52 ± 0.11 | 79.14 ± 0.15 | 70.78 |
| DANCE | 72.51 ± 0.15 | 63.94 ± 0.37 | 64.22 ± 0.24 | 64.88 ± 0.31 | 64.98 ± 0.06 | 75.05 ± 0.17 | 67.6 |
| OVANet | 74.9 ± 0.05 | 64.19 ± 0.25 | 68.84 ± 0.13 | 71.06 ± 0.07 | 67.52 ± 0.27 | 79.11 ± 0.08 | 70.94 |
| UniOT | 75.12 ± 0.17 | 64.5 ± 0.34 | 64.31 ± 0.24 | 67.24 ± 0.32 | 61.14 ± 0.89 | 75.63 ± 0.27 | 67.99 |
| WiSE-FT | 81.23 ± 0.08 | 70.46 ± 0.23 | 72.08 ± 0.11 | 74.16 ± 0.09 | 72.95 ± 0.09 | 83.74 ± 0.02 | 75.77 |
| CLIP cross-model | 81.34 ± 0.08 | 70.44 ± 0.21 | 72.06 ± 0.14 | 73.51 ± 0.06 | 71.97 ± 0.19 | 84.0 ± 0.11 | 75.55 |
| CLIP zero-shot | 87.99 ± 0.0 | 76.44 ± 0.0 | 75.2 ± 0.0 | 76.41 ± 0.0 | 75.2 ± 0.0 | 87.98 ± 0.0 | 79.87 |
| Our calibration | 88.0 ± 0.0 | 76.44 ± 0.0 | 75.2 ± 0.0 | 76.41 ± 0.0 | 75.2 ± 0.0 | 87.99 ± 0.0 | 79.87 |
References
- Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 18–24 July 2021; pp. 8748–8763. [Google Scholar]
- Jia, C.; Yang, Y.; Xia, Y.; Chen, Y.T.; Parekh, Z.; Pham, H.; Le, Q.; Sung, Y.H.; Li, Z.; Duerig, T. Scaling up visual and vision-language representation learning with noisy text supervision. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 18–24 July 2021; pp. 4904–4916. [Google Scholar]
- Oquab, M.; Darcet, T.; Moutakanni, T.; Vo, H.; Szafraniec, M.; Khalidov, V.; Fernandez, P.; Haziza, D.; Massa, F.; El-Nouby, A.; et al. DINOv2: Learning Robust Visual Features without Supervision. arXiv 2023, arXiv:2304.07193. [Google Scholar] [CrossRef]
- Bommasani, R.; Hudson, D.A.; Adeli, E.; Altman, R.; Arora, S.; von Arx, S.; Bernstein, M.S.; Bohg, J.; Bosselut, A.; Brunskill, E.; et al. On the opportunities and risks of foundation models. arXiv 2021, arXiv:2108.07258. [Google Scholar] [CrossRef]
- Lin, Z.; Yu, S.; Kuang, Z.; Pathak, D.; Ramanan, D. Multimodality helps unimodality: Cross-modal few-shot learning with multimodal models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 19325–19337. [Google Scholar]
- Esmaeilpour, S.; Liu, B.; Robertson, E.; Shu, L. Zero-shot out-of-distribution detection based on the pre-trained model clip. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 22 February–1 March 2022. [Google Scholar]
- Wortsman, M.; Ilharco, G.; Kim, J.W.; Li, M.; Kornblith, S.; Roelofs, R.; Lopes, R.G.; Hajishirzi, H.; Farhadi, A.; Namkoong, H.; et al. Robust fine-tuning of zero-shot models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 7959–7971. [Google Scholar]
- Lee, Y.; Chen, A.S.; Tajwar, F.; Kumar, A.; Yao, H.; Liang, P.; Finn, C. Surgical Fine-Tuning Improves Adaptation to Distribution Shifts. In Proceedings of the Eleventh International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
- You, K.; Long, M.; Cao, Z.; Wang, J.; Jordan, M.I. Universal domain adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2720–2729. [Google Scholar]
- Pan, S.J.; Tsang, I.W.; Kwok, J.T.; Yang, Q. Domain Adaptation via Transfer Component Analysis. IEEE Trans. Neural Netw. 2011, 22, 199–210. [Google Scholar] [CrossRef] [PubMed]
- Sun, B.; Saenko, K. Deep Coral: Correlation Alignment for Deep Domain Adaptation. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 443–450. [Google Scholar]
- Ganin, Y.; Ustinova, E.; Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; Marchand, M.; Lempitsky, V. Domain-adversarial training of neural networks. J. Mach. Learn. Res. 2016, 17, 2030–2096. [Google Scholar]
- Long, M.; Cao, Z.; Wang, J.; Jordan, M.I. Conditional Adversarial Domain Adaptation. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; pp. 1640–1650. [Google Scholar]
- Saito, K.; Watanabe, K.; Ushiku, Y.; Harada, T. Maximum Classifier Discrepancy for Unsupervised Domain Adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3723–3732. [Google Scholar]
- Zhang, Y.; Deng, B.; Tang, H.; Zhang, L.; Jia, K. Unsupervised Multi-Class Domain Adaptation: Theory, Algorithms, and Practice. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 2775–2792. [Google Scholar] [CrossRef] [PubMed]
- Shi, L.; Liu, W. Adversarial Self-Training Improves Robustness and Generalization for Gradual Domain Adaptation. In Proceedings of the Thirty-Seventh Conference on Neural Information Processing Systems, New Orleans, LA, USA, 10–16 December 2023. [Google Scholar]
- Ma, W.; Zhang, J.; Li, S.; Liu, C.H.; Wang, Y.; Li, W. Exploiting Both Domain-specific and Invariant Knowledge via a Win-Win Transformer for Unsupervised Domain Adaptation. arXiv 2021, arXiv:2111.12941. [Google Scholar] [CrossRef]
- Xu, T.; Chen, W.; Wang, P.; Wang, F.; Li, H.; Jin, R. CDTrans: Cross-Domain Transformer for Unsupervised Domain Adaptation. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 4 May 2021. [Google Scholar]
- Sun, T.; Lu, C.; Zhang, T.; Ling, H. Safe Self-Refinement for Transformer-Based Domain Adaptation. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 7181–7190. [Google Scholar]
- Ren, C.-X.; Zhai, Y.M.; Luo, Y.W.; Yan, H. Towards Unsupervised Domain Adaptation via Domain-Transformer. Int. J. Comput. Vis. 2024, 132, 6163–6183. [Google Scholar] [CrossRef]
- Fu, B.; Cao, Z.; Long, M.; Wang, J. Learning to detect open classes for universal domain adaptation. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XV 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 567–583. [Google Scholar]
- Saito, K.; Kim, D.; Sclaroff, S.; Saenko, K. Universal domain adaptation through self supervision. In Proceedings of the Advances in Neural Information Processing Systems, Virtual, 6–12 December 2020. [Google Scholar]
- Li, G.; Kang, G.; Zhu, Y.; Wei, Y.; Yang, Y. Domain consensus clustering for universal domain adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 9757–9766. [Google Scholar]
- Chen, L.; Du, Q.; Lou, Y.; He, J.; Bai, T.; Deng, M. Mutual nearest neighbor contrast and hybrid prototype self-training for universal domain adaptation. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 22 February–1 March 2022. [Google Scholar]
- Saito, K.; Saenko, K. Ovanet: One-vs-all network for universal domain adaptation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 9000–9009. [Google Scholar]
- Chang, W.; Shi, Y.; Tuan, H.; Wang, J. Unified optimal transport framework for universal domain adaptation. In Proceedings of the Advances in Neural Information Processing Systems, New Orleans, LA, USA, 28 November–9 December 2022. [Google Scholar]
- Xu, Q.; Zhang, R.; Zhang, Y.; Wang, Y.; Tian, Q. A fourier-based framework for domain generalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 14383–14392. [Google Scholar]
- Zhou, K.; Yang, Y.; Qiao, Y.; Xiang, T. Domain Generalization with MixStyle. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 4 May 2021. [Google Scholar]
- Zhang, X.; Cui, P.; Xu, R.; Zhou, L.; He, Y.; Shen, Z. Deep stable learning for out-of-distribution generalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 5372–5382. [Google Scholar]
- Geirhos, R.; Rubisch, P.; Michaelis, C.; Bethge, M.; Wichmann, F.A.; Brendel, W. ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
- Wang, H.; Ge, S.; Lipton, Z.; Xing, E.P. Learning robust global representations by penalizing local predictive power. In Proceedings of the Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Volume 32. [Google Scholar]
- Gulrajani, I.; Lopez-Paz, D. In Search of Lost Domain Generalization. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 30 April 2020. [Google Scholar]
- Wiles, O.; Gowal, S.; Stimberg, F.; Rebuffi, S.A.; Ktena, I.; Dvijotham, K.D.; Cemgil, A.T. A Fine-Grained Analysis on Distribution Shift. In Proceedings of the International Conference on Learning Representations, Virtual, 25 April 2022. [Google Scholar]
- He, K.; Chen, X.; Xie, S.; Li, Y.; Dollár, P.; Girshick, R. Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 16000–16009. [Google Scholar]
- Neyshabur, B.; Sedghi, H.; Zhang, C. What is being transferred in transfer learning? In Proceedings of the Advances in Neural Information Processing Systems, Virtual, 6–12 December 2020. [Google Scholar]
- Zhou, K.; Yang, J.; Loy, C.C.; Liu, Z. Learning to prompt for vision-language models. Int. J. Comput. Vis. 2022, 130, 2337–2348. [Google Scholar] [CrossRef]
- Zhou, K.; Yang, J.; Loy, C.C.; Liu, Z. Conditional prompt learning for vision-language models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 16816–16825. [Google Scholar]
- Gu, J.; Beirami, A.; Wang, X.; Beutel, A.; Torr, P.; Qin, Y. Towards Robust Prompts on Vision-Language Models. arXiv 2023, arXiv:2304.08479. [Google Scholar] [CrossRef]
- Gao, P.; Geng, S.; Zhang, R.; Ma, T.; Fang, R.; Zhang, Y.; Li, H.; Qiao, Y. Clip-adapter: Better vision-language models with feature adapters. arXiv 2021, arXiv:2110.04544. [Google Scholar] [CrossRef]
- Zhang, R.; Zhang, W.; Fang, R.; Gao, P.; Li, K.; Dai, J.; Qiao, Y.; Li, H. Tip-adapter: Training-free adaption of clip for few-shot classification. In Proceedings of the Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, 23–27 October 2022; Proceedings, Part XXXV. Springer: Berlin/Heidelberg, Germany, 2022; pp. 493–510. [Google Scholar]
- Scheirer, W.J.; de Rezende Rocha, A.; Sapkota, A.; Boult, T.E. Toward Open Set Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1757–1772. [Google Scholar] [CrossRef] [PubMed]
- Hendrycks, D.; Gimpel, K. A baseline for detecting misclassified and out-of-distribution examples in neural networks. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
- Fang, Z.; Li, Y.; Lu, J.; Dong, J.; Han, B.; Liu, F. Is Out-of-Distribution Detection Learnable? In Proceedings of the Advances in Neural Information Processing Systems, New Orleans, LA, USA, 28 November–9 December 2022. [Google Scholar]
- Vaze, S.; Han, K.; Vedaldi, A.; Zisserman, A. Open-Set Recognition: A Good Closed-Set Classifier is All You Need. In Proceedings of the International Conference on Learning Representations, Virtual, 25 April 2022. [Google Scholar]
- DeGroot, M.H.; Fienberg, S.E. The comparison and evaluation of forecasters. J. R. Stat. Soc. Ser. D 1983, 32, 12–22. [Google Scholar] [CrossRef]
- Guo, C.; Pleiss, G.; Sun, Y.; Weinberger, K.Q. On calibration of modern neural networks. In Proceedings of the International Conference on Machine Learning, PMLR, Sydney, Australia, 6–11 August 2017; pp. 1321–1330. [Google Scholar]
- Naeini, M.P.; Cooper, G.; Hauskrecht, M. Obtaining well calibrated probabilities using bayesian binning. In Proceedings of the AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015. [Google Scholar]
- LeVine, W.; Pikus, B.; Raja, P.V.; Amat, F. Enabling Calibration In The Zero-Shot Inference of Large Vision-Language Models. In Proceedings of the ICLR 2023 Workshop on Pitfalls of Limited Data and Computation for Trustworthy ML, Kigali, Rwanda, 5 May 2023. [Google Scholar]
- Hastie, T.; Tibshirani, R.; Friedman, J.H.; Friedman, J.H. The Elements of Statistical Learning: Data Mining, Inference, and Prediction; Springer: Berlin/Heidelberg, Germany, 2009; Volume 2. [Google Scholar]
- Dhamija, A.R.; Günther, M.; Boult, T. Reducing network agnostophobia. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018. [Google Scholar]
- Phillips, P.J.; Grother, P.; Micheals, R. Evaluation Methods in Face Recognition. In Handbook of Face Recognition; Springer: London, UK, 2011; pp. 551–574. [Google Scholar]
- Saenko, K.; Kulis, B.; Fritz, M.; Darrell, T. Adapting visual category models to new domains. In Proceedings of the European Conference on Computer Vision, Heraklion, Greece, 5–11 September 2010; Springer: Berlin/Heidelberg, Germany, 2010; pp. 213–226. [Google Scholar]
- Venkateswara, H.; Eusebio, J.; Chakraborty, S.; Panchanathan, S. Deep hashing network for unsupervised domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5018–5027. [Google Scholar]
- Peng, X.; Usman, B.; Kaushik, N.; Hoffman, J.; Wang, D.; Saenko, K. Visda: The visual domain adaptation challenge. arXiv 2017, arXiv:1710.06924. [Google Scholar] [CrossRef]
- Peng, X.; Bai, Q.; Xia, X.; Huang, Z.; Saenko, K.; Wang, B. Moment matching for multi-source domain adaptation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1406–1415. [Google Scholar]
- Goyal, S.; Kumar, A.; Garg, S.; Kolter, Z.; Raghunathan, A. Finetune like you pretrain: Improved finetuning of zero-shot vision models. arXiv 2022, arXiv:2212.00638. [Google Scholar] [CrossRef]
- Strehl, A.; Ghosh, J. Cluster ensembles—A knowledge reuse framework for combining multiple partitions. J. Mach. Learn. Res. 2002, 3, 583–617. [Google Scholar]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]


| Method Type | Methods | Source | Target | Classifier | Scoring Rule | Threshold Value |
|---|---|---|---|---|---|---|
| Baseline | Source Only (SO) | ✓ | ✗ | softmax | negative entropy | |
| UniDA SOTAs | DANCE [22] | ✓ | ✓ | softmax | negative entropy | |
| OVANet [25] | ✓ | ✓ | softmax | binary softmax prob. | 1/2 | |
| UniOT [26] | ✓ | ✓ | OT | maximum OT mass | ||
| CLIP adaptations | WiSE-FT [7] | ✓ | ✗ | softmax | negative entropy | |
| CLIP cross-model [5] | ✓ | ✗ | softmax | negative entropy | ||
| CLIP zero-shot [1] | ✗ | ✗ | NN | maximum logit | - | |
| Calibration | CLIP calibration [48] | ✓ * | ✗ | softmax | negative entropy | |
| Our calibration | ✓ * | ✗ | softmax | negative entropy |
| Methods | Office | OfficeHome | VisDA | DomainNet | Avg | ||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| (10/10) | (10/0) | (31/0) | (10/21) | (10/5) | (15/0) | (65/0) | (25/40) | (6/3) | (6/0) | (12/0) | (6/6) | (150/50) | (150/0) | (345/0) | (150/195) | ||
| SO | 91.98 | 91.87 | 80.22 | 89.79 | 84.52 | 82.05 | 58.12 | 58.31 | 69.85 | 75.79 | 55.31 | 57.19 | 61.49 | 65.63 | 38.27 | 35.88 | 68.52 |
| DANCE [22] | 94.7 | 96.09 | 75.76 | 66.83 | 89.01 | 83.95 | 55.42 | 46.63 | 71.9 | 74.3 | 58.08 | 49.5 | 60.53 | 65.24 | 37.56 | 30.92 | 66.03 |
| OVANet [25] | 93.36 | 91.16 | 74.64 | 87.53 | 85.42 | 80.29 | 64.65 | 65.92 | 59.47 | 39.27 | 43.55 | 42.58 | 70.7 | 72.4 | 57.22 | 55.86 | 67.75 |
| UniOT [26] | 92.32 | 96.48 | 59.95 | 41.31 | 89.45 | 86.64 | 59.27 | 43.6 | 79.1 | 83.08 | 71.62 | 62.03 | 71.42 | 73.21 | 63.72 | 55.18 | 70.52 |
| WiSE-FT [7] | 82.34 | 94.07 | 47.87 | 53.57 | 79.37 | 73.44 | 13.64 | 16.56 | 62.68 | 72.21 | 30.05 | 27.4 | 3.74 | 7.92 | 0.3 | 0.29 | 41.59 |
| CLIP cross-model [5] | 93.04 | 93.59 | 83.21 | 92.55 | 86.2 | 84.26 | 62.65 | 63.14 | 77.69 | 81.66 | 62.08 | 67.98 | 61.98 | 67.1 | 36.2 | 34.06 | 71.71 |
| w/o calibration () | 0.0 | 0.07 | 0.0 | 0.0 | 0.17 | 1.31 | 0.0 | 0.0 | 0.0 | 0.05 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.1 |
| CLIP calibration [48] | 35.18 | 38.72 | 87.53 | 96.84 | 47.02 | 46.9 | 87.01 | 87.1 | 74.49 | 46.53 | 82.2 | 87.87 | 58.57 | 58.78 | 77.57 | 78.28 | 68.16 |
| Our calibration | 86.74 | 91.89 | 83.8 | 94.39 | 86.4 | 84.77 | 80.58 | 80.81 | 84.74 | 82.26 | 76.08 | 83.12 | 72.37 | 74.22 | 74.93 | 75.22 | 82.02 |
| Methods | Office | OfficeHome | VisDA | DomainNet | Avg | ||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| (10/10) | (10/0) | (31/0) | (10/21) | (10/5) | (15/0) | (65/0) | (25/40) | (6/3) | (6/0) | (12/0) | (6/6) | (150/50) | (150/0) | (345/0) | (150/195) | ||
| SO | 93.98 | 94.95 | 91.44 | 96.99 | 86.89 | 84.53 | 83.55 | 84.44 | 63.46 | 71.17 | 76.66 | 79.51 | 63.19 | 66.01 | 71.26 | 70.78 | 79.93 |
| DANCE [22] | 95.17 | 97.09 | 87.69 | 81.1 | 90.33 | 86.76 | 81.74 | 75.63 | 57.78 | 63.45 | 67.86 | 56.4 | 64.88 | 68.37 | 71.66 | 67.6 | 75.84 |
| OVANet [25] | 95.36 | 95.8 | 91.4 | 96.84 | 88.18 | 85.72 | 83.52 | 84.37 | 68.57 | 66.45 | 76.68 | 79.58 | 64.3 | 66.94 | 71.37 | 70.94 | 80.38 |
| UniOT [26] | 90.62 | 97.2 | 92.14 | 55.94 | 88.85 | 85.59 | 85.5 | 63.61 | 72.22 | 78.8 | 84.79 | 74.78 | 62.88 | 67.33 | 73.85 | 67.99 | 77.63 |
| WiSE-FT [7] | 95.27 | 96.33 | 92.3 | 97.57 | 90.77 | 89.28 | 87.28 | 88.44 | 70.83 | 77.88 | 81.45 | 84.43 | 68.72 | 71.66 | 75.74 | 75.77 | 83.98 |
| CLIP cross-model [5] | 95.38 | 96.18 | 93.24 | 97.58 | 89.71 | 87.82 | 86.95 | 87.97 | 73.22 | 79.06 | 81.15 | 83.76 | 68.81 | 71.53 | 75.57 | 75.55 | 83.97 |
| CLIP zero-shot [1] | 90.1 | 97.68 | 87.69 | 96.61 | 90.21 | 89.67 | 89.08 | 89.43 | 78.6 | 82.86 | 87.56 | 88.1 | 70.78 | 73.34 | 79.48 | 79.87 | 85.69 |
| w/o calibration () | 92.46 | 97.75 | 87.68 | 96.61 | 92.91 | 91.71 | 89.08 | 89.41 | 80.9 | 85.75 | 87.56 | 88.1 | 69.2 | 72.81 | 79.48 | 79.88 | 86.33 |
| CLIP calibration [48] | 90.88 | 97.55 | 87.69 | 96.61 | 92.24 | 90.59 | 89.08 | 89.43 | 82.8 | 86.32 | 87.56 | 88.1 | 74.57 | 76.88 | 79.49 | 79.87 | 86.85 |
| Our calibration | 93.39 | 97.91 | 87.69 | 96.61 | 93.02 | 91.73 | 89.08 | 89.43 | 82.38 | 86.37 | 87.56 | 88.1 | 74.86 | 77.21 | 79.49 | 79.87 | 87.17 |
| Methods | Office | OfficeHome | VisDA | DomainNet | Avg | ||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| (10/10) | (10/0) | (31/0) | (10/21) | (10/5) | (15/0) | (65/0) | (25/40) | (6/3) | (6/0) | (12/0) | (6/6) | (150/50) | (150/0) | (345/0) | (150/195) | ||
| H-score | |||||||||||||||||
| Only IID calibration | 35.18 | 38.72 | 87.53 | 96.84 | 47.02 | 46.9 | 87.01 | 87.1 | 74.49 | 46.53 | 82.2 | 87.87 | 58.57 | 58.78 | 77.57 | 78.28 | 68.16 |
| Only NLL calibration | 36.7 | 66.24 | 87.01 | 96.48 | 53.41 | 53.27 | 87.33 | 87.38 | 72.45 | 45.81 | 85.52 | 89.84 | 58.24 | 59.21 | 77.59 | 78.3 | 70.92 |
| Only OOD calibration | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
| w/o IID calibration | 90.11 | 76.44 | 77.8 | 89.3 | 87.61 | 84.68 | 72.46 | 73.4 | 82.33 | 58.11 | 58.03 | 62.25 | 77.38 | 79.01 | 70.63 | 70.17 | 75.61 |
| w/o NLL calibration | 88.63 | 73.38 | 75.33 | 86.84 | 87.39 | 84.03 | 66.58 | 68.01 | 78.85 | 54.6 | 54.96 | 58.32 | 77.38 | 79.14 | 67.66 | 66.66 | 72.98 |
| w/o OOD calibration | 35.36 | 55.43 | 87.01 | 96.48 | 56.4 | 56.23 | 87.37 | 87.42 | 72.69 | 45.53 | 82.3 | 87.99 | 59.59 | 58.65 | 77.59 | 78.28 | 70.27 |
| Ours | 86.74 | 91.89 | 83.8 | 94.39 | 86.4 | 84.77 | 80.58 | 80.81 | 84.74 | 82.26 | 76.08 | 83.12 | 72.37 | 74.22 | 74.93 | 75.22 | 82.02 |
| UCR | |||||||||||||||||
| Only IID calibration | 90.88 | 97.55 | 87.69 | 96.61 | 92.24 | 90.59 | 89.08 | 89.43 | 82.8 | 86.32 | 87.56 | 88.1 | 74.57 | 76.88 | 79.49 | 79.87 | 86.85 |
| Only NLL calibration | 90.52 | 97.7 | 87.69 | 96.61 | 92.39 | 90.78 | 89.08 | 89.43 | 82.83 | 86.31 | 87.56 | 88.1 | 74.57 | 76.89 | 79.49 | 79.87 | 86.86 |
| Only OOD calibration | 86.98 | 96.86 | 87.69 | 96.61 | 91.69 | 90.49 | 89.08 | 89.43 | 78.91 | 84.55 | 87.56 | 88.1 | 57.83 | 62.48 | 79.49 | 79.87 | 84.23 |
| w/o IID calibration | 93.51 | 97.88 | 87.69 | 96.61 | 93.07 | 91.84 | 89.08 | 89.43 | 82.2 | 86.16 | 87.56 | 88.1 | 74.92 | 77.26 | 79.49 | 79.87 | 87.17 |
| w/o NLL calibration | 93.51 | 97.88 | 87.69 | 96.61 | 93.08 | 91.85 | 89.08 | 89.43 | 82.1 | 86.15 | 87.56 | 88.1 | 74.9 | 77.25 | 79.49 | 79.87 | 87.16 |
| w/o OOD calibration | 90.38 | 97.65 | 87.69 | 96.61 | 92.45 | 90.87 | 89.08 | 89.43 | 82.83 | 86.31 | 87.56 | 88.1 | 74.6 | 76.87 | 79.49 | 79.87 | 86.86 |
| Ours | 93.39 | 97.91 | 87.69 | 96.61 | 93.02 | 91.73 | 89.08 | 89.43 | 82.38 | 86.37 | 87.56 | 88.1 | 74.86 | 77.21 | 79.49 | 79.87 | 87.17 |
| Methods | Office | OfficeHome | VisDA | DomainNet | Avg | ||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| (10/10) | (10/0) | (31/0) | (10/21) | (10/5) | (15/0) | (65/0) | (25/40) | (6/3) | (6/0) | (12/0) | (6/6) | (150/50) | (150/0) | (345/0) | (150/195) | ||
| H-score | |||||||||||||||||
| 0.1 | 17.79 | 27.94 | 87.53 | 96.84 | 20.31 | 20.3 | 88.35 | 88.41 | 32.13 | 41.24 | 89.6 | 92.06 | 4.55 | 6.13 | 79.34 | 80.1 | 54.54 |
| 0.2 | 62.28 | 71.5 | 87.27 | 96.69 | 60.49 | 60.33 | 86.75 | 86.9 | 67.96 | 71.96 | 84.37 | 89.21 | 62.49 | 63.83 | 77.47 | 78.17 | 75.48 |
| 0.3 | 83.28 | 90.45 | 84.7 | 95.16 | 81.09 | 80.49 | 80.06 | 80.24 | 83.41 | 86.6 | 75.01 | 82.37 | 77.54 | 79.58 | 63.02 | 61.75 | 80.3 |
| 0.4 | 90.45 | 95.51 | 75.13 | 88.0 | 88.95 | 87.09 | 62.12 | 64.09 | 83.03 | 87.32 | 62.15 | 67.33 | 56.51 | 63.38 | 26.64 | 26.75 | 70.28 |
| 0.5 | 80.04 | 93.82 | 50.79 | 55.82 | 88.18 | 84.77 | 32.85 | 38.97 | 68.3 | 80.4 | 39.98 | 40.39 | 15.82 | 24.23 | 2.61 | 3.04 | 50.0 |
| 0.6 | 59.36 | 81.49 | 24.95 | 25.91 | 78.42 | 73.0 | 7.9 | 12.97 | 39.74 | 63.18 | 14.31 | 12.63 | 1.61 | 3.1 | 0.1 | 0.14 | 31.18 |
| 0.7 | 33.69 | 64.05 | 6.76 | 3.35 | 57.83 | 50.18 | 1.07 | 2.26 | 10.95 | 34.87 | 3.2 | 1.47 | 0.05 | 0.16 | 0.0 | 0.01 | 16.87 |
| 0.8 | 2.07 | 41.75 | 0.38 | 0.0 | 26.6 | 22.82 | 0.13 | 0.31 | 1.15 | 8.61 | 0.07 | 0.06 | 0.01 | 0.01 | 0.0 | 0.0 | 6.5 |
| 0.9 | 0.0 | 8.68 | 0.03 | 0.0 | 4.64 | 5.09 | 0.0 | 0.02 | 0.0 | 1.38 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 1.24 |
| 1.0 | 0.0 | 0.07 | 0.0 | 0.0 | 0.17 | 1.31 | 0.0 | 0.0 | 0.0 | 0.05 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.1 |
| Ours | 86.74 | 91.89 | 83.8 | 94.39 | 86.4 | 84.77 | 80.58 | 80.81 | 84.74 | 82.26 | 76.08 | 83.12 | 72.37 | 74.22 | 74.93 | 75.22 | 82.02 |
| UCR | |||||||||||||||||
| 0.1 | 90.14 | 97.49 | 87.69 | 96.61 | 91.72 | 89.85 | 89.08 | 89.43 | 82.98 | 86.23 | 87.56 | 88.1 | 72.3 | 74.72 | 79.49 | 79.87 | 86.45 |
| 0.2 | 92.22 | 97.73 | 87.69 | 96.61 | 92.5 | 90.94 | 89.08 | 89.43 | 82.89 | 86.61 | 87.56 | 88.1 | 74.67 | 77.0 | 79.49 | 79.87 | 87.02 |
| 0.3 | 93.26 | 97.86 | 87.69 | 96.61 | 92.92 | 91.56 | 89.08 | 89.43 | 82.57 | 86.63 | 87.56 | 88.1 | 74.86 | 77.23 | 79.49 | 79.87 | 87.17 |
| 0.4 | 93.52 | 97.9 | 87.69 | 96.61 | 93.06 | 91.8 | 89.08 | 89.43 | 82.23 | 86.5 | 87.56 | 88.1 | 74.4 | 76.89 | 79.49 | 79.87 | 87.13 |
| 0.5 | 93.47 | 97.91 | 87.69 | 96.61 | 93.1 | 91.88 | 89.08 | 89.43 | 81.93 | 86.34 | 87.56 | 88.1 | 73.7 | 76.35 | 79.49 | 79.87 | 87.03 |
| 0.6 | 93.31 | 97.89 | 87.69 | 96.61 | 93.08 | 91.88 | 89.08 | 89.43 | 81.66 | 86.19 | 87.56 | 88.1 | 72.87 | 75.7 | 79.49 | 79.87 | 86.9 |
| 0.7 | 93.09 | 97.86 | 87.69 | 96.61 | 93.05 | 91.86 | 89.08 | 89.43 | 81.43 | 86.06 | 87.56 | 88.1 | 71.96 | 74.99 | 79.49 | 79.87 | 86.76 |
| 0.8 | 92.91 | 97.83 | 87.69 | 96.61 | 93.0 | 91.82 | 89.08 | 89.43 | 81.23 | 85.94 | 87.56 | 88.1 | 71.03 | 74.26 | 79.49 | 79.87 | 86.62 |
| 0.9 | 92.69 | 97.79 | 87.69 | 96.61 | 92.96 | 91.77 | 89.08 | 89.43 | 81.06 | 85.84 | 87.56 | 88.1 | 70.11 | 73.53 | 79.49 | 79.87 | 86.47 |
| 1.0 | 92.46 | 97.75 | 87.69 | 96.61 | 92.91 | 91.71 | 89.08 | 89.43 | 80.9 | 85.75 | 87.56 | 88.1 | 69.2 | 72.81 | 79.49 | 79.87 | 86.33 |
| Ours | 93.39 | 97.91 | 87.69 | 96.61 | 93.02 | 91.73 | 89.08 | 89.43 | 82.38 | 86.37 | 87.56 | 88.1 | 74.86 | 77.21 | 79.49 | 79.87 | 87.17 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Deng, B. Exploring Universal Domain Adaptation with CLIP Models: A Calibration Method. Entropy 2025, 27, 1213. https://doi.org/10.3390/e27121213
Deng B. Exploring Universal Domain Adaptation with CLIP Models: A Calibration Method. Entropy. 2025; 27(12):1213. https://doi.org/10.3390/e27121213
Chicago/Turabian StyleDeng, Bin. 2025. "Exploring Universal Domain Adaptation with CLIP Models: A Calibration Method" Entropy 27, no. 12: 1213. https://doi.org/10.3390/e27121213
APA StyleDeng, B. (2025). Exploring Universal Domain Adaptation with CLIP Models: A Calibration Method. Entropy, 27(12), 1213. https://doi.org/10.3390/e27121213

