4.2.1. Dynamic Feature Space Reconstruction
DFSR aims to map the source domain (data-rich countries) and target domain (data-scarce countries) to a shared feature space, making knowledge transfer more effective. Considering the multi-level nature of Olympic prediction, DFSR learns specific feature mapping functions for each prediction level.
Given source domain data
and target domain data
, for prediction level
l ∈ {country level, sport type level, event level}, DFSR learns a feature mapping function
ϕl that maps original features
x to new features
ϕl (
x). The mapping function is determined by optimizing the following objective:
where
is the Maximum Mean Discrepancy loss, which reduces distribution differences between source and target domains:
we use a Gaussian RBF kernel
, with the bandwidth
set via the median heuristic on pairwise distances in each mini-batch.
is the adversarial loss, implemented via a domain classifier
D to encourage domain-invariant features:
is the structure preservation loss, ensuring that the mapping preserves structural relationships from the original feature space:
we grid-searched
: for each value, models were trained on 1896–2016 and evaluated on the 2020 validation set. We selected the setting that minimized the target-domain MAE;
was used thereafter. Here,
is the weight on the structure-preserving term
. Similarly, we grid-searched
for the regularization term
(implemented as L2 weight decay) and
for the consistency loss, using the same training–validation setup, selecting
and
based on minimizing MAE. For the adversarial term, we grid-searched
and selected
using the same criterion.
is a complexity regularization term preventing overfitting; and
is the hierarchical consistency loss, ensuring mapping relationships between different levels:
where
represents levels adjacent to level
l, and
is a level transformation function.
We adopt MMD-based alignment with structural regularization and an adversarial term to explicitly reduce cross-domain distribution shift while preserving task-relevant geometry. This combination is well suited to transferring knowledge from data-rich to data-scarce countries—MMD provides an architecture-agnostic discrepancy reduction [
11], while adversarial training promotes domain-invariant features with minimal overhead [
8]; the cross-level consistency term keeps the learned space coherent across country/sport/event granularities.
To facilitate understanding, we use the United States (source domain) and Nigeria (target domain) as examples to illustrate the working process of DFSR. In the original feature space, these two countries show significant differences in dimensions such as GDP and participation history, making direct comparison difficult. Through DFSR mapping, we obtain new feature representations that make the two countries more comparable regarding relative advantage event patterns and sports development trajectories. For example, the proportion of track and field events in global medals is about 4.2% for the United States and about 2.1% for Nigeria; after mapping, the distance between the two countries in the “relative sports event layout” dimension is reduced from 0.87 to 0.31, achieving feature space alignment.
We also created visualizations to intuitively understand this process, as shown in
Figure 3:
Figure 3 uses t-SNE dimensionality reduction to project high-dimensional features into a two-dimensional space, comparing feature distributions before reconstruction (left) and after reconstruction (right). Blue points represent the source domain (data-rich countries), and red points represent the target domain (data-scarce countries). In
Figure 3a, the original feature space, source, and target domain countries show evident separation, with data-rich countries clustered in the upper right region and data-scarce countries scattered in the left and lower regions; in
Figure 3b, in the feature space after DFSR reconstruction, the two types of countries show significant overlap, with the distribution difference between source and target domains markedly reduced, especially in the dimensions of event layout and development trajectory. The figure marks position changes of typical countries, such as the United States (US), China (CN), Nigeria (NG), and Cambodia (KH), allowing observation of the relative position changes of these countries in the feature space before and after reconstruction. While t-SNE provides an intuitive visualization, we acknowledge that it is a non-linear technique that may distort global distances; to complement this, we quantitatively assess alignment via the MMD metric, which decreased from 0.45 to 0.12 after reconstruction, confirming effective domain alignment.
To implement the adversarial component as part of the optimization, we introduced the Gradient Reversal Layer (GRL) and Domain-Specific Batch Normalization (DSBN) techniques.
Gradient Reversal Layer: The GRL acts as an identity function in forward propagation but reverses the gradient and multiplies it by coefficient during backpropagation. By adding a domain classifier D and a GRL, we can implement adversarial training, encouraging the feature extractor to learn domain-invariant features.
Domain-Specific Batch Normalization: Unlike traditional batch normalization that uses the same scaling and shifting parameters for all data, DSBN learns independent normalization parameters for source and target domains:
where
indicates domain identity,
and
are the mean and standard deviation within the batch, respectively, and
and
are learnable scaling and shifting parameters for domain
d.
The GRL encourages domain-invariant representations without altering the forward pass [
8], and DSBN mitigates covariate shift by decoupling normalization statistics across domains [
9]. Both are lightweight yet effective additions that stabilize transfer under heterogeneous country distributions.
4.2.2. Hierarchical Adaptive Transfer Strategy
The HATS is a core component of the MG-TLC and is responsible for implementing knowledge transfer in the reconstructed feature space. Our HATS method draws inspiration from the bidirectional knowledge transfer framework of Jiang et al. [
13] to optimize predictions for data-scarce countries. The HATS is based on two key ideas: first, different country groups should have different transfer strategies; and second, transfer learning should span multiple levels, including the country level, sport type level, and event level.
The HATS first clusters countries (from both source and target domains) into
K groups, each representing a typical Olympic development pattern, such as the “all-around type,” “single-event specialization type,” etc. Clustering is performed in the reconstructed feature space using the spectral clustering algorithm:
where
is the similarity between countries
i and
j in the reconstructed feature space.
Spectral clustering leverages the affinity graph in the reconstructed space to capture non-linear country similarities; it is well suited to heterogeneous Olympic development patterns and does not presuppose convex clusters. Grouping both source and target countries in this space provides a natural scaffold for targeted transfer within each cluster.
The HATS establishes two models for each cluster
Ck: a source domain base model
and a cross-domain adjustment model
gk. The source domain base model is trained directly on source domain data:
The cross-domain adjustment model learns corrections from source domain predictions to target domain true values:
During prediction, for a country
v in the target domain, the HATS first determines its cluster
and then calculates the predicted value through the source domain base model and cross-domain adjustment model:
A single transfer model is insufficient for the diverse Olympic landscape. Clustering countries into groups with similar development patterns (e.g., all-around vs. single-event specialists) enables targeted transfer. The source-base model captures well-estimated signals from data-rich sources, while the cross-domain adjustment corrects cluster-specific distributional shifts and biases, allowing data-scarce targets to borrow strength from matched sources without washing out heterogeneity.
For example, consider Georgia (target domain), whose feature pattern is similar to that of “single-event-specialization-type” countries like Bulgaria and Hungary. The HATS assigns Georgia to this cluster and applies the specific transfer strategy for this type of country. In the 2024 prediction, the source domain base model gave an initial prediction of 6.3 medals, and the cross-domain adjustment model corrected it to 8.1 medals based on historical performance, close to the actual result (seven medals).
Figure 4 shows radar charts of feature patterns for five country types in the HATS.
Each radar chart represents the feature pattern of a country type, with axes including economic strength, historical performance, event diversity, gender balance, participation stability, and event specialization index. (a) “All-around-type” countries (e.g., the United States, China) are strong across all dimensions, especially economic strength and historical performance; (b) “single-event-specialization-type” countries (e.g., Georgia, Hungary) have high event specialization indices but low event diversity; (c) “team-event-dominant-type” countries (e.g., Serbia, Croatia) excel at team events; (d) “emerging development-type” countries (e.g., Qatar, Azerbaijan) have strong economic strength but weak historical performance; and (e) “initial exploration-type” countries (e.g., Cambodia, Andorra) are weak across all dimensions but improving in participation. Each type is annotated with 2–3 representative countries as examples.
4.2.3. Multi-Level Prediction Consistency Guarantee Mechanism
To achieve multi-level prediction consistency, we employ a three-stage optimization strategy:
1. Independent Training Stage: Separately train country-level, sport type-level, and event-level prediction models to obtain initial prediction results , , and .
2. Consistency Optimization Stage: Introduce an explicit hierarchical consistency loss function to optimize models at all three levels jointly:
where
and
are tuned via a 2D grid search on the 2020 validation set with
and
. The selection criterion is a weighted combination of event-level MAE and the L2 residuals of cross-level consistency (i.e.,
and
). The optimal values are
and
.
3. Post-Processing Calibration Stage: To enforce strict hierarchical consistency, we perform a final calibration on the optimized prediction results. We adopt a top-down, importance-weighted adjustment method, where higher-level predictions are assumed to be more robust and are used to calibrate lower-level estimates.
Then, sport type-level predictions are recalculated based on the adjusted event-level results:
The country-level prediction, serving as the anchor, remains unchanged:
Through this three-stage approach, we ensure that final predictions strictly satisfy hierarchical consistency constraints:
We acknowledge that this deterministic post hoc scaling represents a design trade-off. While it enforces logical consistency—a critical feature for the usability of the forecast—it may alter the probabilistic calibration of the initial, unconstrained model outputs. However, this trade-off is justified, as the process empirically improves the overall accuracy of the point estimates. On the validation set, after adopting this mechanism, the MAE of event-level predictions decreased by a significant 15.2%. This suggests that the information from more stable, higher-level forecasts provides a valuable constraint that refines the more granular predictions, leading to a net improvement in predictive performance. A detailed analysis of the impact on probabilistic calibration is left for future work.
4.2.4. Olympic Development Stage Feature Analysis
In addition to horizontal knowledge transfer, the MG-TLC also considers vertical development stage features. By establishing an Olympic development stage model, different countries can learn from the experiences of countries at similar stages.
We divide a country’s Olympic development into three stages, the initial stage, growth stage, and mature stage, and extract specific feature patterns for each stage:
Initial Stage: Feature patterns include the participation event count, athlete count, GDP growth rate, and other composite features.
Growth Stage: Feature patterns include the medal growth rate, event diversification index, and other composite features.
Mature Stage: Feature patterns include the medal stability index, event specialization index, and other composite features.
Through time series analysis, we identify which development stage a country is in and apply corresponding feature patterns and transfer strategies. For example, for a target domain country in the initial stage, we can focus on learning from source domain countries that were once in the initial stage and successfully transitioned to the growth stage.