Explainable Reciprocal Recommender System for Affiliate–Seller Matching: A Two-Stage Deep Learning Approach
Abstract
1. Introduction
- Multimodal fusion: Effectively combining structured metrics (click-through rate (CTR), earnings per click (EPC), commissions), categorical values (language, category, platform), temporal contexts (season, hour), graph priors (affiliate–affiliate and seller–seller similarity), as well as bios and program description text embeddings is complicated [6,7,8].
- Bias and diversity: Without sufficient countermeasures against popularity bias, head sellers are prioritized, resulting in uneven exposure and downstream discovery, thereby causing systems to miss many valuable niche matches [8].
- Latency and scalability: Scalable and real-time prediction systems require estimated execution times of 100 ms. Although heavier models can yield improved accuracy, they sacrifice speed.
- Trust and explainability: Business users typically ask why certain options are recommended to them. Black-box scores are insufficient, and clear reasons must be provided [9].
- Designing a two-stage reciprocal recommender for affiliate–seller matching with multimodal inputs, evaluated in both affiliate→seller and seller→affiliate directions.
- Using listwise ranking (ListNet) and cross-feature modeling (deep and cross network V2 (DCN-V2)) to rerank retrieved candidates and analyzing why the resulting normalized discounted cumulative gain at 10 (NDCG@10) scores are close to 1.0 under our synthetic evaluation setting, including checks for data leakage and pair overlap [7,12].
2. Background and Related Work
2.1. Background
2.1.1. Recommendation Paradigms
2.1.2. Two-Stage Pipelines
2.1.3. Cold-Start Solutions
2.1.4. Debiasing and Fairness
2.1.5. Explainability in Recommenders
2.2. Literature Review
2.2.1. Collaborative Filtering and Extensions
2.2.2. Content-Based and Hybrid Approaches
2.2.3. Two-Stage Architectures
2.2.4. Cold-Start Strategies
2.2.5. Bias, Fairness, and Diversity
2.2.6. Explainability
3. Materials and Methods
3.1. Exploratory Data Analysis (EDA)
3.1.1. Distribution of Affiliate Categories
3.1.2. Distribution of Affiliate Languages and Locations
3.1.3. Seller Category and Location Distribution
3.1.4. Distribution of Company Tiers
3.1.5. Distribution of Commission Rates
3.1.6. Distribution of Cookie Periods
The Performance Score Graph
3.1.7. Correlation Heatmap
3.1.8. Kruskal–Wallis Test
3.1.9. Revenue, Engagement Rate, and Composite Utility Graphs
Success Probability Based on Signal Strength
3.1.10. Dominance and Strong Pairs
3.1.11. Text Embedding Visualization with t-SNE
3.1.12. Cold-Start Analysis for Affiliates and Sellers
Utility by Hour of Day
Modeling EDA Summary
3.2. Real-World Behavioral Dataset Construction and Preprocessing
3.3. Model Architecture
3.4. Stage A: Setup and Splits
3.5. Stage B: Labels and Sampling
3.5.1. Hard Labels: Conversions
3.5.2. Soft Labels: Weighted Implicit Signals
3.5.3. Hard Negatives: Semantic and Popularity Control
3.6. Stage C: Feature Engineering
3.6.1. Affiliate Features
Numeric Features
Categorical Features
3.6.2. Seller Features
Numeric Features
Categorical Features
3.6.3. Pair Features
3.6.4. Embedding Integration
3.6.5. Final Feature Pack
3.7. Stage D: Candidate Generation
3.7.1. Two-Tower Architecture
3.7.2. InfoNCE Contrastive Learning
3.7.3. ANN Retrieval
3.7.4. Training Curves
3.7.5. Evaluation Metrics
3.8. Stage E: Ranking
3.8.1. Bayesian Personalized Ranking (BPR)
3.8.2. ListNet and LambdaRank (Listwise Models)
3.8.3. DCN-V2
3.8.4. Results from Our Data
Training Results
3.8.5. Why Ranking Is Important
3.8.6. Training Objectives: NDCG and MAP
3.8.7. Ranking Pipeline: Retrieve 50 → Rerank 10
- First, Candidate Generation (Stage D): From the entire catalog, the system first retrieves around 200 candidates using a two-tower and ANN search. Out of these, the top 50 candidates are sent to the ranking step.
- Second, Ranking (Stage E): The ranking models (BPR, ListNet, DCN-V2) then reorder the 50 candidates. The final output is the top 10 sellers displayed to the affiliate.
3.9. Stage F: Postprocessing
3.9.1. MMR Reranking (Diversity)
3.9.2. xQuAD Coverage
3.9.3. Popularity Penalty λ-Sweep
3.9.4. Summary of Postprocessing
3.10. Stage G: Explainability
3.10.1. SHAP Global Importance
3.10.2. Visualizations (Summary and Force Plots)
3.11. Experimental Setup
3.12. Model
4. Results
4.1. Candidate Generation
4.2. Ranking
Segment-Wise Performance
4.3. Postprocessing
Popularity Penalty λ-Sweep
4.4. Explainability
4.4.1. Global SHAP
4.4.2. Local Explanation
4.5. Statistical Validation
4.6. Ablation Studies
4.7. Scalability and Latency
4.8. Evaluation
5. Discussion
5.1. Practical Insights
5.2. Cold Start and Novelty
5.3. Explainability and Fairness
5.4. Limitations and Future Work
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Koren, Y.; Bell, R.; Volinsky, C. Matrix Factorization Techniques for Recommender Systems. Computer 2009, 42, 30–37. [Google Scholar] [CrossRef]
- Hu, Y.; Koren, Y.; Volinsky, C. Collaborative Filtering for Implicit Feedback Datasets. In Proceedings of the 2008 Eighth IEEE International Conference on Data Mining, Pisa, Italy, 15–19 December 2008; IEEE: New York, NY, USA, 2008; pp. 263–272. [Google Scholar] [CrossRef]
- He, X.; Liao, L.; Zhang, H.; Nie, L.; Hu, X.; Chua, T.-S. Neural Collaborative Filtering. In Proceedings of the 26th International Conference on World Wide Web, Perth, Australia, 3–7 April 2017; International World Wide Web Conferences Steering Committee: Geneva, Switzerland, 2017; pp. 173–182. [Google Scholar] [CrossRef]
- Abdollahpouri, H.; Burke, R.; Mobasher, B. Controlling Popularity Bias in Learning-to-Rank Recommendation. In Proceedings of the Eleventh ACM Conference on Recommender Systems, Como, Italy, 27–31 August 2017; ACM: New York, NY, USA, 2017; pp. 42–46. [Google Scholar] [CrossRef]
- Reimers, N.; Gurevych, I. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; pp. 3980–3990. [Google Scholar] [CrossRef]
- Wang, R.; Fu, B.; Fu, G.; Wang, M. Deep & Cross Network for Ad Click Predictions. In Proceedings of the ADKDD’17, Halifax, NS, Canada, 14 August 2017; ACM: New York, NY, USA, 2017; pp. 1–7. [Google Scholar] [CrossRef]
- Wang, R.; Shivanna, R.; Cheng, D.; Jain, S.; Lin, D.; Hong, L.; Chi, E. DCN V2: Improved Deep & Cross Network and Practical Lessons for Web-scale Learning to Rank Systems. In Proceedings of the Web Conference 2021, Ljubljana, Slovenia, 19–23 April 2021; ACM: New York, NY, USA, 2021; pp. 1785–1797. [Google Scholar] [CrossRef]
- Covington, P.; Adams, J.; Sargin, E. Deep neural networks for youtube recommendations. In Proceedings of the RecSys 2016—Proceedings of the 10th ACM Conference on Recommender Systems, Boston, MA, USA, 15–19 September 2016; pp. 191–198. [Google Scholar] [CrossRef]
- Ribeiro, M.T.; Singh, S.; Guestrin, C. Why Should I Trust You? In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; ACM: New York, NY, USA, 2016; pp. 1135–1144. [Google Scholar] [CrossRef]
- Vartak, M.; Thiagarajan, A.; Miranda, C.; Bratman, J.; Larochelle, H. A Meta-Learning Perspective on Cold-Start Recommendations for Items. In Advances in Neural Information Processing Systems (NeurIPS 2017); NeurIPS: Long Beach, CA, USA, 2017; pp. 785–793. [Google Scholar] [CrossRef]
- Malkov, Y.A.; Yashunin, D.A. Efficient and Robust Approximate Nearest Neighbor Search Using Hierarchical Navigable Small World Graphs. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 824–836. [Google Scholar] [CrossRef] [PubMed]
- Cao, Z.; Qin, T.; Liu, T.-Y.; Tsai, M.-F.; Li, H. Learning to rank. In Proceedings of the 24th international conference on Machine learning, Corvallis, OR, USA, 20–24 June 2007; ACM: New York, NY, USA, 2007; pp. 129–136. [Google Scholar] [CrossRef]
- Burges, C.J.C. From RankNet to LambdaRank to LambdaMART: An Overview. IEEE Internet Comput. 2010, 11, 81. [Google Scholar]
- Demšar, J. Statistical Comparisons of Classifiers over Multiple Data Sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
- Carbonell, J.; Goldstein, J. The use of MMR, diversity-based reranking for reordering documents and producing summaries. In Proceedings of the 21st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Melbourne, Australia, 24–28 August 1998; ACM: New York, NY, USA, 1998; pp. 335–336. [Google Scholar] [CrossRef]
- Santos, R.L.T.; Macdonald, C.; Ounis, I. Exploiting query reformulations for web search result diversification. In Proceedings of the 19th International Conference on World Wide Web, Raleigh North, CA, USA, 26–30 April 2010; ACM: New York, NY, USA, 2010; pp. 881–890. [Google Scholar] [CrossRef]
- Tintarev, N.; Masthoff, J. A Survey of Explanations in Recommender Systems. In Proceedings of the 2007 IEEE 23rd International Conference on Data Engineering Workshop, Istanbul, Turkey, 17–20 April 2007; IEEE: New York, NY, USA, 2007; pp. 801–810. [Google Scholar] [CrossRef]
- Henaff, O.J.; Srinivas, A.; Fauw, J.D.; Razavi, A.; Doersch, C.; Eslami, S.M.A.; Oord, A.v.d. Data-Efficient Image Recognition with Contrastive Predictive Coding. In Proceedings of the 37th International Conference on Machine Learning (ICML 2020), Online, 13–18 July 2020; pp. 4182–4192. [Google Scholar]
- Lundberg, S.; Lee, S.-I. A Unified Approach to Interpreting Model Predictions. arXiv 2017, arXiv:1705.07874. [Google Scholar] [CrossRef]
- Adomavicius, G.; Tuzhilin, A. Toward the next generation of recommender systems: A survey of the state-of-the-art and possible extensions. IEEE Trans. Knowl. Data Eng. 2005, 17, 734–749. [Google Scholar] [CrossRef]
- Ricci, F.; Rokach, L.; Shapira, B. Recommender Systems: Techniques, Applications, and Challenges. In Recommender Systems Handbook; Springer: New York, NY, USA, 2022; pp. 1–35. [Google Scholar] [CrossRef]
- Burke, R. Hybrid Recommender Systems: Survey and Experiments. User Model. User-Adapt. Interact. 2002, 12, 331–370. [Google Scholar] [CrossRef]
- van den Berg, R.; Kipf, T.N.; Welling, M. Graph Convolutional Matrix Completion. arXiv 2017, arXiv:1706.02263. [Google Scholar] [CrossRef]
- Sundararajan, M.; Taly, A.; Yan, Q. Axiomatic Attribution for Deep Networks. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 3319–3328. [Google Scholar]
- Fisher, A.; Rudin, C.; Dominici, F. All Models are Wrong, but Many are Useful: Learning a Variable’s Importance by Studying an Entire Class of Prediction Models Simultaneously. J. Mach. Learn. Res. 2019, 20, 1–81. [Google Scholar]
- Rendle, S.; Freudenthaler, C.; Gantner, Z.; Schmidt-Thieme, L. BPR: Bayesian personalized ranking from implicit feedback. arXiv 2012, arXiv:arXiv:1205.2618. [Google Scholar]
- Lops, P.; de Gemmis, M.; Semeraro, G. Content-based Recommender Systems: State of the Art and Trends. In Recommender Systems Handbook; Springer: Boston, MA, USA, 2011. [Google Scholar]
- Zhang, Y.; Chen, X. Explainable Recommendation: A Survey and New Perspectives. Found. Trends® Inf. Retr. 2020, 14, 1–101. [Google Scholar] [CrossRef]
- Amatriain, X.; Basilico, J. Recommender Systems in Industry: A Netflix Case Study. In Recommender Systems Handbook; Springer: Boston, MA, USA, 2015; pp. 385–419. [Google Scholar]
- Mathur, A.; Narayanan, A.; Chetty, M. Endorsements on social media: An empirical study of affiliate marketing disclosures on YouTube and pinterest. Proc. ACM Hum. Comput. Interact. 2018, 2, 119. [Google Scholar] [CrossRef]
- Yi, X.; Yang, J.; Hong, L.; Cheng, D.Z.; Heldt, L.; Kumthekar, A.; Zhao, Z.; Wei, L.; Chi, E. Sampling-bias-corrected neural modeling for large corpus item recommendations. In Proceedings of the 13th ACM Conference on Recommender Systems, Copenhagen, Denmark, 16–20 September 2019; ACM: New York, NY, USA, 2019; pp. 269–277. [Google Scholar] [CrossRef]
- Santos, R.L.T.; Peng, J.; Macdonald, C.; Ounis, I. Explicit Search Result Diversification through Sub-queries. In Proceedings of the 32nd European Conference on IR Research, ECIR 2010, Milton Keynes, UK, 28–31 March 2010; pp. 87–99. [Google Scholar] [CrossRef]
- Zhang, S.; Yao, L.; Sun, A.; Tay, Y. Deep learning based recommender system: A survey and new perspectives. ACM Comput. Surv. 2019, 52, 5. [Google Scholar] [CrossRef]
- Tsai, Y.-H.H.; Bai, S.; Liang, P.P.; Kolter, J.Z.; Morency, L.-P.; Salakhutdinov, R. Multimodal Transformer for Unaligned Multimodal Language Sequences. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; pp. 6558–6569. [Google Scholar] [CrossRef]















| Direction | K | Recall@K | Precision@K | HitRate@K |
|---|---|---|---|---|
| A→S | 10 | 0.0037 | 0.0004 | 0.0037 |
| A→S | 50 | 0.0361 | 0.0008 | 0.0376 |
| S→A | 10 | 0.0116 | 0.0017 | 0.0168 |
| S→A | 50 | 0.0644 | 0.0017 | 0.0842 |
| Ablation Setup | Δ NDCG | Δ Recall | Observation |
|---|---|---|---|
| Text Embeddings | −0.412 | −0.380 | Model losing its understanding of semantic meaning |
| Pair Features | −0.263 | −0.201 | Less personalized matching |
| Structured (numeric) | −0.318 | −0.289 | Missing CTR and EPC weakening precision |
| Only-Text | −0.524 | −0.488 | Semantic only insufficient |
| Only-Pair | −0.469 | −0.422 | Context lost |
| Full Model | 0.9956 | 0.9956 | Best overall |
| Stage | p50 (ms) | p95 (ms) |
| Retrieval A→S | 1.28 | 1.65 |
| Retrieval S→A | 0.86 | 1.12 |
| Ranker (TopK) | 75.09 | 128.55 |
| End-to-End | 76.36 | 130.20 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Almutairi, H.; Ykhlef, M. Explainable Reciprocal Recommender System for Affiliate–Seller Matching: A Two-Stage Deep Learning Approach. Information 2026, 17, 101. https://doi.org/10.3390/info17010101
Almutairi H, Ykhlef M. Explainable Reciprocal Recommender System for Affiliate–Seller Matching: A Two-Stage Deep Learning Approach. Information. 2026; 17(1):101. https://doi.org/10.3390/info17010101
Chicago/Turabian StyleAlmutairi, Hanadi, and Mourad Ykhlef. 2026. "Explainable Reciprocal Recommender System for Affiliate–Seller Matching: A Two-Stage Deep Learning Approach" Information 17, no. 1: 101. https://doi.org/10.3390/info17010101
APA StyleAlmutairi, H., & Ykhlef, M. (2026). Explainable Reciprocal Recommender System for Affiliate–Seller Matching: A Two-Stage Deep Learning Approach. Information, 17(1), 101. https://doi.org/10.3390/info17010101
