Next Article in Journal
Conservation Laws, Soliton Dynamics, and Stability in a Nonlinear Schrödinger Equation with Second-Order Spatiotemporal Dispersion
Previous Article in Journal
Estimating the First, Hyper-Zagreb Index for Direct Product of F-Sum Graphs
Previous Article in Special Issue
A Federated Fine-Tuning Framework for Large Language Models via Graph Representation Learning and Structural Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

An LLM-Powered Framework for Privacy-Preserving and Scalable Labor Market Analysis

1
School of Economics and Finance, Guangdong University of Science and Technology, Dongguan 523083, China
2
Faculty of Education, University of Macau, Macao 999078, China
3
The Faculty of Data Science, City University of Macau, Macao 999078, China
*
Author to whom correspondence should be addressed.
Mathematics 2026, 14(1), 53; https://doi.org/10.3390/math14010053 (registering DOI)
Submission received: 25 October 2025 / Revised: 7 December 2025 / Accepted: 8 December 2025 / Published: 23 December 2025
(This article belongs to the Special Issue Privacy-Preserving Machine Learning in Large Language Models (LLMs))

Abstract

Timely and reliable labor market intelligence is crucial for evidence-based policymaking, workforce planning, and economic forecasting. However, traditional data collection and centralized analytics raise growing concerns about privacy, scalability, and institutional data governance. This paper presents a large language model (LLM)-powered framework for privacy-preserving and scalable labor market analysis, designed to extract, structure, and interpret occupation, skill, and salary information from distributed textual sources. Our framework integrates domain-adapted LLMs with federated learning (FL) and differential privacy (DP) to enable collaborative model training across organizations without exposing sensitive data. The architecture employs secure aggregation and privacy budgets to prevent information leakage during parameter exchange, while maintaining analytical accuracy and interpretability. The system performs multi-task inference—including job classification, skill extraction, and salary estimation—and aligns outputs to standardized taxonomies (e.g., SOC, ISCO, ESCO). Empirical evaluations on both public and semi-private datasets demonstrate that our approach achieves superior performance compared to centralized baselines, while ensuring compliance with privacy and data-sharing regulations. Expert review further confirms that the generated trend analyses are accurate, explainable, and actionable for policy and research. Our results illustrate a practical pathway toward decentralized, privacy-conscious, and large-scale labor market intelligence.
Keywords: machine learning; large language model; privacy protection machine learning; large language model; privacy protection

Share and Cite

MDPI and ACS Style

Ji, W.; Ying, Z. An LLM-Powered Framework for Privacy-Preserving and Scalable Labor Market Analysis. Mathematics 2026, 14, 53. https://doi.org/10.3390/math14010053

AMA Style

Ji W, Ying Z. An LLM-Powered Framework for Privacy-Preserving and Scalable Labor Market Analysis. Mathematics. 2026; 14(1):53. https://doi.org/10.3390/math14010053

Chicago/Turabian Style

Ji, Wei, and Zuobin Ying. 2026. "An LLM-Powered Framework for Privacy-Preserving and Scalable Labor Market Analysis" Mathematics 14, no. 1: 53. https://doi.org/10.3390/math14010053

APA Style

Ji, W., & Ying, Z. (2026). An LLM-Powered Framework for Privacy-Preserving and Scalable Labor Market Analysis. Mathematics, 14(1), 53. https://doi.org/10.3390/math14010053

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop