Next Article in Journal
Recent Progress on Eye-Tracking and Gaze Estimation for AR/VR Applications: A Review
Previous Article in Journal
Design Analysis of an Interior Permanent Magnet Synchronous Motor with Hybrid Hair-Pin and Litz Wire Windings
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

LLM-Guided Weighted Contrastive Learning with Topic-Aware Masking for Efficient Domain Adaptation: A Case Study on Pulp-Era Science Fiction

Department of English Language and Literature, Kyungpook National University, Daegu 41566, Republic of Korea
Electronics 2025, 14(17), 3351; https://doi.org/10.3390/electronics14173351
Submission received: 31 July 2025 / Revised: 19 August 2025 / Accepted: 20 August 2025 / Published: 22 August 2025
(This article belongs to the Section Artificial Intelligence)

Abstract

Domain adaptation of pre-trained language models remains challenging, especially for specialized text collections that include distinct vocabularies and unique semantic structures. Existing contrastive learning methods frequently rely on generic masking techniques and coarse-grained similarity measures, which limit their ability to capture fine-grained, domain-specific linguistic nuances. This paper proposes an enhanced domain adaptation framework by integrating weighted contrastive learning guided by large language model (LLM) feedback and a novel topic-aware masking strategy. Specifically, topic modeling is utilized to systematically identify semantically crucial domain-specific terms, enabling the creation of meaningful contrastive pairs through three targeted masking strategies: single-keyword, multiple-keyword, and partial-keyword masking. Each masked sentence undergoes LLM-guided reconstruction, accompanied by graduated similarity assessments that serve as continuous, fine-grained supervision signals. Experiments conducted on an early 20th-century science fiction corpus demonstrate that the proposed approach consistently outperforms existing baselines, such as SimCSE and DiffCSE, across multiple linguistic probing tasks within the newly introduced SF-ProbeEval benchmark. Furthermore, the proposed method achieves these performance improvements with significantly reduced computational requirements, highlighting its practical applicability for efficient and interpretable adaptation of language models to specialized domains.
Keywords: domain adaptation; contrastive learning; large language models; efficient AI; model optimization; neural networks; computational efficiency domain adaptation; contrastive learning; large language models; efficient AI; model optimization; neural networks; computational efficiency

Share and Cite

MDPI and ACS Style

Kang, S. LLM-Guided Weighted Contrastive Learning with Topic-Aware Masking for Efficient Domain Adaptation: A Case Study on Pulp-Era Science Fiction. Electronics 2025, 14, 3351. https://doi.org/10.3390/electronics14173351

AMA Style

Kang S. LLM-Guided Weighted Contrastive Learning with Topic-Aware Masking for Efficient Domain Adaptation: A Case Study on Pulp-Era Science Fiction. Electronics. 2025; 14(17):3351. https://doi.org/10.3390/electronics14173351

Chicago/Turabian Style

Kang, Sujin. 2025. "LLM-Guided Weighted Contrastive Learning with Topic-Aware Masking for Efficient Domain Adaptation: A Case Study on Pulp-Era Science Fiction" Electronics 14, no. 17: 3351. https://doi.org/10.3390/electronics14173351

APA Style

Kang, S. (2025). LLM-Guided Weighted Contrastive Learning with Topic-Aware Masking for Efficient Domain Adaptation: A Case Study on Pulp-Era Science Fiction. Electronics, 14(17), 3351. https://doi.org/10.3390/electronics14173351

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop