Next Article in Journal
Automated Malware Source Code Generation via Uncensored LLMs and Adversarial Evasion of Censored Model
Previous Article in Journal
Multimodal Emotion Recognition for Seafarers: A Framework Integrating Improved D-S Theory and Calibration: A Case Study of a Real Navigation Experiment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

Enhancing Patent Document Similarity Evaluation and Classification Precision Through a Multimodal AI Approach

by
Hyuna Kim
and
Gwangyong Gim
*
Department of IT Policy and Management, Graduate School, Soongsil University, Seoul 06978, Republic of Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(17), 9254; https://doi.org/10.3390/app15179254
Submission received: 13 July 2025 / Revised: 16 August 2025 / Accepted: 21 August 2025 / Published: 22 August 2025
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

With the global surge in patent filings, accurately evaluating similarity between patent documents has become increasingly critical. Traditional similarity assessment methods—primarily based on unimodal inputs such as text or bibliographic data—often fall short due to the complexity of legal language and the semantic ambiguity that is inherent in technical writing. To address these limitations, this study introduces a novel multimodal patent similarity evaluation framework that integrates weak AI techniques and conceptual similarity analysis of patent drawings. This approach leverages a domain-specific pre-trained language model optimized for patent texts, statistical correlation analysis between textual and bibliographic information, and a rule-based classification strategy. These components, rooted in weak AI methodology, significantly enhance classification precision. Furthermore, the study introduces the concept of conceptual similarity—as distinct from visual similarity—in the analysis of patent drawings, demonstrating its superior ability to capture the underlying technological intent. An empirical evaluation was conducted on 9613 patents in the manipulator technology domain, yielding 668,010 document pairs. Stepwise experiments demonstrated a 13.84% improvement in classification precision. Citation-based similarity assessment further confirmed the superiority of the proposed multimodal approach over existing methods. The findings underscore the potential of the proposed framework to improve prior art searches, patent examination accuracy, and R&D planning.
Keywords: patent document similarity; multimodal approach; weak AI techniques; conceptual similarity; text–image fusion; BERT-for-Patents; CPC classification correlation; prior-art retrieval patent document similarity; multimodal approach; weak AI techniques; conceptual similarity; text–image fusion; BERT-for-Patents; CPC classification correlation; prior-art retrieval

Share and Cite

MDPI and ACS Style

Kim, H.; Gim, G. Enhancing Patent Document Similarity Evaluation and Classification Precision Through a Multimodal AI Approach. Appl. Sci. 2025, 15, 9254. https://doi.org/10.3390/app15179254

AMA Style

Kim H, Gim G. Enhancing Patent Document Similarity Evaluation and Classification Precision Through a Multimodal AI Approach. Applied Sciences. 2025; 15(17):9254. https://doi.org/10.3390/app15179254

Chicago/Turabian Style

Kim, Hyuna, and Gwangyong Gim. 2025. "Enhancing Patent Document Similarity Evaluation and Classification Precision Through a Multimodal AI Approach" Applied Sciences 15, no. 17: 9254. https://doi.org/10.3390/app15179254

APA Style

Kim, H., & Gim, G. (2025). Enhancing Patent Document Similarity Evaluation and Classification Precision Through a Multimodal AI Approach. Applied Sciences, 15(17), 9254. https://doi.org/10.3390/app15179254

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop