Next Article in Journal
Consideration of FedProx in Privacy Protection
Next Article in Special Issue
RoBERTa-Based Keyword Extraction from Small Number of Korean Documents
Previous Article in Journal
FCIHMRT: Feature Cross-Layer Interaction Hybrid Method Based on Res2Net and Transformer for Remote Sensing Scene Classification
Previous Article in Special Issue
Addressing Long-Distance Dependencies in AMR Parsing with Hierarchical Clause Annotation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multi-Modal Retrieval Model for Mathematical Expressions Based on ConvNeXt and Hesitant Fuzzy Set

1
School of Cyber Security and Computer, Hebei University, Baoding 071002, China
2
Institute of Intelligent Image and Document Information Processing, Hebei University, Baoding 071002, China
3
Hebei Machine Vision Engineering Research Center, Hebei University, Baoding 071002, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(20), 4363; https://doi.org/10.3390/electronics12204363
Submission received: 21 September 2023 / Revised: 14 October 2023 / Accepted: 19 October 2023 / Published: 20 October 2023
(This article belongs to the Special Issue Natural Language Processing and Information Retrieval)

Abstract

:
Mathematical expression retrieval is an essential component of mathematical information retrieval. Current mathematical expression retrieval research primarily targets single modalities, particularly text, which can lead to the loss of structural information. On the other hand, multimodal research has demonstrated promising outcomes across different domains, and mathematical expressions in image format are adept at preserving their structural characteristics. So we propose a multi-modal retrieval model for mathematical expressions based on ConvNeXt and HFS to address the limitations of single-modal retrieval. For the image modal, mathematical expression retrieval is based on the similarity of image features and symbol-level features of the expression, where image features of the expression image are extracted by ConvNeXt, while symbol-level features are obtained by the Symbol Level Features Extraction (SLFE) module. For the text modal, the Formula Description Structure (FDS) is employed to analyze expressions and extract their attributes. Additionally, the application of the Hesitant Fuzzy Set (HFS) theory facilitates the computation of hesitant fuzzy similarity between mathematical queries and candidate expressions. Finally, Reciprocal Rank Fusion (RRF) is employed to integrate rankings from image modal and text modal retrieval, yielding the ultimate retrieval list. The experiment was conducted on the publicly accessible ArXiv dataset (containing 592,345 mathematical expressions) and the NTCIR-mair-wikipedia-corpus (NTCIR) dataset.The MAP@10 values for the multimodal RRF fusion approach are recorded as 0.774. These substantiate the efficacy of the multi-modal mathematical expression retrieval approach based on ConvNeXt and HFS.

1. Introduction

Mathematical expressions play a significant role in scientific communication and calculations. Mathematical Information Retrieval (MIR) is a crucial component of Information Retrieval (IR) that deals with searching for specific mathematical expressions, concepts, or objects. Efficient indexing and retrieval of mathematical expressions have become the most challenging part of MIR due to the growing use of mathematical content in various scientific documents, educational materials, and web information in MathML or LaTeX format. Many academics have been working on mathematical information retrieval research recently and have seen some success [1,2,3].
Due to the two-dimensional structure of mathematical expressions and the complexity of mathematical symbol types and semantics, traditional search engines primarily designed for plain text retrieval often struggle to fulfill the required criteria. Mathematical expression retrieval encounters numerous challenges:
  • Mathematical Context: Understanding the context in which a mathematical expression is used is crucial, as the same notation can have different meanings in distinct mathematical subfields or domains.
  • Equivalent transformations of a mathematical expression: Mathematical expressions in different forms may convey the same meaning (e.g., a × ( x + z ) and a × x + a × z , a 1 2 and a ).
  • Multimodal Content: Multimodal content presents mathematical expressions in various forms, in addition to textual modes like MathML and LaTeX, these expressions can also be represented through images. Different modalities of mathematical expressions exhibit distinct characteristics, necessitating the effective handling of diverse data types.
  • Structure Complexity: Mathematical expressions can be symbolically complex, involving nested functions, subscripts, superscripts, and specialized symbols that require intricate parsing and interpretation.
Current research in multimodal studies is expanding rapidly, exploring the integration of various data types and modes to gain a more comprehensive understanding of complex phenomena [4,5]. In response to the challenges outlined above, specifically Challenge 3 and Challenge 4, this paper makes the following contributions:
  • Introducing the image modality in mathematical expression retrieval. Images can capture the visual aspects of mathematical expressions, providing a richer and more comprehensive representation compared to plain text. Combining image modal with text modal allows a more comprehensive understanding of mathematical content. This integration can enhance retrieval accuracy by considering multiple data types simultaneously.
  • Building upon extracting image features from mathematical expression images, we have devised a symbol-level feature extraction method to obtain a more comprehensive set of image feature information. This enhancement ensures that the ranking results produced by the image modality retrieval module are more rational and accurate.
  • We employ algorithmic analysis to extract attributes from the textual modality of mathematical expressions. Subsequently, we construct a table of attributes for textual modality expressions. By introducing hesitant fuzzy sets and leveraging their advantages in handling multi-attribute evaluation criteria, we calculate the similarity between expressions.
  • We opt for an appropriate fusion sorting method to combine and rank the retrieval results from both image and text modalities, resulting in the final ranking outcome. Furthermore, a subset of literature from the publicly accessible ArXiv dataset was extracted and utilized to construct a dataset encompassing 592,345 mathematical expressions.

2. Related Work

Regarding mathematical expression retrieval of the text modal, expression trees are widely used for storing and processing mathematical expressions and have been applied to mathematical expression retrieval by many scholars. Goel et al. [6] have undertaken studies on Math Word Problems, utilizing a tree-matching algorithm for the matching of mathematical expressions; they performed pair-wise matching on expression trees through post-order traversal. Pfahler et al. [7] incorporated unsupervised embedding learning and Graph Convolutional Neural Networks (GCNNs) for learning mathematical representations. In order to facilitate effective nearest-neighbor queries, mathematical operations represented in XML format were processed as graphical data and embedded into a low-dimensional vector space. Schellenberg et al. [8] employed substitution trees to index and retrieve mathematical expressions in LaTeX representation, but the insertion bias limits its performance. Hu et al. [9] proposed WikiMirs, using a method of generalization that is hierarchical to produce subtrees from the representation trees of mathematical expressions, which can support substructure and similarity matching of mathematical expressions. Zhong et al. [10] presented a dynamic pruning algorithm for inverted index, representing mathematical expressions as OPTs (Operator Trees), which improves retrieval efficiency for substructures of mathematical expressions.
Neural network methods have achieved significant progress in natural language-related tasks, but their performance on mathematical language-related tasks remains an active research area. Gao et al. [11] proposed a formula vector generation method based on “formula2vec” by analyzing feature differences between natural and mathematical languages. In pursuit of attaining heightened semantic information during embedding, Dadure et al. [12] proposed a contextual formula embedding method that retrieves syntactically and semantically similar formulas, sub-formulas, and parent formulas, highlighting the importance of formula context in mathematical information retrieval. Peng et al.’s MathBERT pre-training model [13] can capture the semantic structure information of formulas by concurrently training the formulas and the contexts that relate to them. Dai [14] proposed NTFEM, which extends N-ary tree representations of MathML formulas to one-dimensional linear sequences, uses a word embedding model to obtain the sub-structure vector, and applies a weighting function to obtain a weighted average embedding vector.
In the realm of image-based mathematical expression retrieval, Marinai et al. [15] proposed a mathematical symbol retrieval approach based on visual bag-of-words encoding, which employed self-organizing maps to cluster shape context into appropriate visual dictionaries, facilitating efficient retrieval of mathematical symbols. Zanibbi et al. [16] used content-based image retrieval to match binarized and decomposed query images with expression images.
In summary, current mathematical expression retrieval mainly focuses on unimodal approaches. For text modal retrieval, methods based on representation trees and neural networks embedding expressions into vectors predominate. However, these methods may overlook structural information. Image modal retrieval, on the other hand, tends to decompose mathematical expression images into symbol images or connected components for retrieval, often emphasizing symbol similarity at the expense of semantic information. To address these limitations, our study integrates image and text retrieval outcomes, achieving a multimodal retrieval model based on mathematical expression images and text. This approach considers both modalities’ similarities, leading to more rational retrieval outcomes.
In the field of Mathematical Information Retrieval (MIR), precise assessment of expression similarity holds paramount importance. While prevalent models, such as those proposed in [13,14], rely on cosine distance, the intricate two-dimensional characteristics of mathematical expressions demand a more objective approach. The utilization of Hesitant Fuzzy Set (HFS) theory emerges as an apt choice for addressing uncertainty and multi-attribute evaluation, offering a versatile means of representing hesitant information [17,18,19,20,21]. The integration of HFS theory into mathematical expression retrieval enhances the holistic assessment of diverse attribute features.

3. The Proposed Model Overview

Figure 1 represents the framework diagram of the multi-modal retrieval model for mathematical expressions based on ConvNeXt and HFS. Users input images and text of query mathematical expressions and receive a ranked list after retrieval. This system retrieves mathematical expressions from both image and text modalities. In the Image-Modal Retrieval Module, ConvNeXt [22] is used for image feature extraction and Symbol Level Features Extraction (SLFE) for symbol-level features of the mathematical expression images. Initial retrieval is conducted based on image features, followed by ranking based on symbol-level similarity, yielding descending search outcomes. In the Text-Modal Retrieval Module, the Formula Description Structure (FDS) [23] is utilized for the analysis of textual expressions, resulting in multiple attribute features of mathematical symbols. Additionally, the Hesitant Fuzzy Set (HFS) [24] is used to compute hesitant fuzzy similarity between mathematical query and candidate expressions, with results returned in descending similarity order. For result enhancement, the Rank Fusion Module employs Reciprocal Rank Fusion (RRF) to combine rankings generated by distinct modules.
In our research, we broaden the modal of mathematical expression retrieval to include both image and text forms, allowing retrieval from LaTeX and MathML formats. Additionally, HFS enables a flexible representation of uncertain information, facilitating a comprehensive assessment of attribute influences on decision-making. The membership degrees of mathematical expression attributes obtained through FDS analysis are computed using hesitant fuzzy sets, and the similarity between the query and candidate expressions is determined based on the HFS similarity calculation formula.
It is essential to highlight that, as of now, the method proposed in this paper faces limitations in retrieving semantically equivalent formulae to the queried ones. This is attributed to the following factors: The transformation of semantically equivalent mathematical expressions necessitates a profound understanding of domain-specific knowledge, which currently poses challenges for automated processing. The inherent semantics of mathematical expressions can be notably ambiguous, even for human comprehension, particularly when presented in isolation without adequate context. For example, the expression “ m n ” can be interpreted either as a variable “ m n ” or as a multiplication operation between “m” and “n”. This complexity further complicates machine-based semantic equivalence determination.

4. Methods

4.1. Retrieval of Mathematical Expressions in Image Modal

4.1.1. Extraction of Mathematical Expression Image Features

While meeting the requirements, the configuration of ConvNeXt-Tiny is simpler and has fewer parameters compared to other versions. It exhibits faster feature extraction capabilities, facilitating the expedited construction of feature databases. Consequently, this study employs the ConvNeXt-Tiny neural network model [22] for feature extraction in the image modal of mathematical expressions. Figure 2 shows the network structure of ConvNeXt-Tiny. It consists of four stages with different numbers of ConvNeXt blocks, each producing a feature map with different dimensions. It has strong feature extraction ability, few parameters, and low hardware requirements during training.
The Visual Transformer (ViT) [25] has emerged as an effective alternative to convolutional neural networks for various computer vision tasks, exemplified by models such as the Swin Transformer [26]. Leveraging the layer structure, downsampling techniques, activation functions, data processing methods, anti-bottleneck architecture, and deep convolution inspired by the Swin Transformer, ConvNeXt [22] further enhances the image feature extraction performance.
After extracting image features using the Convnext network, an image feature index is constructed. During retrieval, cosine distance is employed to calculate the similarity between feature vectors.

4.1.2. Symbol-Level Feature Extraction Module

Compared with other types of images, mathematical expression images have unique features, and the extraction of symbol-level features is helpful to the retrieval of mathematical expressions. Mathematical expressions are distinct in their two-dimensional structure, complex symbols, and specific spatial arrangements. To accurately retrieve and evaluate mathematical expressions, it is essential to capture the symbol-level details, which are crucial for understanding their meaning and structure.
The Symbol Level Features Extraction (SLFE) module extracts symbol-level features from mathematical expression images by segmenting the input image into individual symbol blocks. Given a query image of a mathematical expression E Q , consider E S = e s 1 , e s 2 , , e s n as the set of symbol blocks representing expression elements and n as representing the total count of these symbol blocks ( e s i R 30 × 30 ). Position vectors are calculated for element symbol blocks to retain spatial information. We use the connected component labeling algorithm to obtain element symbol blocks in mathematical expression images [27].
Figure 3a shows attached symbols “a” and “c”, which are separated using the connected component labeling algorithm (Figure 3b). Each resulting component forms an element symbol block of 30 × 30 pixels. Position vectors P s i = p 1 , p 2 , , p n of the element symbol block e s i are computed for the symbol block set S, where p j = t j , d j , l j , r j , r o j represents the position vector for the j-th element symbol block (Figure 4).
Here t j , d j , l j , and r j represent the distance between the top, down, left, and right edges of the j-th element symbol block and the top and left edges of the query image E Q . The position vector elements are standardized to range between 0 and 1 using Equation (1), where d max = max d 1 , , d n , r max = max r 1 , , r n , d max r max = r o i are the height-to-width ratio of E Q , for recording the size proportion of the expression.
p i = t i d max , d i d max , l i r max , r i r max , d max r max
In this study, we created a database of indexed mathematical expression images with their image features and symbol-level features. During the query process, we retrieve results based on image feature similarity, then sort by symbol-level feature similarity before outputting the final results.

4.2. Retrieval of Mathematical Expressions in Text Modal

4.2.1. Parsing and Extraction of Textual Attributes in Mathematical Expressions

The FDS [23] is used to analyze mathematical expressions and extract their properties. Each symbol in the expression has four attribute values: Number, Level, Operator, and Flag. As shown in Equation (2), the attributes obtained by parsing and extraction using FDS are represented as a quadruple array.
A S i = N i , L i , O i , F i
The meaning of each attribute value is as follows: (1) N i denotes the sequence number of S i in the expression. (2) L i represents the horizontal baseline position of the symbol S i in the expression, as shown in Figure 5. (3) O i is the function code of S i , indicating whether the current symbol is an operator ( O i = 1 ) or an operand ( O i = 0 ). (4) F i is the spatial flag information of S i , which shows how the present symbol and the prior symbol are related. F i can take values from 1 to 8, representing “above”, “superscript”, “right”, “subscript”, “below”, “contains”, “left-superscript”, and “left-subscript”. The main baseline’s symbols have F = 0 .

4.2.2. Text Similarity Calculation of Mathematical Expressions

FDS analysis on mathematical expressions yields four attribute values. A single distance calculation may introduce errors in similarity measurement. A fuzzy set [28] is a group of items with a variety of membership grades. To deal with multi-attribute evaluation indicators, we used hesitant fuzzy sets proposed by Torra [24] to calculate the similarity.
Definition 1. 
Let X be a non-empty set. A hesitant fuzzy set [29] H is defined as follows:
H = < x , h H ( x ) > x X
In this context, h H ( x ) is the hesitant fuzzy element, which is a set of several possible degrees of membership of an element x in the set H of the elements in X = x 1 , x 2 , , x n [29]. Its value range is [0,1].
Let M and N be hesitant fuzzy sets on the non-empty set X = x 1 , x 2 , , x n . The generalized hesitant fuzzy standard distance and similarity [29] between M and N are respectively defined as:
d g h n ( M , N ) = 1 n i = 1 n 1 l x i j = 1 l x i h M σ ( j ) x i h N σ ( j ) x i λ 1 / λ
s i m ( M , N ) = 1 d g h n ( M , N )
In this equation, d g h n ( M , N ) denotes the generalized hesitant fuzzy standard distance between hesitant fuzzy sets M and N, while s i m ( M , N ) represents their corresponding similarity [29]. λ > 0 , when λ = 1 , d g h n ( M , N ) is the hesitant fuzzy Hamming distance, and when λ = 2 , d g h n ( M , N ) is the hesitant fuzzy Euclidean distance; h M σ ( j ) x i and h N σ ( j ) x i refer to the j-th largest element values in h M x i and h N x i , respectively. Additionally, l M x i and l N x i represent the number of elements in h M x i and h N x i , respectively. l x i = max l M x i , l N x i .
The mathematical expression of the four-tuple attribute N i , L i , O i , F i obtained from FDS analysis is used to construct a hesitant fuzzy element set h A ( x ) = μ h N , μ h L , μ h O , μ h F as an evaluation attribute. μ h N , μ h L , μ h O , μ h F represent the hesitant fuzzy membership functions corresponding to each evaluation attribute, as shown in Table 1.
Definition 2. 
Let M E Q be a query expression, M E D B i i = 1 , 2 , , n be a dataset containing n mathematical expressions.
Definition 3. 
The formula for calculating the similarity between two mathematical expressions is as follows:
s i m M E Q , M E D B i = 1 1 4 1 l P p j = 1 l P p h M E Q σ ( j ) P p h M E D B i σ ( j ) P p λ 1 / λ
Here, h M E Q P p and h M E D B i P p represent the sets of hesitant fuzzy elements corresponding to M E Q and M E D B i , respectively. The elements in h M E Q P p and h M E D B i P p represent the membership degrees of each attribute value of expressions M E Q and M E D B i , respectively. l P p represents the number of evaluated attribute values, h M E Q σ ( j ) P p and h M E D B i σ ( j ) P p represent the degree of membership values of the σ ( j ) -th factor in h M E Q P p and h M E D B i P p , respectively.

4.3. Ranking Fusion of Multi-Modal Retrieval Results

The Reciprocal Rank Fusion (RRF) [30] combines the image and text modal retrieval ranking. We use Equation (7) to calculate the fusion score S RRF of a mathematical expression e, based on the retrieval ranking r ( e ) and a constant k to reduce the influence of highly scored documents.
S RRF ( e M E D B i ) = r R 1 k + r ( e )
To determine the optimal value for k, we conducted experiments with various values. The experimental results are presented in Table 2. As seen in Table 2, with the increase in the value of k, the MAP obtained from the fusion results initially increases and then decreases. The maximum MAP value is achieved at k = 60. Therefore, in our experiments, we choose a value of k equal to 60.
Illustratively, taking the query expression “ M E ” as an example, the top 5 mathematical expression IDs, ranks, and fusion scores for multimodal retrieval results are presented in Table 3. This culminates in a retrieval ranking of {3,4,2,1,6,5}. In this case, the top five results in the image-modal retrieval did not include the mathematical expression with ID = 5. However, the top five results in the text-modal retrieval contained the expression with ID = 5. This discrepancy might be due to the image modality retrieval method ranking this candidate expression lower (beyond the 5th position), resulting in a score of 0 in the image modality retrieval calculation. Consequently, S RRF ( i d 5 ) = 1 60 + 5 = 0.01538 .
This method is not affected by similarity scores and only depends on the ranking of retrieval results. It gives higher rankings to result items that a ranking model strongly prefers and ranks result items that are weakly preferred by multiple models less highly.

5. Experimental Results and Discussion

5.1. Experimental Dataset and Environment

We obtained 592,345 mathematical expressions from 14,274 articles in the ArXiv, a free and open dataset created by Cornell University researchers, and extracted 250,045 expressions from 11,770 articles in the NTCIR-mair-Wikipedia-corpus (NTCIR). We employ the method proposed by Xu et al. [31] to extract mathematical expressions images from scientific literature. The expressions are stored in both image and LaTeX text formats. The experimental environment is shown in Table 4.

5.2. Evaluation Protocol and Metrics

We randomly select ten representative mathematical expressions from the dataset for our experiments, as shown in Table 5. We employ Mean Average Precision (MAP) and Discounted Cumulative Gain (DCG) [32] to evaluate the effectiveness of our method in this paper. Three graduate students majoring in computer science have assigned a relevance judgment to the top 20 retrieval results for query expressions. Relevance judgments range from 1 (irrelevant) to 5 (highly relevant) based on the top-k retrieval results for all queries. This approach simulates user assessment of retrieval results and allows us to calculate Average Precision (AP), MAP, and nDCG to evaluate the effectiveness of our model.

5.3. System Experiment

The average precision and MAP were calculated for ten query expressions, as shown in Table 6.
Because there were more similar expressions in the ArXiv dataset, the results were consistently better than those from the NTCIR dataset. The low P@20 values are because there are fewer than 20 expressions in the dataset which are highly similar to some complex queries.

5.4. Ablation Experiments

Through experiments, we observed that the SLFE module prioritizes results that exhibit greater structural and length similarity to the query expression. The relevance scores of the retrieved expressions were used to calculate P@k, MAP@10, and nDCG@10 values. We evaluated the impact of SLFE re-ranking on image-modal retrieval by comparing P@k, MAP@10, and nDCG@10 values. The results in Figure 6 show that SLFE significantly improves both retrieval accuracy and ranking results by attending to symbol-level features.
The image-modal retrieval method is Method 1, the text-modal retrieval method is Method 2, and the image-text modal RRF method is Method 3. For the query O ( log 1 γ ) , the top 5 results from Method 1 and Method 2 are shown in Table 7. Both methods retrieve exact matches as the first formula. Method 1 places a stronger emphasis on the structural information of mathematical expressions. For instance, because the candidate expression O ˜ ( log 1 γ ) closely aligns with the query expression O ( log 1 γ ) in terms of structure, it is ranked higher. Additionally, method 1 can retrieve formulas that share the same structure but have different symbolic representations, such as expression O ( log ε γ ) . On the other hand, method 2 prioritizes the symbols within the expressions, such as candidate expressions that all contain log 1 γ .
The average P@k and MAP@k of the three methods are summarized in Figure 7. Text-modal retrieval outperformed image-modal retrieval because FDS is more concerned with the type of the symbol itself (e.g., whether the symbol is an operator, operand, constant, or variable). However, we found that within the results of image modality retrieval, expressions more structurally similar to the query image are ranked higher. This suggests that image modality retrieval methods are more adept at capturing structural information inherent in expressions. The integration of both modalities can lead to more accurate and effective retrieval results. The image-text modal RRF method was effective, as the fusion ranking of two modal retrieval results was better than the single modal retrieval ranking, demonstrating the complementary results of the two modal retrievals.

5.5. Contrast Experiments

SearchOnMath [33] is a retrieval tool for scientific literature and Wikipedia pages based on mathematical expressions. Tangent-CFT [34] is a mathematical expression embedding model that represents mathematical expressions using Operator trees (OPTs) and Symbol Layout Trees (SLT) and ultimately generates formula embeddings using fastText. Experimental trials were conducted utilizing the mathematical expressions from Table 5. A comparative analysis was performed between our proposed approach and other methods. The results for MAP@k and average P@k are presented in Table 8. The proposed image-text modal RRF method achieves an accuracy of 0.660 and an average precision of 0.774 for the top 10 results. The nDCG@10 results for each expression retrieval outcome under various methods are depicted in Figure 8.
The method proposed in this paper, which combines the attribute features of mathematical expression text and images, is able to simultaneously consider both the holistic characteristics and symbol-level features of the image modality, as well as the symbol types characteristic of the text modality. This results in ranking outcomes that better align with user requirements. Therefore, the NDCG@10 value of the method proposed in this paper is generally higher compared to the contrastive methods. In comparison to the SearchOnMath and Tangent-CFT methods, our approach offers user convenience, avoiding the cumbersome manual input of mathematical expressions. Additionally, it effectively balances the structural information of the image modality and the symbolic attributes of the text modality, optimizing the output ranking. Consequently, it efficiently circumvents the issue of mathematical expression text losing structural features.

6. Conclusions

We have constructed a multimodal mathematical expression retrieval model by leveraging the RRF method to fuse the retrieval results from both image and text modalities. In this model, the image-modal retrieval module employs the ConvNext and SLDE modules to extract mathematical expression image features and symbol-level features, respectively. Based on the similarity of these features, the results are then outputted in descending order. Meanwhile, the text-modal retrieval module analyzes and extracts attribute values of textual expressions using FDS. To determine the hesitant fuzzy similarity between the candidate expression and the mathematical query expression, the hesitant fuzzy set theory is introduced in the meantime. The RRF amalgamates image and text modality retrieval outcomes, culminating in the ultimate combined result. Concurrently, a subset of literature from the publicly accessible ArXiv dataset was extracted and utilized to construct a dataset encompassing 592,345 mathematical expressions.
The multimodal retrieval model eliminates constraints on the user’s input modalities for mathematical query expressions, thereby ensuring a comprehensive evaluation of both modality similarities. This approach enhances the rationality of retrieval outcomes.
In the future, our research will further explore multimodal mathematical expression retrieval in the following areas:
  • Enhancing image modality feature extraction by incorporating contextual parsing of mathematical expressions, thereby preserving a richer semantic understanding.
  • Extending the application of the proposed method to scientific literature retrieval. This involves integrating mathematical expressions with intrinsic attributes of scientific literature, including keywords, to enhance retrieval accuracy during scientific literature searches.

Author Contributions

Data curation, R.L. and J.W.; Formal analysis, R.L., J.W. and X.T.; Funding acquisition, X.T.; Investigation, R.L.; Methodology, R.L., J.W. and X.T.; Software, R.L.; Supervision, X.T.; Validation, R.L. and X.T.; Visualization, R.L. and X.T.; Writing—original draft, R.L.; Writing—review & editing, R.L., J.W. and X.T. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the Natural Science Foundation of Hebei Province of China (Grant No. F2019201329).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available upon request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, X.; Zhu, Y.; Liu, S.; Ju, J.; Qu, Y.; Cheng, G. Dyrren: A dynamic retriever-reranker-generator model for numerical reasoning over tabular and textual data. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; Volume 37, pp. 13139–13147. [Google Scholar]
  2. Satpute, A.; Greiner-Petter, A.; Schubotz, M.; Meuschke, N.; Aizawa, A.; Gipp, B. TEIMMA: The First Content Reuse Annotator for Text, Images, and Math. arXiv 2023, arXiv:2305.13193. [Google Scholar]
  3. Gipp, B.; Greiner-Petter, A.; Schubotz, M.; Meuschke, N. Methods and Tools to Advance the Retrieval of Mathematical Knowledge from Digital Libraries for Search-, Recommendation-, and Assistance-Systems. arXiv 2023, arXiv:2305.07335. [Google Scholar]
  4. Xu, P.; Zhu, X.; Clifton, D.A. Multimodal learning with transformers: A survey. arXiv 2023, arXiv:2206.06488. [Google Scholar] [CrossRef] [PubMed]
  5. Suzuki, M.; Matsuo, Y. A survey of multimodal deep generative models. Adv. Robot. 2022, 36, 261–278. [Google Scholar] [CrossRef]
  6. Goel, M.; Goyal, V.; Venktesh, V. MWPRanker: An Expression Similarity Based Math Word Problem Retriever. arXiv 2023, arXiv:2307.01240. [Google Scholar]
  7. Pfahler, L.; Morik, K. Self-Supervised Pretraining of Graph Neural Network for the Retrieval of Related Mathematical Expressions in Scientific Articles. arXiv 2022, arXiv:2209.00446. [Google Scholar]
  8. Schellenberg, T.; Yuan, B.; Zanibbi, R. Layout-based substitution tree indexing and retrieval for mathematical expressions. In Proceedings of the Document Recognition and Retrieval XIX, SPIE, Burlingame, CA, USA, 25–26 January 2012; Volume 8297, pp. 126–133. [Google Scholar]
  9. Hu, X.; Gao, L.; Lin, X.; Tang, Z.; Lin, X.; Baker, J.B. Wikimirs: A mathematical information retrieval system for wikipedia. In Proceedings of the 13th ACM/IEEE-CS Joint Conference on Digital Libraries, Indianapolis, IN, USA, 22–26 July 2013; pp. 11–20. [Google Scholar]
  10. Zhong, W.; Rohatgi, S.; Wu, J.; Giles, C.L.; Zanibbi, R. Accelerating substructure similarity search for formula retrieval. In Proceedings of the Advances in Information Retrieval: 42nd European Conference on IR Research, ECIR 2020, Lisbon, Portugal, 14–7 April 2020; pp. 714–727. [Google Scholar]
  11. Gao, L.; Jiang, Z.; Yin, Y.; Yuan, K.; Yan, Z.; Tang, Z. Preliminary Exploration of Formula Embedding for Mathematical Information Retrieval: Can mathematical formulae be embedded like a natural language? arXiv 2017, arXiv:1707.05154. [Google Scholar]
  12. Dadure, P.; Pakray, P.; Bandyopadhyay, S. Embedding and generalization of formula with context in the retrieval of mathematical information. J. King Saud-Univ.-Comput. Inf. Sci. 2022, 34, 6624–6634. [Google Scholar] [CrossRef]
  13. Peng, S.; Yuan, K.; Gao, L.; Tang, Z. Mathbert: A pre-trained model for mathematical formula understanding. arXiv 2021, arXiv:2105.00377. [Google Scholar]
  14. Dai, Y.; Chen, L.; Zhang, Z. An N-ary tree-based model for similarity evaluation on mathematical formulae. In Proceedings of the 2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Toronto, ON, Canada, 11–14 October 2020; pp. 2578–2584. [Google Scholar]
  15. Marinai, S.; Miotti, B.; Soda, G. Mathematical symbol indexing using topologically ordered clusters of shape contexts. In Proceedings of the 2009 10th International Conference on Document Analysis and Recognition, Barcelona, Spain, 26–29 July 2009; pp. 1041–1045. [Google Scholar]
  16. Zanibbi, R.; Yuan, B. Keyword and image-based retrieval of mathematical expressions. In Proceedings of the Document Recognition and Retrieval XVIII, SPIE, San Francisco, CA, USA, 26–27 January 2011; Volume 7874, pp. 141–149. [Google Scholar]
  17. Torra, V.; Narukawa, Y. On hesitant fuzzy sets and decision. In Proceedings of the 2009 IEEE International Conference on Fuzzy Systems, Jeju Island, Republic of Korea, 20–24 August 2009; pp. 1378–1382. [Google Scholar]
  18. Farhadinia, B.; Aickelin, U.; Khorshidi, H.A. Uncertainty measures for probabilistic hesitant fuzzy sets in multiple criteria decision making. Int. J. Intell. Syst. 2020, 35, 1646–1679. [Google Scholar] [CrossRef]
  19. Liu, P.; Zhang, X. A new hesitant fuzzy linguistic approach for multiple attribute decision making based on Dempster–Shafer evidence theory. Appl. Soft Comput. 2020, 86, 105897. [Google Scholar] [CrossRef]
  20. Farhadinia, B.; Aickelin, U.; Khorshidi, H.A. Higher order hesitant fuzzy Choquet integral operator and its application to multiple criteria decision making. arXiv 2020, arXiv:2011.08183. [Google Scholar]
  21. Cai, L. Interval-Valued Hesitant Fuzzy Sets and Its Application to Decision Making. Ph.D. Thesis, Zhengzhou University, Zhengzhou, China, 2013. [Google Scholar]
  22. Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A convnet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11976–11986. [Google Scholar]
  23. Tian, X. A mathematical indexing method based on the hierarchical features of operators in formulae. In Proceedings of the 2nd International Conference on Automatic Control and Information Engineering (ICACIE 2017), Hong Kong, China, 26–28 August 2017; pp. 49–52. [Google Scholar]
  24. Torra, V. Hesitant fuzzy sets. Int. J. Intell. Syst. 2010, 25, 529–539. [Google Scholar] [CrossRef]
  25. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  26. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  27. Samet, H.; Tamminen, M. Efficient component labeling of images of arbitrary dimension represented by linear bintrees. IEEE Trans. Pattern Anal. Mach. Intell. 1988, 10, 579–586. [Google Scholar] [CrossRef]
  28. Zadeh, L.A. Fuzzy sets. Inf. Control. 1965, 8, 338–353. [Google Scholar] [CrossRef]
  29. Xu, Z.; Xia, M. Distance and similarity measures for hesitant fuzzy sets. Inf. Sci. 2011, 181, 2128–2138. [Google Scholar] [CrossRef]
  30. Cormack, G.V.; Clarke, C.L.; Buettcher, S. Reciprocal rank fusion outperforms condorcet and individual rank learning methods. In Proceedings of the 32nd International ACM SIGIR Conference on Research and Development in Information Retrieval, Boston, MA, USA, 19–23 July 2009; pp. 758–759. [Google Scholar]
  31. Xu, X.; Tian, X.; Yang, F. A retrieval and ranking method of mathematical documents based on CA-YOLOv5 and HFS. Math. Biosci. Eng. 2022, 19, 4976–4990. [Google Scholar] [CrossRef] [PubMed]
  32. Jarvelin, K.; Kekalainen, J. IR evaluation methods for retrieving highly relevant documents. In Proceedings of the ACM SIGIR Forum; ACM: New York, NY, USA, 2017; Volume 51, pp. 243–250. [Google Scholar]
  33. Oliveira, R.M.; Gonzaga, F.B.; Barbosa, V.C.; Xexéo, G.B. A distributed system for SearchOnMath based on the Microsoft BizSpark program. arXiv 2017, arXiv:1711.04189. [Google Scholar]
  34. Mansouri, B.; Rohatgi, S.; Oard, D.W.; Wu, J.; Giles, C.L.; Zanibbi, R. Tangent-CFT: An embedding model for mathematical formulas. In Proceedings of the 2019 ACM SIGIR International Conference on Theory of Information Retrieval, Santa Clara, CA, USA, 2–5 October 2019; pp. 11–18. [Google Scholar]
Figure 1. Framework diagram of mathematical expression multi-modal retrieval.
Figure 1. Framework diagram of mathematical expression multi-modal retrieval.
Electronics 12 04363 g001
Figure 2. Network structure of ConvNeXt-Tiny.
Figure 2. Network structure of ConvNeXt-Tiny.
Electronics 12 04363 g002
Figure 3. Example image of segmentation (a) adhered grayscale image (b) two components after segmentation (c) symbol block.
Figure 3. Example image of segmentation (a) adhered grayscale image (b) two components after segmentation (c) symbol block.
Electronics 12 04363 g003
Figure 4. Position vector feature map of the element symbol block.
Figure 4. Position vector feature map of the element symbol block.
Electronics 12 04363 g004
Figure 5. Sample plot of the baseline in the mathematical expression.
Figure 5. Sample plot of the baseline in the mathematical expression.
Electronics 12 04363 g005
Figure 6. Ablation experimental results of the SLEF module.
Figure 6. Ablation experimental results of the SLEF module.
Electronics 12 04363 g006
Figure 7. The average P@k and MAP@k values of the three methods.
Figure 7. The average P@k and MAP@k values of the three methods.
Electronics 12 04363 g007
Figure 8. Comparison of nDCG@10 of the proposed method and other methods.
Figure 8. Comparison of nDCG@10 of the proposed method and other methods.
Electronics 12 04363 g008
Table 1. Formulas of hesitant fuzzy membership functions for evaluation attributes.
Table 1. Formulas of hesitant fuzzy membership functions for evaluation attributes.
Membership Function FormulaFunction Description
μ h N M E D B i , M E Q = e Number D B i Number Q σ 2 σ is a balancing factor that ensures the value of μ h N remains in the range of [0,1].
μ h L M E D B i , M E Q = e α Level D B i Level Q α is a balancing factor that ensures the value of μ h L remains in the range of [0,1].
μ h O M E D B i , M E Q = s o , O p e r a t o r x O p e r a t o r x = 1 indicates that the current symbol is an operator, and  s o = 1 . Otherwise, the current symbol is an operand, s o = 0.5 .
μ h F M E D B i , M E Q = f o , F l a g ( DB i , Q ) F l a g ( D B i , Q ) is used to determine the spatial positional relationship of identical terms in two mathematical expressions. If the positional relationship between the two is the same, f o = 1 ; otherwise, f o = 0 .
Table 2. The influence of k on the final MAP value of RRF fusion results.
Table 2. The influence of k on the final MAP value of RRF fusion results.
k20406080100
MAP0.7630.7670.7740.7710.769
Table 3. Illustration of fusion score calculation for multimodal retrieval results.
Table 3. Illustration of fusion score calculation for multimodal retrieval results.
ID of ResultRank of ImageRank of TextFusion Score
154 S RRF ( i d 1 ) = 1 60 + 5 + 1 60 + 4 = 0.03100
242 S RRF ( i d 2 ) = 1 60 + 4 + 1 60 + 2 = 0.03175
331 S RRF ( i d 3 ) = 1 60 + 3 + 1 60 + 1 = 0.03226
423 S RRF ( i d 4 ) = 1 60 + 2 + 1 60 + 3 = 0.03200
5Null5 S RRF ( i d 5 ) = 1 60 + 5 = 0.01538
61Null S RRF ( i d 6 ) = 1 60 + 1 = 0.01639
Table 4. Experimental environment.
Table 4. Experimental environment.
Experimental EnvironmentConfiguration
ProcessorAMD EPYC
RAM16 GB
Operating systemWindows 10
Graphics cardRTX A5000
Video memory24 GB
Python version3.8
Pytorch version1.11.0
Table 5. 10 query expressions in system experiment.
Table 5. 10 query expressions in system experiment.
No.Query ExpressionsNo.Query Expressions
1 M ( x , y ) = 1 2 1 erf π d ( x , y ) δ 6 cos ( θ )
2 a b 7 L = i = 1 N p η i , E i , ϕ i , , η 0
3 K = f · μ x γ u x 0 f · μ y u y 0 0 1 8 Ω ( t ) = k = 0 H k ( t )
4 x + y 9 Ω ^ s = 0 τ P ^ ( S T = t , H ) d t
5 σ ( x ) = 1 1 + exp ( x ) 10 r ˙ ( 0 ) = Δ lim τ 0 d r ( τ ) d τ
Table 6. The average precision and MAP@k of the multi-modal retrieval method on the two datasets.
Table 6. The average precision and MAP@k of the multi-modal retrieval method on the two datasets.
DatasetP@3P@5P@10P@20MAP@3MAP@5MAP@10
ArXiv0.8330.8000.6900.5000.8580.8330.787
NTCIR0.8000.7200.6300.4050.8250.8180.760
Table 7. Method 1 vs. Method 2 results for query: O ( log 1 γ ) .
Table 7. Method 1 vs. Method 2 results for query: O ( log 1 γ ) .
RankMethod 1Method 2
1 O ( log 1 γ ) O ( log 1 γ )
2 O ˜ ( log 1 γ ) d · O ( log 1 γ )
3 O ( log ε γ ) 1 γ 2 O ( log 1 γ )
4 O ( C link · d log 1 γ ) T = O ( log 1 γ )
5 O ˜ ( 1 γ 2 log 2 ( θ ε ) ) O ( C link · d log 1 γ )
Table 8. Comparison of P@k and MAP@k between other methods and the proposed method.
Table 8. Comparison of P@k and MAP@k between other methods and the proposed method.
MethodP@3P@5P@10MAP@3MAP@5MAP@10
Image modal retrieval method0.7840.6800.5350.7960.7870.749
Text modal retrieval method0.8330.7200.5900.8160.8050.760
Image-text RRF method0.8170.7600.6600.8420.8260.774
Search on Math0.6000.6200.5200.7170.7090.664
Tangent-CFT0.8200.6830.5830.8090.7850.739
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, R.; Wang, J.; Tian, X. A Multi-Modal Retrieval Model for Mathematical Expressions Based on ConvNeXt and Hesitant Fuzzy Set. Electronics 2023, 12, 4363. https://doi.org/10.3390/electronics12204363

AMA Style

Li R, Wang J, Tian X. A Multi-Modal Retrieval Model for Mathematical Expressions Based on ConvNeXt and Hesitant Fuzzy Set. Electronics. 2023; 12(20):4363. https://doi.org/10.3390/electronics12204363

Chicago/Turabian Style

Li, Ruxuan, Jingyi Wang, and Xuedong Tian. 2023. "A Multi-Modal Retrieval Model for Mathematical Expressions Based on ConvNeXt and Hesitant Fuzzy Set" Electronics 12, no. 20: 4363. https://doi.org/10.3390/electronics12204363

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop