Next Article in Journal
Mathematical Problem Solving through the Lens of Ethics and Aristotelian Attitude: A Case Study
Next Article in Special Issue
Development of a Multilingual Model for Machine Sentiment Analysis in the Serbian Language
Previous Article in Journal
Stability and Optimal Control of Tree-Insect Model under Forest Fire Disturbance
Previous Article in Special Issue
Short Answer Detection for Open Questions: A Sequence Labeling Approach with Deep Learning Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Entity-Matching System Based on Multimodal Data for Two Major E-Commerce Stores in Mexico

by
Raúl Estrada-Valenciano
1,
Víctor Muñiz-Sánchez
2 and
Héctor De-la-Torre-Gutiérrez
1,*
1
Aguascalientes Campus, Centro de Investigación en Matemáticas, A. C., Calzada de la Plenitud 103, José Vasconcelos Calderón, Aguascalientes 20200, Mexico
2
Monterrey Campus, Centro de Investigación en Matemáticas, A. C., Km. 10 Autopista al Aeropuerto, Parque de Investigación e Innovación Tecnológica (PIIT), Av. Alianza Centro 502, Apodaca 66628, Mexico
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(15), 2564; https://doi.org/10.3390/math10152564
Submission received: 31 May 2022 / Revised: 15 July 2022 / Accepted: 20 July 2022 / Published: 23 July 2022

Abstract

:
E-commerce has grown considerably in Latin America in recent years due to the COVID-19 pandemic. E-commerce users in English-speaking and Chinese-speaking countries have web-based tools to compare the prices of products offered by various retailers. The task of product comparison is known as entity matching in the data-science domain. This paper proposes the first entity-matching system for product comparison in Spanish-speaking e-commerce. Given the lack of uniformity of e-commerce sites in Mexico, we opted for a bimodal entity-matching system that uses the image and textual description of products from two of the largest e-commerce stores in Mexico. State-of-the-art techniques in natural language processing and machine learning were used to develop this research. The resulting system achieves F1 values of approximately 80%, representing a significant step towards consolidating a product-matching system in Spanish-speaking e-commerce.

1. Introduction

Year after year, e-commerce is becoming more common among the general population. On the one hand, users choose this medium for its practicality and efficiency in finding specific products. On the other hand, merchants benefit from the reach and ease of reaching potential buyers. This supply-and-demand phenomenon has generated considerable growth in the number of items available on these platforms. In an environment with millions of products offered every day, providing a good experience in searching and purchasing items brings many opportunities and many challenges.
One challenge that arises with the increasing number of offers is to automatically identify which of them refers to the product of interest. Doing this makes it easier for the buyer to find different proposals for the same product and allows the seller to know the market of offered products. This problem is known as product matching and is a particular case of entity matching.
Entity-matching (EM) and product-matching (PM) problems have been widely studied for several years; EM has the longest history as the general case. Many models have been tested and show a constant improvement in the performance of these tools over time.
Let us place ourselves in the current context of Latin America, and specifically, Mexico. An increase in e-commerce purchases has been observed as a result of the health emergency caused by the SARS CoV-2 virus (COVID-19) and the lockdown imposed by health authorities. However, the use of e-commerce platforms in Latin America clearly lags behind the use of similar platforms in the United States, China and the United Kingdom.
Consequently, in Latin America, the mere design of an e-commerce platform can be very different from platforms targeted at consumers in countries with a more developed market, and, while those platforms are increasingly generalizing the use of schemes such as those provided by schema.org for the presentation of their products, indeed, it is not certain whether the platforms in Latin America, and specifically in Mexico, will follow the same trend.
The task of PM becomes more complicated within platforms that offer products whose textual descriptions are not clear and homogeneous due to the impossibility of automatically locating specific characteristics of the product that are essential to compare with similar products. It is possible, for example, that the position of the name within the web page changes with each product or that the brand name is explicitly displayed for certain products and not for others. Perhaps this is why PM has focused mainly on Chinese or English-speaking platforms. The closest thing we have in Latin America is the implementation of a tool for the extraction of product features [1]; however, the PM problem is not addressed as such. This scarcity of research in the Latin American region opens an interesting field of study because several existing tools can be applied to this type of problem. However, it is not expected that the tools will work the same way they do on non-Spanish platforms, so it is essential to test them and develop a unique methodology for this case.
An important aspect to consider is the multimodal nature of the data related to the PM task. When the information related to some phenomenon of interest comes from different sources, sensors, or instruments, we say it is multimodal [2]. It is particularly evident with the information shared by users or companies on digital media or social networks, where the information related to some item (user, product, news, post, etc.) is composed of images, text, audio, or video. Traditionally, many tasks related to this kind of data (PM included) have been addressed by considering only one modality of the information. In recent years, there has been an increasing number of research efforts to model and combine the information from each modality of data to solve specific tasks related to them. Methodologically, there is a preference for classical-machine-learning (CML) algorithms within the literature for solving the PM task, whether a unimodal or multimodal approach is considered, and state-of-the-art (SOTA) results have been achieved by using deep-learning (DL) methods. There are many advantages to using CML and DL methods in both performance metrics and in the fact that the amount of training data required is greatly reduced due to the pre-trained models available, which allows transfer learning, and, as a consequence, we can obtain better generalization. For these reasons, we propose a multimodal data approach based on DL and CML methodologies to solve the PM task for the particular case of Mexican e-commerce.
Given the lack of resources and research for the Mexican e-commerce market, this work aims to develop and test a system capable of recognizing the similarity between products offered on different e-commerce sites in Mexico. This system will be based on ML techniques and will offer the consumer and the business a clearer picture of the products they intend to obtain and trade, respectively, by automatically identifying, within a large number of offers, which one corresponds to the product of interest. We build a database of products offered on e-commerce websites operating in Mexico to this end. As the PM problem has not been studied in the Spanish language, any database with this characteristic will have to be built from scratch. This is not an easy task; due to the complexity of the PM problem, it will not be possible to use data from all e-commerce stores operating in Mexico. Therefore, it is necessary to define some criteria. In our case, we decided to focus on the two most popular e-commerce stores in Mexico. Names are not disclosed due to the confidentiality of the information. The stores will be referred to as e-shop 1 and e-shop 2, hereafter. In addition, we restrict ourselves to the most popular products offered in these e-commerce stores. We obtained two modalities of data for each product: text and images.
With this multimodal data, we implemented and compared several ML methodologies to address the PM problem as a binary classification problem: the objective is to identify if a given pair of offers corresponds to the same product. Initially, we used some CML algorithms based on a particular representation of the text description of the data. In addition, we used some of the SOTA models for PM based on DL architectures that use only one modality of the data (image or text) independently. We propose some multimodal algorithms based on the intermediate fusion of some useful vector representations of images and text.
Another objective we pursue with this research is to know how the Mexican market and e-commerce platforms present their products and what structure (if any) they follow to organize their characteristics. We think this work could set a precedent for future research on PM and the obstacles to working with these data for the Mexican market and the Spanish language.
In summary, this research develops a product-matching scheme, which is the first focused on e-commerce in Spanish-speaking countries and considers the Mexican market’s particularities. The proposed scheme, which also considers the difficulties of the Mexican market, is bimodal and has not been explored in Chinese- or English-speaking e-commerce markets. The proposed bimodal product-matching system also explores a wide range of SOTA, CML, and DL algorithms in EM, which will allow measuring their efficiency in the Mexican market and the Spanish language.
Section 2 provides an overview of related papers that deal with similar topics. The definition of the problem, the data set, and the methodology we propose are described in Section 3. Section 4 presents the result of our study with regard to the stated objectives. Based on these results, Section 5 provides further key findings and challenges and envisions future research directions to master them.

2. Related Work

Since the advent of e-commerce, several ML-based solutions to specific tasks have been proposed, ranging from prediction, web mining, data gathering and preprocessing [3,4,5,6]. With the rise in the big-data era and disruptive information technologies (IT), methodologies for solving traditional and new tasks related to e-commerce have evolved too, from classical-ML to artificial-intelligence (AI) approaches [7]. This evolution allows us to face large-scale prediction problems such as extreme classification, where multi-label prediction with hundreds of thousands or even millions of labels are involved [8,9,10,11], and is particularly relevant in PM, when there may not be an exact match but may be an approximate match, as is the case in recommendation systems.
To give some context to our particular PM problem and to reinforce ideas that were discussed earlier, a brief literature review will follow. Starting with classical methods, ref. [12] addresses the MS problem by asking the question of how similar is similar? The question is answered by constructing an algorithm capable of finding the appropriate similarity function to remove ambiguity. Ref. [13] proposed a methodology for the formalization and evaluation of product-category matching schemes, cost-effectively and reliably. Another classical approach is given by [14], who bases the study on Markov logic and combines first-order logic and probabilistic models for solving the EM task.
As for ML-based approaches, ref. [15] performs work for the PM task using a document representation called term frequency–inverse document frequency (TF-IDF). This representation is described in Section 3.3.1.2. Roughly speaking, it is a numerical measure expressing how relevant a word is within a document. Similarly, the authors use the Word2vec technique [16], which assigns each word a vector in space so that words that share the same context are closer than others. Both techniques, together with a measure for the similarity between two vectors called cosine similarity, make it possible to create a model capable of calculating the similarity between two objects. The results showed that TF-IDF is the more efficient and straightforward of the two techniques and that Word2Vec requires a large amount of data to obtain good results.
A work that employs both classical and ML-based approaches was developed by [17]. Here, several problems within the EM domain, including PM, are addressed by using the structure to evaluate the relationship between entities (FEVER) to implement and adjust each model used. Classical methods include PPJoin [18] which, by defining a threshold and a similarity function, determines how similar two entities are. The method proposed by [19], also classical, works similarly but defines lower and upper bounds. For methods based on ML techniques, FEBRL [20] and MARLIN [21] were used, which are based on support vector machine (SVM) and decision trees, respectively. Ref. [22] developed a neural-network-based product-matching system using only the title and attributes of the offered product.
A vital system based on ML techniques is the system designed by [23] (Magellan), as it provides a step-by-step guide on how to act in each EM scenario, as well as proposing techniques and models for data cleaning, visualization, and learning. The above proposals help the user to address any problem related to EM.
For further references to these and other models, ref. [24] provides an overview of the aspects to consider for the resolution of this task, from basic concepts to a vast number of strategies and methods to follow for the type of task to be solved.
One of the first DL works carried out concerning the EM task was developed by [25] and presents a system that they called DeepER, which uses a compendium of DL techniques and models, achieving excellent performance. In this system, recurrent neural networks (RNN), bidirectional recurrent networks (BRNN), and long short-term memory (LSTM) units convert each pair of products into a vector that can be used to capture similarities between pairs. In addition, that system considers cases where there are pre-trained word embeddings (vector representations of words that are derived from training a neural network), as well as cases where they are not. Methods to improve the vectors for each special case of MS are presented. The results of implementing this system on several databases showed that DeepER performs better than previously developed systems. Ref. [26] utilized EM techniques to develop a team-manager matching scheme that includes performance metrics and information preprocessing.
Another DL-based model is developed by [27], which tests the benefits and limitations derived from using DL methods for the EM task. Their work examines several DL-based solutions that have been developed for similar tasks, creating a solution space shaped by DL methods: smooth inverse frequency (SIF), RNN, attention models, and a combination of the latter two. The study reports the benchmark for this task and some exceptional cases, such as PM. The authors of [28] present an approach that combines DL techniques, natural language processing (NLP), and standard PM methods. For this, structured data are used as training for feature-extraction models in unstructured databases. These models are: dictionary-based [29], conditional random field (CRF) [30], CRF with word embeddings to improve generalization (http://nlp.stanford.edu/software/CRF-NER.shtml, accessed on 5 October 2021) [16,31,32] and an attribute extraction model for images. Once the unstructured data set features are obtained, a similarity vector is computed for each pair, which will be the input information for classification models such as decision trees, SVMs, logistic regression, and Bayesian classifiers. A portion of the databases is manually labeled as matched or mismatched pairs to train these models. The results obtained indicate that the CRF model incorporating word embeddings and feature extraction from images, in addition to the classifier based on decision trees, has the best performance. The authors of [33] highlight the fact that several e-commerce sites use a scheme given by schema.org (https://schema.org/, accessed on 12 January 2022) to publish offers for their products. Thanks to this, the information and structure that the Internet offers are being homogenized little by little. Among the information that more and more sites are including are unique product identifiers in their offers, which allow identical products to be grouped without manual work. With the help of this identifier, the authors formed a database with pairs of products, either identical (positive) or different (negative) from each other, belonging to four categories: computers, cameras, watches, and shoes. Subsequently, several models are tested for PM, such as the one proposed by [23], word co-occurrence, and several models with the structure of the Deepmacher [27]. The results show that Deepmacher models are the most efficient. However, their generalization ability is limited. The authors of [34] address the PM problem through two different approaches, modeling it as a classification task using a deep neural network and as a similarity task using a Siamese neural network. The latter consists of two parallel neural networks, each fed with one of the two products to be compared with the corresponding binary label. In this case, the model is not learning to classify but learning to differentiate the two inputs. The Siamese network obtained the best results. However, the deep neural network was faster to train and achieved similar results. The authors of [35] propose a model for PM that consists of two modules, one in charge of title matching and the other for attribute matching. The first one combines an RNN with a convolutional neural network (CNN). The second one only uses a CNN to perform the task, resulting in better performance than other SOTA EM models.
Recently, the field of DL for problem solving related to NLP has been revolutionized with the creation of BERT, a DL model based on transformers, positioning itself as the state-of-the-art approach for many of these tasks. The authors of [36] are the first to implement BERT for the PM task, demonstrating that, simply by performing fine tuning on a pre-trained BERT model, comparable results to other SOTA models can be achieved because the BERT model is more data-efficient. The authors also show that, by training the pre-trained network with product-offering information, the performance of the network on unobserved data can be improved without the need for fine tuning. Overall, it is shown that using BERT for the PM task is quite efficient in terms of data used and generalization.
The scientific community quickly improved BERT by creating several architectures based on it. These were exploited by [37] to solve the PM task, specifically using two of the BERT enhancements, RoBERTa [38] and DistilBERT [39]. Such models take a pre-trained model and connect it to a simple network that performs the classification task. This work also addresses the problem of product classification, which consists of extracting attributes from the description of a product. For this, [37] employ the architecture of a CNN [40] because the task is understood as a schema in which one has one input text and multiple labels as output. The results for the PM task show good performance for the combination of RoBERTa and DistilBERT, achieving up to an F1 of 95.8 for training with 15 epochs and considering the inclusion of all attributes.
As in the work performed by [35,37] construct a system based on pre-trained transformer models, which they call DITTO. For this, it performs a fine-tuning process on models such as BERT, DistilBert and RoBERTa, pretrained on a large amount of data, which greatly improves the performance of the classifier. In addition, hierarchical optimization techniques were developed for the classifier to improve its performance. For learning DITTO on the specific EM task, data-augmentation techniques are implemented by including ‘difficult’ instances, further increasing its classification capability. Consequently, DITTO can reach similar results to other SOTA models using almost half of the labeled data, being quite efficient in real large-scale problems.
Finally, the study by [41] describes their work carried out for the Semantic Web Challenge on the PM task. In this work, four SOTA models are implemented for the MS task: MPM [42], Seq2SeqMatcher [43], HierMatcher [44] and DITTO [25]. In addition, post-processing was performed by means of a heuristic to detect false positives or false negatives.
The outstanding performance shown by researchers at the University of Mannheim is noteworthy in the area of PM because, in addition to having several publications on the subject [28,33,36], they also make public their WDC database (http://webdatacommons.org/largescaleproductcorpus/v2/index.html, accessed on 20 September 2021) for the use of the scientific community, as well as challenges. This keeps the research in the area active and encourages research in the area.

3. Material and Methods

3.1. Problem Definition

We will start by formally defining the EM problem the same way as [27] presents it in the article on the DeepMatcher tool. For this, any observable object in the real world (e.g., person, organization, etc.) will be called an entity. Likewise, any reference to an entity, for example, in the form of a record in a database, will be called an entity representation.
Let D and D be two collections of representations. It is assumed that the elements in D and D follow the same structure (e.g., each has specific attributes A 1 , , A N if the collection is structured data). The goal of EM is to find all pairs of representations between D and D that refer to the same entity. These pairs will be called positive pairs; conversely, pairs of distinct entities will be called negative pairs. Usually, the product D   ×   D is filtered to obtain the candidate set C, which only includes pairs that have been manually identified as positive or negative. Once the set C is created, a model is designed to identify these pairs.
For the matching model, two collections, D and D , and a candidate set C, which contain entity representations ( e 1 D , e 2 D ), will be used as input. We define T, a set of tuples { ( e 1 i , e 2 i , l ) } i = 1 | T | where { ( e 1 i , e 2 i ) } i = 1 | T | C , and l is a label that can take the values {“Positive”, “Negative”}, if they are identical entities or if they are distinct, respectively. Thus, given the labeled set T, the goal is to design an algorithm M that can distinguish between positive and negative pairs ( e 1 , e 2 ) in C.
Now, for the particular case of the PM task, we have that the entity to observe is an offer, and its representation is everything that comprises the web page where it is advertised. That is to say, an offer can be represented through its name, brand, price, characteristics, and even its image, as long as all this information is found on the reference page. However, the same product can be advertised on different web pages within different e-commerce platforms; this causes us to have different representations for the same product, and, hence, we can build sets D and D of representations that correspond to two different e-commerce platforms. Thus, depending on how similar these e-commerce platforms are or how much scope they have, they will come to offer, to a greater or lesser extent, identical products, being able to construct a set T of pairs of labeled offers. The problem remains the same: build a matching model from the set T that can distinguish whether two pairs of representations refer to the same product.

3.2. The Data Set

A crucial point in the solution to the PM problem is the construction of the data set. As mentioned above, the University of Mannheim freely offers unique databases for training and testing PM algorithms. In turn, these databases are obtained from the web-page repository Common Crawl (https://commoncrawl.org/, accessed on 1 October 2021), which is a non-profit organization that collects web pages from different platforms and makes them available to the general public. However, this repository crawls a limited number of websites, none of which are written in Spanish, and is only updated once a month. As a result, the information collected is slightly outdated and does not contain websites written in Spanish or targeted toward buyers in Latin America. Therefore, it was necessary to extract product information using web-scraping tools to obtain the most current data possible and from web pages of interest.
As it is impossible to extract products for all the e-commerce stores operating in Mexico, the three most popular e-shops with a significant presence in Mexico and Latin America, which we named e-shop 1, e-shop 2 and e-shop 3, were initially selected. One was discarded because we observed that the information of most of the products did not follow any standard structure on the web page, making it difficult to gather the data with the web-scraping procedures we implemented. Then, we focused only on e-shop 1 and e-shop 2.
Before performing the offers extraction, a market analysis was carried out to test the EM system on the products that are most often searched for on the Internet. For this purpose, the trends page from e-shop 2 was used. This page contains a list of 40 phrases or words that were the most popular the day before the visit. Portable fans, remote control carts, and blankets for babies are some examples of these searches. Each popular target search is linked to a certain number of products so that when using the search engine of e-shop 2, the products related to the search performed will be displayed. The analysis consisted of extracting the name, price, category, and quantity of items sold from the first 50 products of each search, approximately 2000 items. This process was performed daily (although not exhaustively), from 1 September 2020, to 31 January 2021. We decided to gather and process data from footwear products based on this analysis. The details and results of this analysis can be found in the Appendix A.

3.2.1. Pair Extraction and Labeling Process

According to the definition of PM presented in Section 3.1, it is necessary to form the set T of labeled tuples. For this, the sets of representations were constructed from offers within the e-commerce platforms. The extraction process starts from e-shop 1, extracting all available footwear items (excluding children’s footwear) from its online catalog. Due to the format of e-shop 1’s page and the information provided about the product, it is possible to extract five features: offer title, brand, category, price, and the image of the product (Figure 1). Once the entire e-shop 1 catalog is extracted, the title of each of the offers is used as a query to search for the product in e-shop 2 (Figure 2). From the search results, the first three matches are extracted with the same features as the query product from e-shop 1, as is shown in Figure 3. As a result, we have pairs of entities with two modalities of data: text (title, brand, category, price) and the product’s image.
Data extraction was performed with the programming languages R and P y t h o n due to the benefits of both languages for certain tasks. In the case of R, the web-scraping package called rvest was used, in addition to the tidyverse set of libraries. For P y t h o n , web-scraping was carried out with the Selenium library, which, although it is not a package designed for web-scraping, is useful for this purpose because it can help us to display menus that do not appear when parsing the html with other packages.
The set T has a label associated with each pair of offers. This label is already included in the databases designed for the PM problem; however, they are not available in this case, and the only way to obtain them is by manual labeling. This results in the set T, a database in which each element is a pair of representations accompanied by a label. These representations include structured and nonstructured data: name, brand, category, price, and image. Figure 4 shows a representation of the elements of the database. The set T contains 3489 pairs of products, from which 941 are positive (i.e., correspond to the same product) and 2548 are negative. Then, we have an unbalanced data set. To use this database in the labeling algorithms, a partition must be performed to create the training, validation, and test data sets. However, because the categories are unbalanced, the partitioning of the sets will have to preserve the same proportion of classes within each set. For this purpose, stratified sampling was performed with 60%, 20% and 20% for the training, validation and test sets, respectively.

3.3. A Multimodal Approach

Having defined the PM problem as a binary classification task, we aim to take advantage of two modalities of data (text and images) related to the pairs of entities, as described in Section 3.2.1.
Although there are many advantages to incorporating more channels of information into our data through their different modalities, more data are also a challenge. In classification tasks, two key aspects of multimodal data are representation (or encoding) and combination. We will describe each briefly.

3.3.1. Feature Extraction and Encoding

In this part, the goal is to obtain useful vector representations of the different modalities of the data prior to the fusion and classification process, capable of capturing the main features they possess. Arguably, this is the most challenging part, given the complex nonstructured nature of the data. Modern approaches are based on deep encoder or encoder–decoder neural networks pretrained on large data sets for transfer learning in specific tasks [45].

3.3.1.1. Image Modality

Particularly for images, CNNs ([46,47,48,49]) are popular models for feature extraction and representation. There are a wide variety of pretrained CNN architectures [50]. We used a CNN with the ResNet50 architecture [51], pretrained on the ImageNet classification data set [52], which has shown remarkable results in several competitions and has shown a good generalization performance. The input for the encoding process on this data modality consists of pairs of images of the products. The output will be an embedding for each offer that encodes the images’ information based on their similarity. Based on that, two encoding architectures are proposed:
  • 2-CNN architecture: The embeddings of both images are concatenated to form a single embedding. After this, it is connected to a series of regularized dense layers, whose output is a binary classifier that identifies if the images correspond to the same product. This architecture is shown in Figure 5.
  • image-Siamese: This architecture is based on a Siamese network similar to the one shown by Ralph Peeters [34]. In this case, the embeddings are created for each image from the connection between the pretrained CNN and the Siamese network consisting of regularized dense layers. Subsequently, a binary classifier is computed based on the Euclidean distance obtained from the final embeddings of the input images given by the Siamese network. The structure of the network is shown in Figure 6.
It was decided to use these two models because, on the one hand, CNNs are the most popular and efficient when it comes to image-related tasks. On the other hand, Siamese networks are used explicitly for image similarity recognition. Both architectures are very different; in the first one, a single embedding is obtained, and, based on its values, a label is assigned to the embedding. In contrast, the Siamese network creates new vector representations for each image, and then a label is assigned to indicate if they are the same products, based on the distance of their embeddings.

3.3.1.2. Text Modality

There are several options to obtain vector representations capable of modeling similarities between the textual descriptions of the products. A popular encoding approach is based on the occurrence of words in documents (in our case, textual descriptions), such as the vector space model [53] or TF-IDF ([54]). Here, the vector representation is obtained from a term-frequency matrix, where documents are vectors with lengths equal to the size of some predefined vocabulary. These representations are easy to obtain and have proven to be very efficient in some NLP-related tasks such as text classification [55], where even CML models perform better than sophisticated models or some ensemble of CML and DL models with TF-IDF [56]. Additionally, this encoding approach allows some interpretability in the sense that each entry of the encoding vector is related to a word in the vocabulary (row vector of the TF-IDF matrix) or a document in the corpus (column vector of the TF-IDF matrix), which makes it very useful for information retrieval tasks [57] or feature selection [56]. However, there are some drawbacks due to the high dimensionality and sparsity of the embedding vectors, and it is impossible to model relevant language characteristics, such as the sequential order of the words or semantic properties. Particularly for classification tasks based solely on text information, some studies suggest that the performance of the different word and document representations are highly dependent on the characteristics of the documents, that is, the number of tokens, size of the data set, and unbalance of categories [58,59], in that these characteristics are even more critical than the classifier itself. Still, for some DL-based classifiers, the data set size is crucial [56]. In recent years, word embeddings based on probabilistic language models have been building blocks of many SOTA models for various NLP tasks. They are dense vector representations of a predefined length and are capable of modeling complex properties of the semantics of words, such as polysemy, synonymy, and antonymy, among others [60]. Nowadays, word embeddings based on neural network language models are the preferred approach [31,61,62,63], particularly, contextualized word embeddings based on deep neural networks [64,65] have shown the best performance. Another advantage of word and document embeddings is the ability to perform transfer learning and fine tuning, which somewhat alleviates the problem of small data sets, as in our case.
Without a doubt, contextual word embeddings based on the transformer architecture [66] such as BERT [65], RoBERTa [38] or DistilBERT [39] are the SOTA for many NLP-related tasks. As they rely on self-attention mechanisms, these approaches allow us to obtain interpretable models, where we can find insight into which tokens are relevant for the given task. This is why we decided to use BERT embeddings as the encoding method for texts. Certainly, simpler document representations such as TF-IDF might give us the same or better results for the PM task than BERT embeddings if we only consider textual information. In our case, this is only one modality of our data, and our main objective is to exploit image and text representations. Therefore, we decided to use the best representation reported in the literature for most NLP tasks. However, a comparison of the different text representations (including other contextual embeddings, such as ELMo [64], DistilBERT, etc.) on a multimodal setting is very interesting, as is the combination of other DL classification algorithms different from dense layers, such as convolutional, LSTM-based or GRU-based neural networks. These options, suggested by a referee, will be considered as future work.
BERT (bidirectional encoder representations from transformers) is a language model pre-trained on Wikipedia’s BookCorpus [67] on two main tasks: masked language modeling (MLM) and next sentence prediction (NSP). We used the latter. In this task, the input consists of two text sequences with the following format: [CLS] Text 1 [SEP] Text 2 [SEP]; where [SEP] and [CLS] are special tokens denoting the start and separation of the text sequences, respectively. In our case, the text sequences are formed for each pair of offers based on their attributes: name, brand, category, and price, and the output is the encoding of the token [CLS] obtained with BETO [68] (https://huggingface.co/dccuchile/bert-base-spanish-wwm-uncased, accessed on 1 March 2021), the pretrained model of BERT in the Spanish language. The final embedding is obtained with a model inspired by the architecture designed by Mohamad Merchant for the semantic similarity task (https://keras.io/examples/nlp/semantic_similarity_with_bert/, accessed on 10 February 2021), which consists of a bidirectional LSTM layer followed by a hybrid pooling layer (max-pooling and average-pooling); then, both are concatenated and passed through regularized dense layers to the output layer, as shown in Figure 7. It should be noted that different structures were tested for these last dense layers to improve the model.

3.3.2. Fusion

After obtaining the embeddings for each modality of data, the second key aspect is the combination of the vector representations of the different modalities to have a global set of features for each item in our database. Sometimes, multimodal data can be simplified by translating the modalities in only one, such as in the case of audio transcription or image captioning [69,70,71]. Then, models based on CML, NLP, or DL can be used. However, a loss of valuable information can occur on that simplification. For classification tasks, as in our case, the popular approach is data fusion techniques [2]. Data or information fusion is defined in [72] as “the study of efficient methods for automatically or semi-automatically transforming information from different sources and different points in time into a representation that provides effective support for human or automated decision making.” Although initially two main strategies for fusion were considered, early and late fusion [73,74,75], for a wide range of ML tasks, a third one, intermediate fusion, was included [76]. Early or input-level fusion integrates the different modalities of raw or preprocessed data before any learning process is carried out. Late or decision-level fusion operates on each modality of the data, where a classifier is learned independently, and the final prediction is made by combining the outcomes of each one of them, which is very similar to the way ensemble methods work [77,78,79]. In the case of intermediate or feature-level fusion, there is a transformation of the raw inputs of each modality into a higher level representation [76], followed by a fusion of these representations, where a learning method is applied on the global feature map. Intermediate fusion is very flexible because it allows combining different features or representations, even from the same modality of data, which could enrich the information we obtain from the phenomenon of interest through the data sample. In this work, we used this approach for the fusion of the two modalities of data we have in our database. The architectures of the models we propose for PM based on multimodal data are listed below.
  • ImageBERT: This model consists of the fusion of text and image embeddings obtained with BERT and 2-CNN architectures as described in Section 3.3.1. Thus, a single embedding is obtained, which, in turn, is the input of a regularized neural network as a classifier. This architecture is shown in Figure 8.
  • BERTSiamese: This model consists of the fusion of the BERT-based embeddings for text and the image-Siamese embeddings for images, as was described in Section 3.3.1. Unlike the previous model, we concatenate the Euclidean distance of the resulting image embeddings from the Siamese network and the text embedding from the BERT-based architecture to feed a neural network classifier. Note that the text embedding is smaller than the individual version because of the addition of the distance; we do not want to lose its importance. The structure is shown in Figure 9.
Some relevant aspects should be mentioned. First, we introduce class weights to consider the imbalance in the categories of our data set, which implies the use of the weighted cross-entropy loss function. This function can be defined for a K categories classification problem, and for some training pair ( x i , y i ) , as L ( x i , y i )   =   w k log exp ( z i , y i ) k = 1 K exp ( z i , k ) y i , where the score z i , y i = x i β y i , for some parameters β , and w k is the weight for each class, which in our case is defined according to the class proportions. It is easy to show that minimizing this loss function over the entire training set is equivalent to obtaining the maximum likelihood estimator of the parameters of the model [80] but taking into account the imbalance of categories. For practical purposes, it is a way to tell the model to "pay more attention" to samples from an under-represented class (https://www.tensorflow.org/api_docs/python/tf/keras/Model, accessed on 15 March 2022).
Second, we choose the architectures for the encoding of multimodal data and the fusion models after an extensive set of experiments, consisting of a grid search over different sizes and numbers of dense layers. The optimal architecture was chosen based on the performance metrics described in Section 3.4.3.

3.4. Methodology

3.4.1. Data Preprocessing

We used some standard preprocessing for the data. We converted text to lowercase, removed special characters and stop words, and applied word tokenization. We resized images to 200   ×   200 pixels.

3.4.2. First Approach, CML Algorithms

Our first approach consisted of some classical classification algorithms based on textual information only. The data structure includes the following characteristics for each pair of products: name, category, brand, price, and image. The similarity between each pair of products will be calculated for each of these characteristics except the image, and they will be, in turn, the covariates of each labeled tuple. In the case of text, we used the Jaccard distance, defined by D J   =   1     J ( d o c 1 , d o c 2 ) , where J ( d o c 1 , d o c 2 ) is the Jaccard index, computed by J ( d o c 1 , d o c 2 )   =   | d o c 1 d o c 2 | | d o c 1 | + | d o c 2 | | d o c 1 d o c 2 | . The Jaccard index gives us the similarity between two documents d o c 1 and d o c 2 as the ratio of intersection over the union of the set of tokens of which they are composed. The difference between prices will be calculated and normalized, and the result will be subtracted from 1 so that all distances are in the same range. This results in four independent variables corresponding to the distance between each attribute. In all cases, the dependent variable is binary, indicating whether they are the same products or not. An example of the resulting vector of covariates is shown in Figure 10.
Our CML classification algorithms are based on logistic regression, SVMs, naive Bayes, k nearest neighbors, and random forest, using the default settings implemented in the Scikit-learn Python module [81].

3.4.3. Performance Metrics

We used standard metrics for classification tasks to assess the performance of the models we proposed.
  • Accuracy: Determines the proportion of correct predictions.
    Accuracy = TP + TN TP + TN + FP + FN
    where:
    -
    TP = true positives;
    -
    TN = true negatives;
    -
    FP = false positives;
    -
    FN = false negatives.
  • Precision: answers the question: what proportion of positive predictions are actually correct?
    Precision = TP TP + FP
  • Sensitivity (Recall): similar to Accuracy, Sensitivity answers the question: how many true positives were correctly classified?
    Sensitivity = TP TP + FN
  • F1 value: represents a weighted average between Accuracy and Sensitivity.
    F 1 = 2 ( Precision * Sensitivity ) ( Precision + Sensitivity ) .

4. Results

All models were implemented in the Python programming language using the Scikit-learn module for the CML algorithms and Keras-TensorFlow [82,83,84] for the DL models. All training and inference processes were performed on Google Colab (https://research.google.com/colaboratory/, accessed on 15 January 2021) with 12 GB RAM and using the GPU provided by this service. Training, validation, and test data sets were created for all models. For DL models, we used binary-weighted cross-entropy as the cost function (Section 3.3.2), the Adam optimizer, and all models were trained with 160 epochs. Additionally, we used early stopping with a patience parameter of 20 epochs to avoid overfitting. The bidirectional layer was chosen for the BERT model to take better advantage of the type of output of the pre-trained model. Hybrid pooling was selected to increase the number of elements that the classifier would receive as input. In general, the choice was made to keep the structures simple, so there is room for further improvements.
In Table 1, we show the results obtained for the CML models. Special attention will be paid to the F1 metric to evaluate the performance of all classification models, as it contains information on accuracy and sensitivity, measures directly related to the ability of the models to classify positive pairs correctly. It is worth mentioning that the training time for all models is very fast (less than a second) in all cases.
Table 1 shows that the CML model that achieved the best performance according to the F1 metric is random forest. Although the accuracy is above 0.7 in all models, meaning that more than 70% of the pairs were correctly classified, this metric should not be used alone due to the imbalance of the categories, as explained in Section 3.2.1. To understand why, consider that an algorithm that classified all pairs as negative would have an accuracy of 73%. However, it cannot distinguish one category from another. Therefore, knowing the accuracy performance of all the models is not a decisive point. Note also that there are no significant changes in the accuracy across the CML algorithms. In all cases, more than 50% of the predictions of positive pairs were correct. However, a drastic sensitivity change is observed among the methods. The lowest in sensitivity is logistic regression, with 38% of true positives classified correctly, while random forest scores 73%, a considerable difference. In addition to the F1 metric, this positions random forest as the best among the classical models.
For comparison purposes, we test the performance of the unimodal DL models proposed in Section 3.3.1. In this case, we used these models as classification algorithms, which is different from the multimodal approach described in Section 3.3.2, where we used only the embeddings of the unimodal representations of the data to combine them later with fusion techniques. The results are shown in Table 2.
A clear improvement is observed compared to the CML algorithms. Here, the accuracy is above 91%, the precision is 81%, and the sensitivity is 92%. In particular, the image-Siamese algorithm is the best in all metrics, achieving an accuracy of 91.1%, a precision of 81.8%, a sensitivity of 86.0%, and an F1 of 83.22%. An important observation in the unimodal models is that the algorithms behaved similarly despite using different types of data, that is, images and text, independently. In other words, a single data type, either images or text, is sufficient to build a model that can identify similarities between products. This, in part, responds to the doubt that arose as to whether, by having pairs of products with very similar textual descriptions, the algorithm fed only with texts would have problems identifying whether they were the same. Furthermore, this problem does not influence the performance of the models based on textual information because it behaves very similarly to the algorithms with only images as input and is even superior to one of them. However, this is not conclusive; more robust algorithms will have to be developed, and more tests must be carried out with each type of data to identify if there is indeed a problem.
Finally, we show the results for the multimodal approach we proposed in Table 3. It is worth analyzing each models’ behavior separately. Remember that the ImageBERT algorithm consists of the fusion of BERT and 2-CNN embeddings; we can observe an improvement in their performance, especially when we compare it to the unimodal 2-CNN architecture. Regarding their metrics, the accuracy is 92.41%, an improvement by 5.06% for BERT and 6.99% for 2-CNN; the precision is 86.26%, an improvement by 16% for BERT and 14.16% for 2-CNN; the sensitivity is 86.45%. These values suggest that the model could better identify positive pairs by sacrificing some accuracy. In contrast, for the BERTSiamese model, a loss in performance is observed compared to its unimodal algorithms BERT and image-Siamese in most cases. Although the image-Siamese model obtained the best performance within the unimodal models, its performance decreases for all metrics when we fuse it with BERT. This may be due to the final representation obtained from the Siamese model because it is only a distance value, which may not provide enough information from this modality of data compared to the textual representation given by the BERT embedding.
We can see that that the training time for all DL models was greater than one hour, whereas the training of the CML models was practically immediate. This is a disadvantage for DL-based models, however, once the models are trained, the inference procedure for new data is immediate. Furthermore, the specified epochs were not completed in any of the DL models due to the early-stopping procedure included.

5. Conclusions

In this article, a first approximation was made to understand the behavior of the Mexican e-commerce market, both in terms of buyers’ preferences and the structure in which sellers offer their products. We saw that the lack of resources and the heterogeneous structure of e-commerce platforms in Mexico make it particularly difficult to deal with the PM problem. For this reason, we developed a system capable of solving this specific problem based on modern deep-learning architectures and multimodal data consisting of text and image information. An extensive set of experiments were carried out to compare our proposal with different classical DL algorithms and even with modern DL models based on unimodal data.
Random forest presented the best performance of the CML algorithms because it was able to classify 73% of the positive pairs correctly.
During the unimodal models testing, and contrary to what was initially believed during the manual labeling of pairs, the performance of the algorithms trained only with text or with images was similar. This was observed despite the existence of certain pairs where the description is very similar. The previous leads us to conclude that it was equally feasible to create a model fed either by images only or by text only in terms of modeling.
The results indicate that multimodal models that combine text and images to construct a classification model can improve the performance of models based on unimodal data, as seen in the ImageBERT algorithm, which showed the best performance metrics.
Based on the above points, and despite the peculiarities and difficulties of the database, the objective of creating a system capable of identifying the similarity between products was achieved. This proposed system corresponds to the first system, as far as we know, that focused entirely on products with descriptions in Spanish and, specifically, for the Mexican e-commerce market. As a consequence of the lack of uniformity among the analyzed e-commerce sites, the input of the proposed system was multimodal. Furthermore, in addition to being the first of its kind in Spanish-speaking countries, the present multimodal product matching system is also the first in the scientific literature.

Future Work

This work was intended to be a starting point for further research in the field. For this reason, and based on the observations made throughout the process and the limitations of this study, the following is a description of specific issues that would be especially important to attack in future research.
  • In addition to the features used to describe the algorithms (title, category, price, brand, and image), a feature table with various product information was also extracted. This information was not used in the proposed system as its characteristics were not the same for all products. However, the BERT model uses text strings as input, so with the correct data preparation, it would be easy to use this feature table to make the product description more robust.
  • To improve the performance of the models, it is also crucial to have a sufficiently large database to train them. Therefore, increasing the amount of data that the current database contains is indispensable for future work.
  • Although the use of the BERT model was based on the results achieved in the literature, its performance in the PM problem studied here was not certain. For this reason, other types of algorithms could be used to see if they improve the results for this particular problem.
  • Although the problem of class imbalance present in the database generated in this study was treated here by means of various techniques, it is recommended to study other classification algorithms designed with this problem in mind, as well as to incorporate data-augmentation techniques to generate synthetic data of the minority class.
  • The precisions, accuracies and F1 values obtained are far from those achieved in the e-commerce PM literature for platforms written in English or Chinese. Future studies should focus on generating schemes that include more than two data modalities, as well as text preprocessing, other types of information fusion and classification-algorithm types such as multimodal extreme learning and shallow trees.

Author Contributions

Conceptualization, R.E.-V., V.M.-S. and H.D.-l.-T.-G.; methodology, R.E.-V., V.M.-S. and H.D.-l.-T.-G.; software, R.E.-V.; validation, R.E.-V., V.M.-S. and H.D.-l.-T.-G.; formal analysis, R.E.-V., V.M.-S. and H.D.-l.-T.-G.; investigation, R.E.-V., V.M.-S. and H.D.-l.-T.-G.; resources, R.E.-V.; data curation, R.E.-V.; writing—original draft preparation, R.E.-V. and H.D.-l.-T.-G.; writing—review and editing, R.E.-V., V.M.-S. and H.D.-l.-T.-G.; visualization, R.E.-V., V.M.-S. and H.D.-l.-T.-G.; supervision, V.M.-S. and H.D.-l.-T.-G.; project administration, R.E.-V., V.M.-S. and H.D.-l.-T.-G.; funding acquisition, V.M.-S. and H.D.-l.-T.-G. All authors have read and agreed to the published version of the manuscript.

Funding

The APC was funded by the Centro de Investigación en Matemáticas, A. C.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

R.E.V. acknowledges CONACyT for funding his postgraduate studies. H.T.G. would like to acknowledge the Catedras CONACyT fellowship program (project number 720) and Sistema Nacional de Investigadores (548421).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Market Study

For the market study, a database of 178,371 products was obtained at the end of the extraction period. Figure A1 shows the numbers of products, by category, obtained throughout the period. The category of Clothing, Bags, and Footwear has the most significant number of products (38,848). In other words, it is the most popular category, followed by Home, Furniture, and Garden with 20,739 products.
Within the Clothing, Bags and Footwear category, as shown in Figure A2, there is a clear trend for footwear, as it is by far the most popular subcategory. An analysis of the subcategories of Home, Furniture & Garden (Figure A3) shows the difference between them is not as clear; kitchen products are slightly more popular. In addition, as shown in Figure A4, among all the subcategories, footwear is the most popular, followed by Cell Phone Accessories, with a difference of 5246 products.
Figure A1. Number of products within each category.
Figure A1. Number of products within each category.
Mathematics 10 02564 g0a1
Figure A2. Number of products within the subcategories of Clothing, Bags, and Footwear.
Figure A2. Number of products within the subcategories of Clothing, Bags, and Footwear.
Mathematics 10 02564 g0a2
Figure A3. Number of products within the subcategories of Home, Furniture, and Garden.
Figure A3. Number of products within the subcategories of Home, Furniture, and Garden.
Mathematics 10 02564 g0a3
Figure A4. Number of products within the main subcategories.
Figure A4. Number of products within the main subcategories.
Mathematics 10 02564 g0a4
Analyzing the temporal behavior of the categories and subcategories mentioned above might also be worthwhile. First, for the case of the categories Clothing, Bags, and Footwear and Home, Furniture, and Garden, Figure A5 shows the number of daily products obtained in each category. Here, it can be seen how Clothing, Bags, and Shoes remains above the others for most of the time, except for the last few days where a small decrease is observed. For the sub-categories, we contrasted Shoes, for being the most popular, against Cell Phone Accessories (Figure A6) and Cooking (Figure A7). Both graphs show how footwear remains above the others over time and, equally, a decrease is observed towards the end.
Figure A5. Products per day by category.
Figure A5. Products per day by category.
Mathematics 10 02564 g0a5
Figure A6. Products per day; Footwear vs. Cell Phone Accessories.
Figure A6. Products per day; Footwear vs. Cell Phone Accessories.
Mathematics 10 02564 g0a6
Figure A7. Products per day; Footwear vs. Kitchen.
Figure A7. Products per day; Footwear vs. Kitchen.
Mathematics 10 02564 g0a7

Appendix A.1. Data Extraction and Manual Tagging

Two online stores were selected for the extraction process. A third e-shop was discarded after analyzing the web-page structure because there was no standard or structure for the information about most products. This was surprising because of what [33] mentions about the structure that e-commerce sites are adopting to present their products, especially because it is a multinational company whose roots are in the United States. Another critical point to highlight is that several of the products offered, when searching, had suppliers outside Mexico and presented descriptions in English.
Continuing with the proposed scheme, we will now address the problem of selecting e-commerce pages. In the case of e-shop 1, an analysis of the pages for characteristics that were repeated across the different products (to standardize the database) found that only the name, category, price, and brand appeared in all of them. However, the features table contained very variable information across all products, even within the same category. In addition, the description section included much information that was not relevant to our study (e.g., a phrase like “Gain a point in your favor to combine your formal attire with the elegance you deserve...’’) or contained errors (e.g., Style: $$ Dressing $$ Color: Black $$ Sole: Synthetic Anti-skid...). Fortunately, e-shop 2 presented a similar enough structure to be able to extract the same characteristics of the products.
Continuing with the manual labeling process, it was observed that the name and brand were practically identical for some pairs of products, and their price and category were not enough to differentiate them. Hence, the only way to tell them apart was through the image of each offer. An example of this case is shown in Figure A8. It was also observed that some of the pairs formed contained products that did not even belong to the same category (e.g., tennis ball and shoes). Although they could have been kept in the base, it was decided to eliminate them. We obtained a base with 3489 pairs, of which 943 are positive and the rest negative.
Figure A8. Different products with almost identical names.
Figure A8. Different products with almost identical names.
Mathematics 10 02564 g0a8

Appendix A.2. Discussion

The choice of products analyzed and matched between the two e-commerce stores was made by market research over a long period. Detailed product information was obtained from e-shop 2’s trends page. The fact that Clothing, Bags, and Footwear and, specifically, footwear, is the category with the highest number of products in the extraction period means that the most searched products daily were those belonging to that category. In other words, they were the most popular products day by day. This idea is reinforced by observing how, in Figure A6 and Figure A7, the footwear category remains positioned with products on the trending page throughout every day. Most of the time, this category had more products than any other category.
It is also important to note that the popularity of footwear products declines towards the end of the extraction days. This may be due to a seasonal phenomenon because these are festive dates where it is logical that other types of products have more prominence. Despite this, it is still a tiny fraction of the time when this decrease in popularity originates, compared to the number of days footwear remains the most popular category. Given the above, there is no doubt that the footwear category plays an essential role in the e-commerce market, and although the analysis was only carried out in e-shop 2, the presence and popularity of this platform means that it can significantly represent the consumption habits of the population that shops online.
As mentioned before, we chose e-shop 1 and e-shop 2 because their web pages are more homogeneous and offer easy access to product features. Those problems are common because some e-shops provide a platform for other sellers to offer their products. This solves the initial question of whether the e-commerce platforms operating in Mexico would employ a general structure the same way platforms in other countries do. This indicates that, for e-shop 1 and e-shop 2, the organization of their products is sufficiently clear to find all the characteristics of their products. Even across platforms, the products presented the same features and a similar organization.
Another aspect to highlight in the information-extraction process is related to the triplet formed by e-shop 1, e-shop 2 and Shoes, which was ideal for the development of this research in PM. On the one hand, e-shop 1 and e-shop 2 present products with well-structured characteristics and, on the other hand, the footwear products, in addition to the undoubted popularity that was observed, are similar enough to each other so that they present the same type of characteristics, but without being very complicated to distinguish one from the other. To understand this, if you look at the products for the cell phone accessories category, which had popularity close to that of footwear, you will see that there is a lot of difference between them: cell phone cases, chargers, protective covers, etc. However, if one were to choose only cell phones, two phones that were mostly identical except for the amount of storage would have to be treated as a negative pair.
As for the database in general, the number of pairs it contains is relatively tiny compared to those used in the literature. However, using images for the analysis makes the data more varied and contain more information. Even so, it was initially expected that the performance of the models would not be optimal due to the amount of data.

References

  1. Guerra López, R.A. Diseño de un Sistema de Catalogación de Productos de E-Commerce Utilizando Procesamiento de Lenguaje Natural (NLP) y Machine learning. Master’s Thesis, Universidad de Chile, Santiago, Chile, 2019. [Google Scholar]
  2. Lahat, D.; Adali, T.; Jutten, C. Multimodal Data Fusion: An Overview of Methods, Challenges, and Prospects. Proc. IEEE 2015, 103, 1449–1477. [Google Scholar] [CrossRef] [Green Version]
  3. Ngai, E.; Wat, F. A literature review and classification of electronic commerce research. Inf. Manag. 2002, 39, 415–429. [Google Scholar] [CrossRef]
  4. Sans, O.; Agnew, G.B. Electronic Commerce Technologies, 1st ed.; Kou, W., Yesha, Y., Tan, C.J., Eds.; Lecture Notes in Computer Science No.2040 2001-apr 10; Springer: Berlin/Heidelberg, Germany, 2001. [Google Scholar]
  5. Graef, G.; Gaedke, M. Electronic Commerce and Web Technologies: First International Conference, EC-Web 2000 London, UK, September 4–6, 2000 Proceedings, 1st ed.; Bauknecht, K., Madria, S.K., Pernul, G., Eds.; Lecture Notes in Computer Science No. 1875; Springer: Berlin/Heidelberg, Germany, 2000. [Google Scholar]
  6. Yokoo, M.; Ito, T.; Zhang, M.; Matsuo, T.; Lee, J. Electronic Commerce: Theory and Practice, 1st ed.; Studies in Computational Intelligence No. 110; Springer: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
  7. Bawack, R.E.; Wamba, S.F.; Carillo, K.D.A.; Akter, S. Artificial intelligence in E-Commerce: A bibliometric study and literature review. Electron. Mark. 2022, 32, 297–338. [Google Scholar] [CrossRef] [PubMed]
  8. Babbar, R.; Schölkopf, B. DiSMEC: Distributed Sparse Machines for Extreme Multi-Label Classification. In Proceedings of the Tenth ACM International Conference on Web Search and Data Mining (WSDM ’17), Cambridge, UK, 6–10 February 2017; Association for Computing Machinery: New York, NY, USA, 2017; pp. 721–729. [Google Scholar] [CrossRef]
  9. Khandagale, S.; Xiao, H.; Babbar, R. Bonsai: Diverse and shallow trees for extreme multi-label classification. Mach. Learn. 2020, 109, 2099–2119. [Google Scholar] [CrossRef]
  10. Dahiya, K.; Agarwal, A.; Saini, D.; Gururaj, K.; Jiao, J.; Singh, A.; Agarwal, S.; Kar, P.; Varma, M. SiameseXML: Siamese Networks meet Extreme Classifiers with 100M Labels. In Proceedings of the International Conference on Machine Learning, Vienna, Austria, 18–24 July 2021. [Google Scholar]
  11. Mittal, A.; Dahiya, K.; Malani, S.; Ramaswamy, J.; Kuruvilla, S.; Ajmera, J.; Chang, K.; Agrawal, S.; Kar, P.; Varma, M. Multimodal extreme classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LO, USA, 19–24 June 2022. [Google Scholar]
  12. Wang, J.; Li, G.; Yu, J.X.; Feng, J. Entity matching: How similar is similar. Proc. VLDB Endow. 2011, 4, 622–633. [Google Scholar] [CrossRef] [Green Version]
  13. Kejriwal, M.; Shen, K.; Ni, C.C.; Torzec, N. An evaluation and annotation methodology for product category matching in e-commerce. Comput. Ind. 2021, 131, 103497. [Google Scholar] [CrossRef]
  14. Singla, P.; Domingos, P. Entity resolution with markov logic. In Proceedings of the Sixth International Conference on Data Mining (ICDM’06), Hong Kong, China, 18–22 December 2006; pp. 572–582. [Google Scholar]
  15. Kusniyati, H.; Nugraha, A.A. Analysis of Matric Product Matching between Cosine Similarity with Term Frequency-Inverse Document Frequency (TF-IDF) and Word2Vec in PT. Pricebook Digital Indonesia. Int. J. Sci. Res. Comput. Sci. Eng. Inf. Technol. 2020, 6, 105–112. [Google Scholar] [CrossRef]
  16. Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient estimation of word representations in vector space. arXiv 2013, arXiv:1301.3781. [Google Scholar]
  17. Köpcke, H.; Thor, A.; Rahm, E. Evaluation of entity resolution approaches on real-world match problems. Proc. VLDB Endow. 2010, 3, 484–493. [Google Scholar] [CrossRef] [Green Version]
  18. Xiao, C.; Wang, W.; Lin, X.; Yu, J.X.; Wang, G. Efficient similarity joins for near-duplicate detection. ACM Trans. Database Syst. (TODS) 2011, 36, 1–41. [Google Scholar] [CrossRef]
  19. Fellegi, I.P.; Sunter, A.B. A theory for record linkage. J. Am. Stat. Assoc. 1969, 64, 1183–1210. [Google Scholar] [CrossRef]
  20. Christen, P. Febrl: A freely available record linkage system with a graphical user interface. In Proceedings of the Second Australasian Workshop on Health Data and Knowledge Management, Wollongong, Australia, January 2008; Volume 80, pp. 14–25. [Google Scholar]
  21. Bilenko, M.; Mooney, R.J. Adaptive duplicate detection using learnable string similarity measures. In Proceedings of the Ninth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, 24–27 August 2003; pp. 39–48. [Google Scholar]
  22. Li, J.; Dou, Z.; Zhu, Y.; Zuo, X.; Wen, J.R. Deep cross-platform product matching in e-commerce. Inf. Retr. J. 2020, 23, 136–158. [Google Scholar] [CrossRef]
  23. Konda, P.; Das, S.; Suganthan GC, P.; Doan, A.; Ardalan, A.; Ballard, J.R.; Li, H.; Panahi, F.; Zhang, H.; Naughton, J.; et al. Magellan: Toward building entity matching management systems. Proc. VLDB Endow. 2016, 9, 1197–1208. [Google Scholar] [CrossRef]
  24. Christophides, V.; Efthymiou, V.; Palpanas, T.; Papadakis, G.; Stefanidis, K. End-to-end entity resolution for big data: A survey. arXiv 2019, arXiv:1905.06397. [Google Scholar]
  25. Ebraheem, M.; Thirumuruganathan, S.; Joty, S.; Ouzzani, M.; Tang, N. Distributed representations of tuples for entity resolution. Proc. VLDB Endow. 2018, 11, 1454–1467. [Google Scholar] [CrossRef] [Green Version]
  26. Yang, C.; Flak, O.; Grzegorzek, M. Representation and matching of team managers: An experimental research. IEEE Trans. Comput. Soc. Syst. 2018, 5, 311–323. [Google Scholar] [CrossRef]
  27. Mudgal, S.; Li, H.; Rekatsinas, T.; Doan, A.; Park, Y.; Krishnan, G.; Deep, R.; Arcaute, E.; Raghavendra, V. Deep learning for entity matching: A design space exploration. In Proceedings of the 2018 International Conference on Management of Data, Houston, TX, USA, 10–15 June 2018; pp. 19–34. [Google Scholar]
  28. Ristoski, P.; Petrovski, P.; Mika, P.; Paulheim, H. A machine learning approach for product matching and categorization. Semant. Web 2018, 9, 707–728. [Google Scholar] [CrossRef] [Green Version]
  29. Kannan, A.; Givoni, I.E.; Agrawal, R.; Fuxman, A. Matching unstructured product offers to structured product specifications. In Proceedings of the 17th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Diego, CA, USA, 21–24 August 2011; pp. 404–412. [Google Scholar]
  30. Finkel, J.R.; Grenager, T.; Manning, C.D. Incorporating non-local information into information extraction systems by gibbs sampling. In Proceedings of the 43rd Annual Meeting of the Association for Computational Linguistics (ACL’05), Ann Arbor, MI, USA, 25–30 June 2005; pp. 363–370. [Google Scholar]
  31. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.; Dean, J. Distributed representations of words and phrases and their compositionality. arXiv 2013, arXiv:1310.4546. [Google Scholar]
  32. Wang, M.; Manning, C.D. Effect of non-linear deep architecture in sequence labeling. In Proceedings of the Sixth International Joint Conference on Natural Language Processing, Nagoya, Japan, 14–19 October 2013; pp. 1285–1291. [Google Scholar]
  33. Peeters, R.; Primpeli, A.; Wichtlhuber, B.; Bizer, C. Using schema. org annotations for training and maintaining product matchers. In Proceedings of the 10th International Conference on Web Intelligence, Mining and Semantics, Biarritz, France, 30 June 2020; pp. 195–204. [Google Scholar]
  34. Shah, K.; Kopru, S.; Ruvini, J.D. Neural network based extreme classification and similarity models for product matching. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 3 (Industry Papers); Association for Computational Linguistics: Dublin, Ireland, 2018; pp. 8–15. [Google Scholar]
  35. Li, Y.; Li, J.; Suhara, Y.; Doan, A.; Tan, W.C. Deep entity matching with pre-trained language models. arXiv 2020, arXiv:2004.00584. [Google Scholar] [CrossRef]
  36. Peeters, R.; Bizer, C.; Glavaš, G. Intermediate training of BERT for product matching. Small 2020, 745, 2–112. [Google Scholar]
  37. Borst, J.; Koerner, E.; Opasjumruskit, K.; Niekler, A. Language Model CNN-driven similarity matching and classification for HTML-embedded Product Data. In Proceedings of the ISWC 2020, virtual, 12–17 September 2020. [Google Scholar]
  38. Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; Stoyanov, V. Roberta: A robustly optimized bert pretraining approach. arXiv 2019, arXiv:1907.11692. [Google Scholar]
  39. Sanh, V.; Debut, L.; Chaumond, J.; Wolf, T. DistilBERT, a distilled version of BERT: Smaller, faster, cheaper and lighter. arXiv 2019, arXiv:1910.01108. [Google Scholar]
  40. Zhang, Y.; Wallace, B. A sensitivity analysis of (and practitioners’ guide to) convolutional neural networks for sentence classification. arXiv 2015, arXiv:1510.03820. [Google Scholar]
  41. Fu, C.; Wang, T.; Nie, H.; Han, X. 0001, L.S. ISCAS_ICIP at MWPD-2020 Task 1: Product Matching Based on Deep Entity Matching Frameworks. In Proceedings of the ISWC 2020, virtual, 12–17 September 2020. MWPD@ISWC. [Google Scholar]
  42. Fu, C.; Han, X.; Sun, L.; Chen, B.; Zhang, W.; Wu, S.; Kong, H. End-to-End Multi-Perspective Matching for Entity Resolution. In Proceedings of the IJCAI, Macao, China, 10–16 August 2019; pp. 4961–4967. [Google Scholar]
  43. Nie, H.; Han, X.; He, B.; Sun, L.; Chen, B.; Zhang, W.; Wu, S.; Kong, H. Deep sequence-to-sequence entity matching for heterogeneous entity resolution. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, Beijing, China, 3–7 November 2019; pp. 629–638. [Google Scholar]
  44. Fu, C.; Han, X.; He, J.; Sun, L. Hierarchical matching network for heterogeneous entity resolution. In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, Yokohama, Japan, 7–15 January 2021; pp. 3665–3671. [Google Scholar]
  45. Torrey, L.; Shavlik, J. Transfer learning. In Handbook of Research on Machine Learning Applications and Trends: Algorithms, Methods, and Techniques; IGI Global: Hershey, PA, USA, 2010; pp. 242–264. [Google Scholar]
  46. LeCun, Y. Generalization and Network Design Strategies. In Connectionism in Perspective; Pfeifer, R., Schreter, Z., Fogelman, F., Steels, L., Eds.; An Extended Version Was Published as a Technical Report of the University of Toronto; Elsevier: Zurich, Switzerland, 1989. [Google Scholar]
  47. LeCun, Y.; Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jackel, L.D. Backpropagation applied to handwritten zip code recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
  48. LeCun, Y.; Bengio, Y. Convolutional networks for images, speech, and time series. Handb. Brain Theory Neural Netw. 1995, 3361, 1995. [Google Scholar]
  49. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  50. Khan, A.; Sohail, A.; Zahoora, U.; Qureshi, A.S. A survey of the recent architectures of deep convolutional neural networks. Artif. Intell. Rev. 2020, 53, 5455–5516. [Google Scholar] [CrossRef] [Green Version]
  51. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  52. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. (IJCV) 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  53. Salton, G.; Wong, A.; Yang, C.S. A Vector Space Model for Automatic Indexing. Commun. ACM 1975, 18, 613–620. [Google Scholar] [CrossRef] [Green Version]
  54. Salton, G.; Buckley, C. Term-Weighting Approaches in Automatic Text Retrieval. Inf. Process. Manag. 1988, 24, 513–523. [Google Scholar] [CrossRef] [Green Version]
  55. Dessi, D.; Helaoui, R.; Kumar, V.; Recupero, D.R.; Riboni, D. TF-IDF vs word embeddings for morbidity identification in clinical notes: An initial study. In Proceedings of the First Workshop on Smart Personal Health Interfaces Co-Located with 25th International Conference on Intelligent User Interfaces, SmartPhil@IUI 2020, Cagliari, Italy, 17 March 2020; Consoli, S., Recupero, D.R., Riboni, D., Eds.; CEUR-WS.org: Aachen, Germany, 2020; Volume 2596, pp. 1–12. [Google Scholar]
  56. Kumar, V.; Recupero, D.R.; Riboni, D.; Helaoui, R. Ensembling classical machine learning and deep learning approaches for morbidity identification from clinical notes. IEEE Access 2020, 9, 7107–7126. [Google Scholar] [CrossRef]
  57. Manning, C.D.; Raghavan, P.; Schütze, H. Introduction to Information Retrieval; Cambridge University Press: Cambridge, UK, 2008. [Google Scholar]
  58. Piskorski, J.; Jacquet, G. TF-IDF Character N-grams versus Word Embedding-based Models for Fine-grained Event Classification: A Preliminary Study. In Proceedings of the Workshop on Automated Extraction of Socio-Political Events from News 2020, Marseille, France, 11–16 May 2020; European Language Resources Association (ELRA): Marseille, France, 2020; pp. 26–34. [Google Scholar]
  59. Walkowiak, T. Subject Classification of Texts in Polish—From TF-IDF to Transformers. In Theory and Engineering of Dependable Computer Systems and Networks; Zamojski, W., Mazurkiewicz, J., Sugier, J., Walkowiak, T., Kacprzyk, J., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 457–465. [Google Scholar]
  60. Senel, L.K.; Utlu, I.; Yucesoy, V.; Koc, A.; Cukur, T. Semantic Structure and Interpretability of Word Embeddings. IEEE/ACM Trans. Audio Speech Lang. Proc. 2018, 26, 1769–1779. [Google Scholar] [CrossRef] [Green Version]
  61. Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient Estimation of Word Representations in Vector Space. In Workshop Track Proceedings, Proceedings of the 1st International Conference on Learning Representations (ICLR 2013), Scottsdale, AZ, USA, 2–4 May 2013; Bengio, Y., LeCun, Y., Eds. arXiv 2013, arXiv:1301.3781. [Google Scholar]
  62. Pennington, J.; Socher, R.; Manning, C.D. GloVe: Global Vectors for Word Representation. In Proceedings of the Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1532–1543. [Google Scholar]
  63. Bojanowski, P.; Grave, E.; Joulin, A.; Mikolov, T. Enriching Word Vectors with Subword Information. Trans. Assoc. Comput. Linguist. 2017, 5, 135–146. [Google Scholar] [CrossRef] [Green Version]
  64. Peters, M.E.; Neumann, M.; Iyyer, M.; Gardner, M.; Clark, C.; Lee, K.; Zettlemoyer, L. Deep Contextualized Word Representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers); Association for Computational Linguistics: New Orleans, LA, USA, 2018; pp. 2227–2237. [Google Scholar]
  65. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  66. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
  67. Zhu, Y.; Kiros, R.; Zemel, R.S.; Salakhutdinov, R.; Urtasun, R.; Torralba, A.; Fidler, S. Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books. In Proceedings of the ICCV 2015, Santiago, Chile, 7–13 December 2015; IEEE Computer Society: Manhattan, NY, USA, 2015; pp. 19–27. [Google Scholar]
  68. Cañete, J.; Chaperon, G.; Fuentes, R.; Ho, J.H.; Kang, H.; Pérez, J. Spanish Pre-Trained BERT Model and Evaluation Data. PML4DC ICLR 2020 2020, 2020, 1–10. [Google Scholar]
  69. Mao, J.; Xu, W.; Yang, Y.; Wang, J.; Huang, Z.; Yuille, A. Deep captioning with multimodal recurrent neural networks (m-rnn). arXiv 2014, arXiv:1412.6632. [Google Scholar]
  70. Kiros, R.; Salakhutdinov, R.; Zemel, R. Multimodal neural language models. International conference on machine learning. PMLR 2014, 32, 595. [Google Scholar]
  71. Xu, K.; Ba, J.; Kiros, R.; Cho, K.; Courville, A.; Salakhudinov, R.; Zemel, R.; Bengio, Y. Show, attend and tell: Neural image caption generation with visual attention. In Proceedings of the International Conference on Machine Learning (PMLR 2015), Lille, France, 7–9 July 2015; pp. 2048–2057. [Google Scholar]
  72. Khaleghi, B.; Khamis, A.M.; Karray, F.; Razavi, S.N. Multisensor data fusion: A review of the state-of-the-art. Inf. Fusion 2013, 14, 28–44. [Google Scholar] [CrossRef]
  73. Atrey, P.K.; Hossain, M.A.; Saddik, A.E.; Kankanhalli, M. Multimodal fusion for multimedia analysis: A survey. Multimed. Syst. 2010, 16, 345–379. [Google Scholar] [CrossRef]
  74. Yao, W.; Moumtzidou, A.; Dumitru, C.O.; Andreadis, S.; Gialampoukidis, I.; Vrochidis, S.; Datcu, M.; Kompatsiaris, Y. Early and Late Fusion of Multiple Modalities in Sentinel Imagery and Social Media Retrieval. In Proceedings of the ICPR Workshops 2020, virtual, 10–15 January 2021. [Google Scholar]
  75. Gadzicki, K.; Khamsehashari, R.; Zetzsche, C. Early vs Late Fusion in Multimodal Convolutional Neural Networks. In Proceedings of the 2020 IEEE 23rd International Conference on Information Fusion (FUSION), Rustenburg, South Africa, 6–9 July 2020; pp. 1–6. [Google Scholar] [CrossRef]
  76. Boulahia, S.Y.; Amamra, A.; Madi, M.R.; Daikh, S. Early, intermediate and late fusion strategies for robust deep learning-based multimodal action recognition. Mach. Vis. Appl. 2021, 32, 121. [Google Scholar] [CrossRef]
  77. Breiman, L. Bagging Predictors. Mach. Learn. 2005, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  78. Polikar, R. Ensemble based systems in decision making. IEEE Circuits Syst. Mag. 2006, 6, 21–45. [Google Scholar] [CrossRef]
  79. Rokach, L. Ensemble-based classifiers. Artif. Intell. Rev. 2009, 33, 1–39. [Google Scholar] [CrossRef]
  80. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  81. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  82. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems, 2015. Available online: tensorflow.org (accessed on 5 January 2021).
  83. Abadi, M.; Barham, P.; Chen, J.; Chen, Z.; Davis, A.; Dean, J.; Devin, M.; Ghemawat, S.; Irving, G.; Isard, M.; et al. TensorFlow: A System for Large-Scale Machine Learning. In Proceedings of the 12th USENIX Conference on Operating Systems Design and Implementation (OSDI’16), Savannah, GA, USA, 2–4 November 2016; USENIX Association: Berkeley, CAL, USA, 2016; pp. 265–283. [Google Scholar]
  84. Chollet, F. Keras. 2015. Available online: https://keras.io (accessed on 1 October 2021).
Figure 1. Example of e-shop 1 product. Title, brand, category, price and image.
Figure 1. Example of e-shop 1 product. Title, brand, category, price and image.
Mathematics 10 02564 g001
Figure 2. Example of e-shop 2 product. Title, brand, category, price and image.
Figure 2. Example of e-shop 2 product. Title, brand, category, price and image.
Mathematics 10 02564 g002
Figure 3. Information extraction process.
Figure 3. Information extraction process.
Mathematics 10 02564 g003
Figure 4. Database structure.
Figure 4. Database structure.
Mathematics 10 02564 g004
Figure 5. 2-CNN architecture for images of products.
Figure 5. 2-CNN architecture for images of products.
Mathematics 10 02564 g005
Figure 6. Image-Siamese architecture for images of products.
Figure 6. Image-Siamese architecture for images of products.
Mathematics 10 02564 g006
Figure 7. BERT-based architecture for textual descriptions of products.
Figure 7. BERT-based architecture for textual descriptions of products.
Mathematics 10 02564 g007
Figure 8. ImageBERT architecture.
Figure 8. ImageBERT architecture.
Mathematics 10 02564 g008
Figure 9. BERTSiamese architecture.
Figure 9. BERTSiamese architecture.
Mathematics 10 02564 g009
Figure 10. Vector of covariates for the labeled tuples used in our CML algorithm.
Figure 10. Vector of covariates for the labeled tuples used in our CML algorithm.
Mathematics 10 02564 g010
Table 1. Results of the classical ML models.
Table 1. Results of the classical ML models.
ModelAccuracyPrecisionSensitivityF1Training Time (s)
Logistic regression0.7720.6280.3780.4720.0250
SVM0.7780.6740.3400.4520.1408
Naive Bayes0.7340.5040.7290.5960.0073
KNN0.7940.6360.5480.5890.0261
Random forest0.8380.6440.7250.6820.2984
Table 2. Result of unimodal DL models.
Table 2. Result of unimodal DL models.
ModelAccuracyPrecisionSensitivityF1Training Time (s)
2-CNN0.85420.72100.80410.74926200.72
Image-Siamese0.91070.81810.86000.83224411.63
BERT0.87350.70260.92140.7887618.77
Table 3. Results of multimodal DL models.
Table 3. Results of multimodal DL models.
ModelAccuracyPrecisionSensitivityF1Training Time (s)
ImageBERT0.92410.86260.86450.85844597.17
BERTSiamese0.87200.70340.89800.783515,129.38
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Estrada-Valenciano, R.; Muñiz-Sánchez, V.; De-la-Torre-Gutiérrez, H. An Entity-Matching System Based on Multimodal Data for Two Major E-Commerce Stores in Mexico. Mathematics 2022, 10, 2564. https://doi.org/10.3390/math10152564

AMA Style

Estrada-Valenciano R, Muñiz-Sánchez V, De-la-Torre-Gutiérrez H. An Entity-Matching System Based on Multimodal Data for Two Major E-Commerce Stores in Mexico. Mathematics. 2022; 10(15):2564. https://doi.org/10.3390/math10152564

Chicago/Turabian Style

Estrada-Valenciano, Raúl, Víctor Muñiz-Sánchez, and Héctor De-la-Torre-Gutiérrez. 2022. "An Entity-Matching System Based on Multimodal Data for Two Major E-Commerce Stores in Mexico" Mathematics 10, no. 15: 2564. https://doi.org/10.3390/math10152564

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop