You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

8 February 2022

Trademark Image Similarity Detection Using Convolutional Neural Network

,
and
1
College of Computer and Information Sciences, King Saud University, Riyadh 11543, Saudi Arabia
2
College of Computer and Information Sciences, Imam Muhammad bin Saud University, Riyadh 11432, Saudi Arabia
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Applications of Deep Learning and Artificial Intelligence Methods

Abstract

A trademark is any recognizable sign that identifies products/services and distinguishes them from others. Many regional and international intellectual property offices are dedicated to dealing with trademark registration processes. The registration process involves examining the trademark to ensure there is no confusion or interference similarity to any other prior registered trademark. Due to the increasing number of registered trademarks annually, the current manual examining approach is becoming insufficient and more susceptible to human error. As such, there is potential for machine learning applications and deep learning, in particular, to enhance the examination process by providing an automated image detection system to be used by the examiners to facilitate and improve the accuracy of the examination process. Therefore, this paper proposed a trademark similarity detection system using deep-learning techniques to extract image features automatically in order to retrieve a trademark based on shape similarity. Two pretrained convolutional neural networks (VGGNet and ResNet) were individually used to extract image features. Then, their performance was compared. Subsequently, the extracted features were used to calculate the similarity between a new trademark and each of those registered using the Euclidean distance. Thereafter, the system retrieved the most similar trademark to the query according to the smallest distances. As a result, the system achieved an average rank of 67,067.788, a normalized average rank of 0.0725, and a mean average precision of 0.774 on the Middle East Technical University dataset, which displays a promising application in detecting trademark similarity.

1. Introduction

Object detection is an important field in artificial intelligence (AI). Today, there are many computer vision applications based on object detection in different contexts. Accordingly, a wide variety of detection models have been proposed in the literature [1,2]. Similarity detection and measures is an emerging field of research in object detection.
A trademark is a form of intellectual property (IP) consisting of a recognizable sign that identifies and distinguishes a product or service from others. It can be text, a phrase, a figurative symbol, a design, or any combination of these. The importance of logos and trademarks in building a brand is unquestionable, even for small businesses. A trademark mirrors the quality and professionalism of the product or service provided. Most customers rely on the trademark’s reputation to make purchase decisions and compare the quality of two products [3,4]. Therefore, the added value and additional profit that trademarks endow to their products exceed their real functional benefits [4]. In 2018, the number of recorded trademark registrations worldwide was approximately equal to 10.9 million, an increase of 19.2%, compared with 2017 [5].
Some processes should be conducted to ensure that the new mark does not cause confusion with prior registered trademarks and does not match any others in the case of a new trademark [6]. In visual terms, the meaning of “matching” includes a combination of different parameters, such as colors, shape, and texture.
Nowadays, a number of regional and national IP offices around the world deal with the processes of trademark examination and registration [6,7]. Examples include the European Union Intellectual Property Office and the United States Patent and Trademark Office (USPTO), which serve these functions. In Saudi Arabia, the Saudi Authority for Intellectual Property is the official government body responsible for protecting and supporting IP rights in the Kingdom [7].
Any new trademark should be examined to ensure no confusion or close similarity to any other prior registered marks. In a manual approach, the examination process for any trademark requires a long time as a result of the increasing annual number of registered trademarks. Additionally, the process could be susceptible to error because judging whether two trademarks are similar is subjective and varies from one examiner to another depending on their experience.
The automation of the examination process using AI that facilitates the retrieval of similarity-based trademarks can provide significant help for examiners to achieve higher accuracy. The AI can provide a limited number of most similar trademarks that need to be examined rather than a large database of marks. Although there are already some systems intended to support examiners in the trademark examination process, these systems depend on manually annotated trademarks. Therefore, they are subjective, time-consuming, and error-prone. Consequently, this study develops a trademark similarity detection system using deep learning that can automatically and accurately detect and retrieve registered trademarks that are closest in similarity to an input mark.
The contribution of this study is to provide a framework and comprehensive background for trademark similarity experiments as an examination tool in the trademark registration process. The proposed framework presents a system that reads an image (including the various image formats), extracts its features, and finds the similarity between a given image and each of those stored based on their distance. We investigated the use of convolutional neural network (CNN) to extract image features. Two pre-trained CNNs were utilized individually as features extractors: ResNet-50 and VGG-16. All of the experiments were conducted on a portion of the METU test set, equal to 10,000 images, to accelerate the process, along with the injected query set, to evaluate the retrieval performance.
The following sections illustrate our proposed system framework for the trademark similarity retrieval system. Section 2 presents related work on this topic. Section 3 describes our proposed system. Section 4 presents experiments and results. Section 5 presents our discussion. Finally, Section 6 presents the study’s conclusions.

3. Materials and Methods

Deep learning allows a wide range of intelligent tasks to be automatically and efficiently performed. It has allowed for significant advancements in solving many difficult problems with performance that could exceed human capabilities. The proposed trademark similarity detection system will help the examiners in the IP offices and individual users guarantee that a given trademark does not match any prior registered trademarks by harnessing the potential of deep learning to automate extracting the image’s features. Two pretrained CNNs are individually utilized: VGG-16 and ResNet-50. The distance between the query image and each previously registered trademark is measured using the Euclidean distance metric. Finally, the system retrieves the images with the smallest distance, meaning they are the most similar to the given query.
The system architecture consists of four main modules: preprocessing, feature extraction, feature reduction, and similarity measurement. Figure 1 presents the system’s architecture.
Figure 1. System architecture.
The system was developed using Python version 2.7.12, with the help of Google Colaboratory from Google Research, a hosted Jupyter Notebook service that provides access to graphics processing units. Moreover, the dataset was uploaded into Google Drive to manage storage.
The following subsection briefly describes our proposed system. We start by describing the benchmark dataset, followed by the utilized techniques of the essential system components. Finally, it presents the metrics used to evaluate the system performance.

3.1. Dataset

Trademark similarity detection is an image retrieval operation used to guarantee that a new trademark does not match any previously registered marks. The METU dataset is considered one of the best datasets due to its large scale, consisting of more than 923,000 trademark images containing text only, figures only, both figure and text, or unknown content [21]. All the images are in JPEG format at various sizes and resolutions, from 30 × 30 pixels to 180 × 180 pixels. It has two main sets: the query set and test set. The query set is an independent labeled set containing 417 trademark images that belong to 35 different classes based on shape similarity. The primary role of the query set is to evaluate such a system. It could be injected into the entire or part of the test set, and then the image retrieval approach can be applied for the whole dataset and the rank of the relevant images that belong to the same query class can be looked at. Figure 2 shows a sample of the METU dataset.
Figure 2. Middle East Technical University data samples that contain figures only, text only, and both.

3.2. Preprocessing

The nature of trademark images is artificial and as such they require little or no processing. Moreover, any pretrained CNNs require a specific input dimension. Due to the dataset images being a variety of sizes, all images were resized to the accepted dimension by a given CNN, which is in our case equal to 224 × 224. Therefore, two approaches were conducted: scaling and padding to choose the best approach that does not lose the image’s information. The image scaling changes the input image’s height h and width w to the target dimension n × n.
In contrast, the image padding process involves finding the max value m of the height h and width w of the input image and then computing the ratio of n to the max value m. Thereafter, it changes the input image size to the integer value of (h × ratio, w × ratio). Finally, it creates the blank image with the dimension of n × n and pastes the newly resized image into the blank image (Figure 3).
Figure 3. Image scaling and padding.

3.3. Features Extraction

As mentioned in Section 2, using deep learning to extract image features improves related system results due to deep learning’s ability to extract deep features, such as efficiently detecting an intended object within a short time frame. Therefore, two pretrained CNNs were individually utilized to extract image features. The features were extracted from the pooling layer of a ResNet-50 and VGG16 with an output dimension of 2048 and 512, respectively. Then, their performance was compared.

3.4. Features Reduction

Many problems arise when working on high-dimensional features, i.e., adding too many features worsens recognition performance; this problem is known as the curse of dimensionality [22]. Feature reduction refers to the techniques used to reduce the size of the vectors. As mentioned above, the dimensions of the features produced by ResNet-50 and VGG-16 are too large; therefore, reducing the features is essential to overcome future issues.
PCA is an unsupervised technique used in machine learning for reducing dimensionality [22]. It aims to reduce the data dimensions from n variables to m components, where m < n. This means representing the data using m components (instead of n), accounting for the most variance [23], i.e., whenever the correlation between features is high, the PCA will combine these correlated features within a smaller number of uncorrelated features.
In this study, PCA was used to reduce the features extracted by two models to improve the recognition/detection performance. PCA works better if a linear structure in the data and variables are scaled in the same unit. Since the features were already normalized, features scaling was unnecessary. Moreover, the features extracted by the ResNet-50 and VGG-16 models were represented by 1450 and 410 components, respectively. These components explained more than 95% of the variance.

3.5. Similarity Measurement

As mentioned above, the output of the preceding module is the feature vector. Thus, in order to compute the similarity between the query and the stored dataset images, some computation should be conducted to calculate the distance between their feature vectors. If the distance is too small, then the images will be considered similar to each other. According to the literature, the Euclidean distance is one of the most effective measures in terms of classifying trademarks. The distance is computed from the query image to each image in the database. Next, the images are ranked based on their distance, in increasing order. Then, the system retrieves n ranked images corresponding to the user input. Consider the feature vector q for a given query and set of database features vectors a1, a2,…, an. The system will measure the similarity between q and every feature vector from a1 to an. The Euclidean distance will be calculated using the following equation [24]:
d ( a i , q ) = ( i = 1 n | a i q | 2 ) 1 2

3.6. Evaluation

Evaluation is a process that is intended to ensure that the system performs in the way it should. In similarity detection, the system should be able to retrieve the images most similar to the query, or at least one of them. To this end, the proposed system was evaluated using three measures: average rank, normalized average rank, and mAP. The average rank metric computes the average ranks of the image belonging to the same query class [21]. As shown in Equation (2), Nrel refers to the number of relevant images in the query class, and Ri the rank of the ith relevant query.
A v e r a g e   R a n k = 1 N r e l   i = 1 N r e l R i
The normalized average rank value is between [0,1], where zero represents the best retrieval performance and 0.5 is random performance [21]. The normalized average rank equation is given below, where N refers to the size of the dataset.
N o r m a l i z e d   A v e r a g e   R a n k :   R a n k ˜ = 1 N × N r e l   ( i = 1 N r e l R i N r e l ( N r e l + 1 ) 2 )  
The value mAP is the most frequently metric used to evaluate object-detection systems [25]. Average precision (AP) is the positive value over the predicted positives, where a true positive equals 1 for an image relevant to the query q; otherwise, it is 0. Therefore, mAP is the mean of the AP for each q made over the total number of queries. The higher the mAP value, the better the performance.
m A P = 1 Q   i = 1 Q A P i
where Q is the number of queries in the set.

4. Experiments and Results

Several parameters and procedures played essential roles in the final results. Thus, many experiments with different techniques should be conducted to consider all possible situations. Some of these procedures include the methods used to process the input images and extract the image features, as well as additional processes to improve the system performance.
There is no single optimal or standard CNN structure. Each problem might be solved using various CNN structures, but the best-fitting one can be obtained through experiments for a given problem. Indeed, several CNN architectures that were pretrained on a large dataset for image classification purposes and achieved high accuracy are available to reuse for related problems.
With reference to one of the most valuable competitions that encourages researchers and developers to develop such a solution, is the ImageNet Large Scale Visual Recognition Challenge (ILSVRC). It is an annual competition that was conducted from 2010 to 2017 for object detection and image classification. Many of the current innovations have sprung from the ILSVRC competition. AlexNet, VGGNet, and ResNet are some of the popular CNN architectures that were developed during an ILSVRC [26].
AlexNet was introduced in 2012 by Krizhevsky et al. [27]. AlexNet was trained on more than 1.2 million images of the ImageNet dataset to classify them into 1000 classes belonging to different natural objects. It consists of 650,000 neurons and eight layers in total (five convolutional layers and three fully connected layers). The max-pooling layer follows the first, second, and fifth convolutional layers. Moreover, the rectified linear unit (ReLU) is applied to each neuron’s output. The ReLU is the factor that makes AlexNet more efficient compared with other activation functions, as it helps to accelerate the learning process and computational efficiency. To reduce the overfitting, data augmentation and dropout methods were applied. As a result, AlexNet won the ILSVRC-2012 competition by achieving top-1 and top-5 test set error rates of 37.5% and 17%, respectively.
VGGNet was inspired by the AlexNet architecture [28]. It consists of a 3 × 3 kernel-size filter to produce a better recognition rate than the large-size filter. There are two different architectures of VGGNet: VGGNet16 and VGGNet19; these consist of 16 and 19 layers, respectively. Both architectures are composed of 3 × 3 convolutional, 2 × 2 max-pooling, and three fully connected layers. The last of the fully connected layers utilizes a softmax for classification purposes. Consequently, VGGNet achieved a top-5 test set error rate of 6.8%.
ResNet stands for Residual Network, a neural network introduced by K. He et al. [29]. It was inspired by the VGG network. The main idea behind ResNet is to “identity shortcut connections,” which skip one or more network layers. The ResNet architecture consists of a 34-layer plain network, along with the shortcut connection. Most convolution layers have 3 × 3 filters and perform downsampling with a stride of two. The network ends with a global average pooling layer, followed by a fully connected layer that utilizes a softmax. As a result, ResNet won first place in the ILSVRC-2015 classification competition by achieving a top-5 test set error rate of 3.57%.
The following subsections briefly describe each experiment along with the corresponding evaluation results. The experiments are divided based on the image features extracted to make them easy to read.

4.1. Experiments

The main component of a similarity detection system is a feature extractor that involves image-fetching features to help distinguish similar images from others. In this study, two pretrained CNNs were individually utilized as features extractors: ResNet-50 and VGG-16. Moreover, trademark images often require much less preprocessing due to them being artificial and many processes being conducted automatically using the pretrained CNNs. Furthermore, all of the experiments were conducted on a portion of the METU test set, equaling 10,000 images, to accelerate the process, along with the injected query set, to evaluate the retrieval performance. The following two sections describe the experiments conducted with the help of two feature extractors.

4.1.1. ResNet-50 Features Extractor

All of the following experiments utilized the ResNet-50 model to extract image features.
1.
ResNet-50-based similarity detection (RSD) with image scaling-up: This experiment utilized the ResNet-50 model as a feature extractor. Since the images are a variety of sizes and the pretrained CNNs required a particular input size, in this experiment, all of the images were scaled to the size required by a model, specifically 224 × 224.
2.
RSD with image scaling-up and feature reduction using PCA: The dimension of the extracted features by the ResNet-50 model is equal to 2048, which is large to handle and process. So, there is a benefit to reducing the overall features. Therefore, this experiment used features reduced by PCA of scaling images.
3.
RSD with image padding: In this experiment, each image was resized and padded into a 224 × 224 image, which is the size required by ResNet-50.
4.
RSD with image padding and feature reduction using PCA: This experiment used features reduced by PCA of the padding images.

4.1.2. VGG-16 Features Extractor

All of the following experiments were conducted based on the features extracted by VGG-16.
1.
VGG-16-based similarity detection (VSD) with image scaling-up: This experiment utilized the VGG-16 model to extract all of the images’ features. Moreover, all images were scaled to the size required by a VGG-16, which is equal to 224 × 224.
2.
VSD with image scaling-up and feature reduction using PCA: As mentioned above, the dimension of the extracted features should be reduced. So, this experiment used features reduced by PCA of scaling images.
3.
VSD with image padding: In this experiment, each image was padded into a 224 × 224 image, which is the input size required by VGG-16.
4.
VSD with image padding and features reduction using PCA: As with the previous experiments, this experiment investigated the effect of the feature reduction technique for the features extracted from the padded images.
Table 1 shows the average and normalized average ranks for each conducted experiment using the ResNet-50 features extractor and VGG-16 features extractor, respectively.
Table 1. Result of the average rank and normalized average rank for each experiment.
The above table shows that Exp#1 and Exp#2 achieved results that are close to each other; the average rank and R of the related experiment were improved by applying the PCA reduction technique. In contrast, the worst results were obtained from the features extracted by the VGG-16 pooling layer for padded images. Table 2 presents the mAP value for each experiment.
Table 2. Mean average precision results for each experiment.
As shown in Table 2, the mAP values of the first four experiments were close to each other, which represent the experiments that utilized ResNet-50 as a feature extractor. Hence, the ResNet-50 feature extractor overcomes the VGG-16 performance in the conducted experiments.
Figure 4 presents some examples of the developed trademark similarity detection system. The green rectangles identify the true positives (images belong to the same query class); the remaining images are bounded into red rectangles. A yellow background was used to clearly show the results.
Figure 4. Some examples of the images retrieved for a given query.
Table 3 presents the comparison between the performance of the proposed approach and other studies other previous studies that use similar approaches, where they used the same dataset.
Table 3. Comparison between the proposed approach and sate-of-the-art approaches.
Tursan et al. [21] used pre-trained models with different parameters over the METU dataset, and the output of FC7 of VGG16 achieved the best retrieval performance, which was 0.086 of NAR.
Feng et al. [30] proposed a method that extracted features starting from trademarks’ edge images until the enhanced SIFT features. As a result, their model achieved an NAR of 0.083, outperforming some of the deep-learning methods.
Cao et al. [31] proposed using the ResNet50 model on the METU dataset by considering features from the FC layer. Their propsed model achieved a NAR of 0.110.
As shown in Table 3, the retrieving performance of the relation database method overcomes the methods proposed by Tursan et al., Feng et al., and Cao et al. with the fully-connected layer. Additionally, it achieved an mAP of 0.774, which means that the proposed approach gives a promising result in detecting trademark similarity.

5. Discussion

In our study, we used mAP to evaluate the retrieval performance of our proposed model. We also compared our proposed method with other studies using the average rank and normalized average rank. As shown in Table 1 and Table 2, the best performance was achieved by Exp#2, which was based on reducing the features extracted by ResNet-50 for rescaled dataset images. It achieved an average rank of 774.447, a R a n k ˜ of 0.0737, and an mAP of 0.759, which were the best results compared to the other experiments. Moreover, 85% of the relevant images to a given query appeared on the top 20 retrieved images.
Visual analysis helped in further understanding the system’s performance. Whenever the query was a trademark with a clear shape, the system was either able to retrieve all relevant images or most of them correctly (see Figure 5). In contrast, the text included in the query image could mislead the system to detect the core component because it was dealing with each letter as a shape (Figure 5a). Therefore, the system must be improved to consider semantic similarities. Moreover, the detection performance was primarily affected by some parameters, such as the arrangement of the objects and spaces (Figure 5b). Additionally, when evaluating this system, some of the most similar images to the query existed in the test set (large unclassified dataset), but they were not considered to be similar because they did not belong to the same query class. For example, as shown in Figure 5c, the images in the red rectangles appeared in the top 15 retrieved images for a related query, but they were not considered relevant.
Figure 5. Examples of samples retrieved from the test set (in red rectangles), with a related query: (a) Trademark containing text. (b) Trademarks have the same arrangement of objects. (c) Similar retrieved images.

6. Conclusions

There is potential for machine learning applications to enhance the examination process by providing examiners with an automated image detection system that would facilitate improved accuracy in the detection process. This study aimed to provide a trademark similarity detection system based on deep learning that is able to automatically and accurately detect and retrieve the registered trademarks that are most similar to the input mark. The target system users are both IP examiners and individual users. The system provides assessments in the examination process by retrieving the trademark images that most closely resemble the input for the IP examiners. Thus, the examiners compare the query input with the limited number of most similar marks retrieved by the system. It helps the individual users by providing them the guarantee that a trademark is a right reserved exclusively by them and is not similar to any prior ones. As a result, the system achieved an average rank of 67,067.788, a normalized average rank of 0.073, and a mAP of 0.774.
ResNet50 model was used only by a few studies for image features extraction, mainly the trademarks. In our study, we found that the deep architecture of ResNet50 could help extract the distinct features, even without training on an annotated dataset.
Future directions for our work include considering the semantic similarities to be able to detect a similarity among text-only trademarks. Moreover, a large annotated trademark dataset will be built to fine-tune the model in order to improve the model’s performance.

Author Contributions

H.A. conceived, designed, and performed the experiment; analyzed and interpreted the data; and drafted the manuscript. A.A.-N. co-supervised the analysis, reviewed the manuscript, and contributed to the discussion. Y.A.-O. supervised this study. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

The authors would like to thank the Deanship of Scientific Research (DSR) in King Saud University for funding and supporting this research through the initiative of DSR Graduate Students Research Support (GSR). Also, our thanks go out to the staff of the Saudi authority for intellectual property (SAIP) for their help and support.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Roy, A.M. A deep learning enabled multi-class plant disease detection model based on computer vision. AI 2021, 2, 413–428. [Google Scholar] [CrossRef]
  2. Bhaduri, R.J. Real-time growth stage detection model for high degree of occultation using DenseNet-fused YOLOv4. Comput. Electron. Agric. 2022, 193, 106694. [Google Scholar] [CrossRef]
  3. Trappey, C.V.; Trappey, A.J.C.; Lin, S.C.C. Intelligent trademark similarity analysis of image, spelling, and phonetic features using machine learning methodologies. Adv. Eng. Informatics 2020, 45, 101120. [Google Scholar] [CrossRef]
  4. Macías, W.; Cerviño, J. Trademark dilution and its practical effect on purchase decision. Spanish J. Mark. ESIC 2017, 21, 1–13. [Google Scholar] [CrossRef]
  5. WIPO. Trademarks; WIPO: Geneva, Switzerland, 2019. [Google Scholar]
  6. Showkatramani, G.; Nareddi, S.; Doninger, C.; Gabel, G.; Krishna, A. Trademark Image Similarity Search. In HCI International 2018—Posters’ Extended Abstracts; Springer International Publishing: Berlin/Heidelberg, Germany, 2018; pp. 199–205. ISBN 9783319922706. [Google Scholar]
  7. Saudi Authority for Intellectual Property. Available online: https://www.saip.gov.sa (accessed on 21 October 2021).
  8. Eakins, J.P.; Boardman, J.M.; Graham, M.E. Similarity retrieval of trademark images. IEEE Multimed. 1998, 5, 53–63. [Google Scholar] [CrossRef]
  9. Haddad, M.; Saleh, A.; Jaber, Y. Trademark image recognition utilizing deep convolutional neural network. Ann. Electr. Electron. Eng. 2019, 2, 13–20. [Google Scholar]
  10. Perez, C.A.; Estévez, P.A.; Galdames, F.J.; Schulz, D.A.; Perez, J.P.; Bastías, D.; Vilar, D.R. Trademark Image Retrieval Using a Combination of Deep Convolutional Neural Networks. In Proceedings of the 2018 International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018. [Google Scholar]
  11. Lan, T.; Feng, X.; Li, L.; Xia, Z. Similar trademark image retrieval based on convolutional neural network and constraint theory. In Proceedings of the 2018 Eighth International Conference on Image Processing Theory, Tools and Applications (IPTA), Xi’an, China, 7–10 November 2018; pp. 1–6. [Google Scholar]
  12. Bhunia, A.K.; Bhunia, A.K.; Ghose, S.; Das, A.; Roy, P.P.; Pal, U. A deep one-shot network for query-based logo retrieval. Pattern Recognit. 2019, 96, 106965. [Google Scholar] [CrossRef] [Green Version]
  13. Tursun, O.; Denman, S.; Sivapalan, S.; Sridharan, S.; Fookes, C.; Mau, S. Component-based Attention for Large-scale Trademark Retrieval. IEEE Trans. Inf. Forensics Secur. 2019, 1. [Google Scholar] [CrossRef] [Green Version]
  14. Mosseri, I.; Rusanovsky, M.; Oren, G. TradeMarker—Artificial Intelligence Based Trademarks Similarity Search Engine; Springer International Publishing: Berlin/Heidelberg, Germany, 2019; Volume 1034, ISBN 978-3-030-23524-6. [Google Scholar]
  15. Xia, Z.; Lin, J.; Feng, X. Trademark image retrieval via transformation-invariant deep hashing. J. Vis. Commun. Image Represent. 2019, 59, 108–116. [Google Scholar] [CrossRef]
  16. Morita, H.; Abe, K.; Hayashi, T. A similarity retrieval of trademark images considering similarity for local objects using vector images. In Proceedings of the 8th International Conference on Signal Processing Systems, Auckland, New Zealand, 21–24 November 2016; pp. 79–82. [Google Scholar]
  17. Abe, K.; Morita, H.; Hayashi, T. Similarity retrieval of trademark images by vector graphics based on shape characteristics of components. In Proceedings of the 2018 10th International Conference on Computer and Automation Engineering, Brisbane, Australia, 24–26 February 2018; pp. 82–86. [Google Scholar]
  18. Khamkar, S. User Retrieval of Trademarks System Using Conceptual Similarity Approach. Helix 2018, 8, 3754–3758. [Google Scholar] [CrossRef] [Green Version]
  19. Aires, S.B.K.; Freitas, C.O.A.; Sguario, M.L. Dynamic selection feature extractor for trademark retrieval. Lect. Notes Comput. Sci. (including Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinformatics) 2018, 11288, 219–231. [Google Scholar]
  20. Lan, T.; Feng, X.; Xia, Z.; Pan, S.; Peng, J. Similar trademark image retrieval integrating LBP and convolutional neural network. Lect. Notes Comput. Sci. (including Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinformatics) 2017, 10668, 231–242. [Google Scholar]
  21. Tursun, O.; Aker, C.; Kalkan, S. A Large-scale Dataset and Benchmark for Similar Trademark Retrieval. arXiv 2017, arXiv:1701.05766, 1–33. [Google Scholar]
  22. Goonewardana, H. PCA: Application in Machine Learning. medium, 2019. Available online: https://medium.com/apprentice-journal/pca-application-in-machine-learning-4827c07a61db (accessed on 21 October 2021).
  23. Arguello, J.; Crescenzi, A. Using principal component analysis to better understand behavioral measures and their effects. In Proceedings of the 2019 ACM SIGIR International Conference on Theory of Information Retrieval, Santa Clara, CA, USA, 2–5 October 2019; pp. 177–184. [Google Scholar]
  24. Russell, S.; Norvig, P. Artificial Intelligence A Modern Approach; Prentice Hall: Hoboken, NJ, USA, 2020; Volume 53, ISBN 9788578110796. [Google Scholar]
  25. Tan, R.J. Breaking Down Mean Average Precision (mAP). Towards Data Science. 2019. Available online: https://towardsdatascience.com/breaking-down-mean-average-precision-map-ae462f623a52 (accessed on 21 October 2021).
  26. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  27. Kumar, G.; Bhatia, P.K. A detailed review of feature extraction in image processing systems. In Proceedings of the 2014 Fourth international conference on advanced computing & communication technologies, Rohtak, India, 8–9 February 2014. [Google Scholar] [CrossRef]
  28. Trappey, J.C.; Trappey, C.V.; Lin, S.C.C. Detecting trademark image infringement using convolutional neural networks. Adv. Transdiscipl. Eng. 2019, 10, 477–486. [Google Scholar]
  29. Czum, J.M. Dive Into Deep Learning. J. Am. Coll. Radiol. 2020, 17, 637–638. [Google Scholar] [CrossRef]
  30. Feng, Y.; Shi, C.; Qi, C.; Xu, J.; Xiao, B.; Wang, C. Aggregation of reversal invariant features from edge images for large-scale trademark retrieval. In Proceedings of the 2018 4th International Conference on Control, Automation and Robotics (ICCAR), Auckland, New Zealand, 20–23 April 2018. [Google Scholar]
  31. Cao, J.; Huang, Y.; Dai, Q.; Ling, W.K. Unsupervised trademark retrieval method based on attention mechanism. Sensors 2021, 21, 1894. [Google Scholar] [CrossRef] [PubMed]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.