Skip Content
You are currently on the new version of our website. Access the old version .
Applied SciencesApplied Sciences
  • Article
  • Open Access

8 October 2024

Combination Pattern Method Using Deep Learning for Pill Classification

,
,
and
1
Department of AI Convergence & Engineering, Open Cyber University of Korea, C-9F, 353, Mangu-ro, Jungnang-gu, Seoul 02087, Republic of Korea
2
Department of Visual Design, Hyupsung University, 72, Choerubaek-ro, Bongdam-eup, Hwaseong-si 18330, Republic of Korea
3
Department of Computer Engineering, Pai Chai University, 155-40 Baejae-ro, Seo-gu, Daejeon 35345, Republic of Korea
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Integrating Artificial Intelligence in Renewable Energy Systems

Abstract

The accurate identification of pills is essential for their safe administration in the medical field. Despite technological advancements, pill classification encounters hurdles such as ambiguous images, pattern similarities, mixed pills, and variations in pill shapes. A significant factor is the inability of 2D imaging to capture a pill’s 3D structure efficiently. Additionally, the scarcity of diverse datasets reflecting various pill shapes and colors hampers accurate prediction. Our experimental investigation shows that while color-based classification obtains a 95% accuracy rate, shape-based classification only reaches 66%, underscoring the inherent difficulty distinguishing between pills with similar patterns. In response to these challenges, we propose a novel system integrating Multi Combination Pattern Labeling (MCPL), a new method designed to accurately extract feature points and pill patterns. MCPL extracts feature points invariant to rotation and scale and effectively identifies unique edges, thereby emphasizing pills’ contour and structural features. This innovative approach enables the robust extraction of information regarding various shapes, sizes, and complex pill patterns, considering even the 3D structure of the pills. Experimental results show that the proposed method improves the existing recognition performance by about 1.2 times. By improving the accuracy and reliability of pill classification and recognition, MCPL can significantly enhance patient safety and medical efficiency. By overcoming the limitations inherent in existing classification methods, MCPL provides high-accuracy pill classification, even with constrained datasets. It substantially enhances the reliability of pill classification and recognition, contributing to improved patient safety and medical efficiency.

1. Introduction

Medication safety is a critical concern in healthcare due to its direct impact on patient health. The accurate identification of medications is essential, as errors can lead to serious adverse effects, including injuries and complications. Oral pills, widely used because of their stability and ease of administration, pose significant challenges, particularly for patients with memory problems. Reports from the Organization for Economic Co-operation and Development (OECD) indicate that medication errors contribute to 18.3% of adverse patient events [1], with a significant portion resulting from confusion over pill shapes and colors, as highlighted by studies in the British Medical Journal (BMJ) [2] and by the World Health Organization (WHO) [3]. This problem is exacerbated during the medication cycle—prescribing, dispensing, and administration—where physical characteristics of pills often lead to errors [4,5].
The variability in pill shapes, sizes, and colors, which may change over time due to factors like wear and discoloration, complicates reliable identification. Traditional machine learning models used for pill detection struggle to keep pace with the rapid introduction of new drugs, often failing to adapt to these changes, as evidenced by studies showing that nearly 25% of medications go unidentified in university hospitals [6,7]. Pentagonal pills present unique challenges among the various pill shapes due to their complex geometry, including pointed edges and a mix of curves and straight lines, as illustrated in Figure 1.
Figure 1. Pentagonal pill shapes with subtle differences.
These identification challenges are compounded by rotation invariance and perspective distortion, where a pill’s orientation in an image can drastically alter its perceived shape. Variable lighting conditions that distort visual features further exacerbate such issues, underscoring the need for more sophisticated algorithms that can effectively navigate these complexities.
To address these issues, numerous studies have explored the potential of computer vision technology in pill detection. A widely used analytical technique involves comparing similarities between input and reference images based on shape, color, and imprinting features. These methods are broadly categorized into pixel-based and shape-based matching. Pixel-based methods calculate differences between corresponding pixels and compute similarity metrics such as the sum of squared differences or normalized cross-correlation [8,9]. Although effective against image distortion, they cannot detect changes in size or rotation. In contrast, shape-based matching extracts the subject’s region of interest (ROI) from a reference image, better accommodating object size and orientation variations, and excels in variable illumination settings [10,11,12].
Recent advancements in deep learning-based object-detection algorithms have been pivotal in developing self-learning and decision-making capabilities for pill-recognition solutions [13]. These neural networks excel in tasks such as object detection, biomedical diagnosis, and variant detection thanks to their ability to process and learn from large datasets [14,15,16,17]. Popular methods include You Only Look Once (YOLO), convolutional neural networks (CNNs), and recurrent neural networks (RNNs). Achieving high detection performance, however, necessitates extensive data due to the nuanced variability in pill shapes, textures, and colors. This variability demands a comprehensive approach that accounts for these subtle differences, making pill identification a complex visual classification challenge that requires precise differentiation between pill fragments.
To address these challenges, we propose the Multi-Combination Pattern Labeling (MCPL) method, a novel approach designed to enhance the accuracy and reliability of pill classification in challenging image conditions. The MCPL method extracts unique and representative features from pill images, providing a rich representation of the pill’s characteristics to improve classification performance. To accomplish this, MCPL combines features obtained from multiple established image feature-extraction algorithms: KAZE, Oriented FAST and Rotated BRIEF (ORB), Shi-Tomasi, and Harris. KAZE offers scale and rotation-invariant feature-detection robustness, capturing intricate details and broader patterns within images. ORB combines the FAST key point detector with the BRIEF descriptor for fast and efficient feature matching. Shi-Tomasi and Harris focus on corner detection, with Harris excelling under changes in rotation and illumination.
By combining these algorithms, MCPL comprehensively represents pill images, capturing a wide array of features, including shapes, patterns, and details that may be missed by using a single extraction method.
To effectively tackle these identification challenges, we employ the YOLO version 5 object-detection model to initially identify pills within images. This step is crucial as it sets the stage for more detailed analysis. Following this initial detection, our convolutional networks undertake a thorough analysis of the pills’ shape, color, and form, focusing particularly on those with complex geometries. This targeted approach is vital for the MCPL method to refine the feature-extraction processes, significantly enhancing classification accuracy and adapting to the diverse challenges presented by different pill types. The contributions of this study are as follows.
  • Proposal of a new approach, Multi-Combination Pattern Labeling method, designed to enhance the accuracy and reliability of pill classification in challenging image conditions. The proposed method extracts unique and representative features from pill images, providing a rich representation of the pill’s characteristics to improve classification performance.
  • Proposal of a pill-detection model-based MCPL to extract pills’ shape, color, and form, focusing particularly on those with complex geometries.
  • Verification of the proposed methods through the experimental results.
This paper is organized as follows: Section 2 introduces the proposed method and summarizes related studies, providing the necessary context and background for our approach. Section 3 details the deep learning-based pill-classification system developed in this study, elucidating technical details and integration of the MCPL method. Section 4 elaborates further on the MCPL method, describing how it utilizes various image feature-extraction algorithms to surmount the identified challenges. Section 5 examines the experimental results, assessing the efficacy of the MCPL method across various scenarios and pill types. Finally, Section 6 presents our concluding remarks, discussing the implications of our findings and suggesting potential avenues for future research.

3. The Proposed Pill-Classification System

The following sections elaborate on the deep learning-based pill-classification system developed in this research, incorporating feedback to enhance the clarity and depth of the methodologies employed. The MCPL algorithm is pivotal in our approach, aiming to improve identification accuracy by extracting discernible features from pills, particularly those that are challenging to classify based on limited training data. We first collected around 25,000 pill images for building the dataset, either by directly downloading them or by crawling various sites that provide pill images. As illustrated in Figure 4, the system architecture is structured first to detect and analyze the colors and shapes of pills—fundamental attributes for initial sorting and identification.
Figure 4. Overview of proposed pill-classification system.

3.1. Pill Detection

We utilize the YOLOv5 [S] architecture for pill detection, composed of four distinct subnetworks: the backbone, neck, head, and detect layer. YOLOv5 models—available in sizes S, M, L, and X—differ in capacity and computational efficiency, each tailored to specific operational requirements. Our implementation in PyTorch version 1. 7 leverages the generality of the trained YOLOv5 model, making it highly adaptable across varied pill-identification scenarios. As depicted in Figure 5, the pill-detection workflow entails three primary steps:
Figure 5. YOLOv5-based pill-detection step for pill classification.
-
Step1: Localization and Configuration: Initial detection of pill locations and configurations crucial for setting up the training dataset.
-
Step2: Feature Extraction: The backbone network, utilizing a Darknet architecture enhanced by a Cross Stage Partial Network (CSPNet), extracts a broad spectrum of features from pill images, including edges, textures, and distinctive patterns.
-
Step3: Detection and Classification: The combined efforts of the neck and head networks, which employ spatial pyramid pooling (SPP) and anchor boxes, refine the detection and classification processes. This stage culminates with applying non-maximum suppression (NMS) to accurately identify and categorize pill types based on the extracted features.

3.2. Color and Shape Detection

The identification of pills also heavily relies on distinguishing features such as color and shape. For example, Amaryl pills, used in diabetes management, share the same shape and size but are differentiated by color based on dosage. We employ a CNN model to detect these features, focusing mainly on color because of its significance in the identification of the pill. The CNN’s ability to process and classify images based on learned features allows it to effectively separate pills that appear similar but differ in critical attributes like color. This section details how the CNN model is specifically trained to recognize and categorize pills based on these attributes, providing a foundational step in the classification process.
The CNN model utilized for color detection has also been adapted for shape identification, leveraging the robust feature-detection capabilities of CNNs. This approach is supported by extensive research in the field. References [42,43] applied neural-network-based methods to detect pills in images. References [44,45] focused on identifying pills in images portraying multiple different pills. Reference [46] aimed to detect the presence of pills in blister images for pill tracking, as well as identify pills in images of the mouth. Reference [47] attempted to learn potential associations between pills using information extracted from prescriptions. Although the problems covered in previous studies differ from the present study’s objective, our work aims to complement the results of these studies by presenting an accurate and efficient pill-identification method.
In this study, two primary methods are implemented to classify pills using the CNN framework. The first method processes entire images as input, dividing them into multiple layers and applying a specific filter to each layer. This technique demonstrates high accuracy but requires extensive processing time and significant data. The second method focuses on extracting key features from pill images and using these as direct inputs to the CNN. This method significantly reduces processing time and increases accuracy by concentrating on essential information extracted directly from the pills.
After verifying the accuracy of both methods, we introduced an additional feature-extraction process tailored for pills that are difficult to classify based on standard attributes. This enhancement improves the robustness of our pill-classification system, allowing for more precise identification across a broader range of pill types and conditions.

3.3. Feature Detection

In the proposed MCPL-based pill-classification system, feature detection plays a crucial role in enhancing the accuracy and reliability of pill identification. The feature-detection process is designed to capture and analyze the overt and subtle characteristics of pills that are significant for their classification.
First, we detect color and texture as specific features of the pill. These are essential for distinguishing pills with similar shapes but different therapeutic uses. The system identifies subtle variations in color gradients and texture patterns characteristic of different pill coatings and materials. We also detect the shape and size of the pill, as contour analysis techniques detect the geometric features of pills, including round, oval, rectangular, and other irregular shapes. Size measurements help differentiate between dosage forms.
For feature-extraction techniques, we first perform initial image processing, then utilize advanced algorithms for feature detection, and finally apply deep learning models. High-resolution images are pre-processed using techniques such as denoising, normalization, and color correction to enhance the quality of the input data for feature extraction. Next, the proposed method employs a combination of edge detection, corner detection, and blob-detection algorithms to identify and outline critical features. Algorithms like the Shi-Tomasi and Harris corner detectors identify key points on the pill images that are invariant to rotation and scale. CNNs are used for more complex feature detection, such as identifying textural patterns and classifying pills based on learned image features from extensive datasets. This approach enhances the pill-classification system’s ability to handle variability in pill appearances and ensures high identification accuracy.

4. Multi-Combination Pattern Labeling

In this section, we propose the MCPL method to achieve accurate pill classification based on extracted features. In pill-classification systems, the feature extraction of hard-to-classify pills is indispensable in maximizing accuracy and reliability by minimizing the risks of drug misuse and adverse effects. To this end, we implemented appropriate feature-extraction procedures to supplement the information needed to identify hard-to-classify pills.
The MCPL model detects features and utilizes them as criteria for the classification of complex pill shapes. To achieve this, the KAZE algorithm, which responds to various changes in image size, rotation, and orientation, is used in conjunction with the ORB algorithm, which performs fast computation and matching. Furthermore, kernel features are found using the Shi-Tomasi and Harris algorithms to ensure more accurate feature extraction and matching.

4.1. KAZE

The KAZE algorithm employs a nonlinear scale space, which utilizes nonlinear scale filtering to make image feature blurring locally adaptive. This approach helps in reducing noise while preserving image boundaries. The KAZE detector is configured based on a scale-normalized determinant of the Hessian matrix computed at multiple scale levels. The maxima of the detector response are selected as features using a moving window. Feature description introduces the rotation invariance property by finding the dominant orientation in the circular neighborhood around each detected feature. Although KAZE features are invariant to rotation, scale, and limited affine transformations and have more distinct characteristics at different scales, they incur a longer computational time. Figure 6 illustrates the structure of KAZE-based feature detection. The process begins with image pyramid generation, creating multiple image scales. Scale extrema are identified across these scales, followed by orientation estimation represented by the colored lines, each indicating different gradient directions at specific points. Next, descriptor computation generates feature vectors that capture key point information, while feature matching compares these descriptors across different images. In the transformation computation phase, matching key points determine geometric transformations between the images. The final step, orientation estimation, refines the key points matches, ensuring accurate alignment and orientation, visualized by the connected key points and transformation overlays.
Figure 6. Structure of KAZE-based feature detection.
The KAZE detector initially generates an image pyramid, wherein the input image is repeatedly down-sampled to detect features at different scales. The scale extrema step detects the maximum and minimum points in the scale space of each image based on the LoG filter. Orientation estimation is performed using the gradient orientation of the image to approximate the orientation at each key point. The descriptor computation step obtains the path of the region around the key point to extract local features from that key point. The rotation is corrected using the orientation information from the path and computed based on the SIFT descriptor. In the feature-matching step, distances are calculated between descriptors to match key points between two images via Nearest-Neighbor (NN) and the Nearest-Neighbor-Ratio (NNR) matching. To compute the transformation, the NNR matching results are used to calculate the affine transformation matrix, which is used to align features between the two images and subsequently identify the key points between them. Figure 7 presents sample results of the KAZE algorithm. The left side of the Figure 7 shows the extracted features as colored circles, and the right side shows the matching result based on the features.
Figure 7. KAZE-based feature extraction and matching.

4.2. ORB

The ORB algorithm is a combination of the modified FAST detection and orientation-normalized BRIEF techniques. FAST detects possible corner points at each layer of the scale pyramid, and each point is then evaluated using the Harris corner score to filter points with the highest quality. The features of ORB are points that are invariant to scalar, rotational, and limited affine changes.
The ORB algorithm comprises the following procedure: (1) configure a multi-resolution pyramid of the image; (2) perform FAST corner detection at each resolution; (3) evaluate the ‘cornerness’ value by calculating the Harris corner score from the detected corners; (4) remove redundant corners by selecting only the corners with the highest Harris corner scores; (5) estimate the orientation from the selected corners; (6) generate a BRIEF descriptor based on the orientation to ensure invariance to rotation; (7) display the features using the descriptors. Figure 8 presents a sample result of ORB-based feature detection.
Figure 8. ORB-based feature extraction and matching.

4.3. Shi-Tomasi and Harris Algorithms

The Shi-Tomasi and Harris algorithms are among the most popular corner-detection algorithms. To identify the corners of pills in an input image, the Shi-Tomasi algorithm creates a grayscale version of the image, selects a window in the image, and calculates the gradient at every pixel inside the window. It then applies the derivative at each pixel to compute a feature value from the magnitude and direction of the gradient. The feature value is the smallest of the gradient magnitudes of the pixels inside the window. The top N features are selected by comparing the magnitudes of feature values over the entire image.
Although the Harris algorithm is similar to the Shi-Tomasi algorithm, it uses a corner-response function to detect corners. In Figure 9, Shi-Tomasi (left), the blue circles show corner points detected based on the smallest eigenvalue of the gradient matrix. This method’s sensitivity to subtle image variations results in more scattered corners. In Harris (right), the blue circles indicate fewer but stronger corner detections. The algorithm focuses on areas with significant intensity changes in multiple directions, offering more stable corner identification.
Figure 9. Shi-Tomasi and Harris algorithm for corner detection.

5. Experimental Results

This section presents and interprets the experimental results of our proposed deep learning-based pill-classification system. First, we describe the results of the proposed base pill-recognition system, and then the experimental results based on MCPL. The proposed system uses a CNN to recognize the colors and shapes of pills. Although CNNs have exhibited high performance in tasks including image classification, object detection, and segmentation, they require large quantities of data for training. Accordingly, we applied the image-augmentation technique to generate new training data from existing images via transformations, allowing the CNN model to learn more diverse patterns. To ensure a sufficiently large and diverse dataset, we rotated the image data by random angles between 0 and 25 degrees, shifted the orientation in horizontal and vertical directions by random values within a range of 0.05, performed scaling by random values between 0.8 and 1.2, and inverted the original images. Figure 10 illustrates the CNN-facilitated processes of image augmentation and color detection. Five color classes were used for classification: class = [‘red’, ‘green’, ‘blue’, ‘orange’, ‘white’]. The experimental results show a pill color detection accuracy of 95%.
Figure 10. Image augmentation and color detection using CNN.
Figure 11 illustrates the CNN-based shape-detection procedure. In our experimental setup, we designated Models A and B to compare detection accuracy according to class. In the classification results of the figure, the blue graph indicates correct classifications, while the red graph signifies incorrect classifications. Figure 11a presents the pill shape-detection results obtained by Model A. For shape detection, six classes were initially defined—class = [‘circle’, ‘hexagon’, ‘pentagon’, ‘rectangle’, ‘capsular’, ‘triangle’]—yielding an accuracy of 66%. Specifically, the detection accuracies of circle, triangle, rectangle, and capsular shapes were 95%, 65%, 62%, and 75%, respectively, whereas those of hexagon and pentagon shapes were 29% and 32%, respectively.
Figure 11. CNN-facilitated shape detection.
Figure 11b presents the pill shape-detection results obtained by Model B, where in four classes were defined: class = [‘circle’, ‘rectangle’, ‘triangle’, ‘etc’]. The number of classes was reduced to make the pill classification simpler and more intuitive. The etc class was added to allow shapes that were not correctly classified by Model A to be organized within a single category. The resulting detection accuracy was 75%, which is an improvement over Model A. The experimental results once again confirm the importance of identifying features for accurate classification, as complex shapes may negatively impact overall model performance. Accordingly, features identified as etc require further detection based on complex shapes.
We detect shapes of pill using MCPL for the hexagon and pentagon classes, which are not well classified in the proposed base system, as shown in Figure 11a. We experimented with KAZE, ORB, Shi-Tomasi and Harris for feature extraction. Figure 12 shows the experimental results of MCPL-based pill detection and the colored lines represent similar features.
Figure 12. Experimental results of MCPL-based pill detection.
The experimental results comparing the combination of feature-extraction algorithms and the proposed system are shown in Table 1. In response to the lower accuracy in detecting complex shapes like hexagons and pentagons, we implemented the MCPL method, incorporating advanced feature-extraction algorithms (KAZE, ORB, Shi-Tomasi, Harris). This corresponds to a relative performance enhancement of approximately 1.2 times the base accuracy, providing a clearer and more concrete demonstration of the MCPL method’s effectiveness.
Table 1. Comparative experimental results of the combination of feature-extraction algorithm and proposed method.

6. Conclusions

In this study, we introduced a deep learning-based pill-classification system leveraging Multi-Combination Pattern Labeling (MCPL) to enhance the identification of complex-shaped pills such as hexagons and pentagons. While our approach improved classification accuracy significantly, achieving about 35% accuracy for these shapes, there is room for further enhancement. The MCPL system, integrating algorithms like KAZE, ORB, Shi-Tomasi, and Harris, proved effective in refining feature detection, thus improving the system’s accuracy and potentially reducing medication errors due to misclassification. Moreover, image quality variations such as blur and noise were recognized as crucial factors affecting performance, suggesting the need for robust preprocessing techniques to standardize input data and enhance the system’s reliability. This study not only pushed the boundaries of current pill-classification systems by employing advanced deep learning methods but also highlighted several areas for further investigation to optimize performance and expand usability.
In future research, we aim to integrate the latest advancements in deep learning architectures to enhance the classification accuracy of complex-shaped pills. We plan to experiment with newer object-detection models, such as improved versions of YOLO and EfficientDet, comparing these against our current MCPL-based system to quantify any performance improvements. Additionally, recognizing the critical impact of image quality on classification success, we will conduct a systematic study to evaluate how variations in image quality—such as blur and noise—affect the system’s performance. This will include testing advanced preprocessing techniques to standardize input data quality. To clarify and expand upon the dataset used, we will provide a detailed subsection in our study that describes the dataset’s publicly available or proprietary sources and analyze the distribution of different pill types better to understand their impact on our system’s effectiveness. Moreover, we will undertake an ablation study to distinctly showcase the contribution of each algorithmic component to the overall system efficacy. Finally, our method will be benchmarked against other state-of-the-art pill-classification systems to place our findings in a broader context and outline our system’s specific limitations and challenges, setting the stage for ongoing improvements in automated pill identification.

Author Contributions

Conceptualization, S.K. and S.-Y.I.; methodology, S.-Y.I.; software, J.-S.K.; validation, S.K., E.-Y.P. and S.-Y.I.; formal analysis, S.-Y.I.; resources, E.-Y.P.; data curation, J.-S.K. and S.-Y.I.; writing—original draft preparation, S.K.; writing—review and editing, S.-Y.I.; visualization, S.-Y.I.; supervision, S.-Y.I.; project administration, S.-Y.I.; funding acquisition, S.-Y.I. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. 2021R1C1C2011105). This work was supported by Innovative Human Resource Development for Local Intellectualization program through the Institute of Information & Communications Technology Planning & Evaluation(IITP) grant funded by the Korea government(MSIT)(IITP-2024-RS-2022-00156334).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Slawomirski, L.; Auraaen, A.; Klazinga, N. The Economics of Patient Safety: Strengthening a Value-Based Approach to Reducing Patient Harm at National Level; OECD Health Working Papers, no.96; OECD: Paris, France, 2017. [Google Scholar]
  2. Alrabadi, N.; Shawagfeh, S.; Haddad, R.; Mukattash, T.; Abuhammad, S.; Al-rabadi, D.; Farha, R.A.; AlRabadi, S.; Al-Faouri, I. Medication errors: A focus on nursing practice. J. Pharm. Health Serv. Res. 2021, 12, 78–86. [Google Scholar] [CrossRef]
  3. Isaacs, A.N.; Ch’ng, K.; Delhiwale, N.; Taylor, K.; Kent, B.; Raymond, A. Hospital medication errors: A cross-sectional study. Int. J. Qual. Health Care 2021, 33, mzaa136. [Google Scholar] [CrossRef] [PubMed]
  4. Naseralallah, L.; Stewart, D.; Price, M.; Paudyal, V. Prevalence, contributing factors, and interventions to reduce medication errors in outpatient and ambulatory settings: A systematic review. Int. J. Clin. Pharm. 2023, 45, 1359–1377. [Google Scholar] [CrossRef] [PubMed]
  5. Keers, R.N.; Williams, S.D.; Cooke, J.; Ashcroft, D.M. Causes of Medication Administration Errors in Hospitals: A Systematic Review of Quantitative and Qualitative Evidence. Drug Saf. 2013, 36, 1045–1067. [Google Scholar] [CrossRef]
  6. Tefera, Y.G.; Gebresillassie, B.M.; Ayele, A.A.; Belay, A.B.; Emiru, A.K. The characteristics of drug information inquiries in an Ethiopian university hospital: A two-year observational study. Sci. Rep. 2019, 9, 13835. [Google Scholar] [CrossRef]
  7. Almazrou, D.A.; Ali, S.; Alzhrani, J.A. Assessment of queries received by the drug information center at King Saud Medical City. J. Pharm. BioAllied Sci. 2017, 9, 246–250. [Google Scholar]
  8. Chantara, W.; Mun, J.H.; Shin, D.W.; Ho, Y.S. Object tracking using adaptive template matching. IEIE Trans. Smart Process. Comput. 2015, 4, 1–9. [Google Scholar] [CrossRef]
  9. Zhou, X.; Wang, Y.; Xiao, C.; Zhu, Q.; Lu, X.; Zhang, H.; Ge, J.; Zhao, H. Automated visual inspection of glass bottle bottom with saliency detection and template matching. IEEE Trans. Instrum. Meas. 2019, 68, 4253–4267. [Google Scholar] [CrossRef]
  10. Ushizima, D.; Carneiro, A.; Souza, M.; Medeiros, F. Investigating pill recognition methods for a new national library of medicine image dataset. In Proceedings of the ISVC 2015, Las Vegas, NV, USA, 14–16 December 2015; pp. 410–419. [Google Scholar]
  11. Ling, S.; Pastor, A.; Li, J.; Che, Z.; Wang, J.; Kim, J.; Callet, P.L. Few-Shot Pill Recognition. In Proceedings of the CVPR 2020, Seattle, WA, USA, 13–19 June 2020; pp. 9789–9798. [Google Scholar]
  12. Yaniv, Z.; Faruque, J.; Howe, S.; Dunn, K.; Sharlip, D.; Bond, A.; Perillan, P.; Bodenreider, O.; Ackerman, M.J.; Yoo, T.S. The national library of medicine pill image recognition challenge: An initial report. In Proceedings of the IEEE Applied Imagery Pattern Recognition Workshop, Washington, DC, USA, 18–20 October 2016; pp. 1–9. [Google Scholar]
  13. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  14. Sahraeian, S.M.E.; Liu, R.; Lau, B.; Podesta, K.; Mohiyuddin, M.; Lam, H.Y.K. Deep convolutional neural networks for accurate somatic mutation detection. Nat. Commun. 2019, 10, 1041. [Google Scholar] [CrossRef]
  15. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef]
  16. Kermany, D.S.; Goldbaum, M.; Cai, W.; Valentim, C.C.S.; Liang, H.; Baxter, S.L.; McKeown, A.; Yang, G.; Wu, X.; Yan, F.; et al. Identifying medical diagnoses and treatable diseases by image-based deep learning. Cell 2018, 172, 1122–1131. [Google Scholar] [CrossRef] [PubMed]
  17. Eraslan, G.; Avsec, Z.; Gagneur, J.; Theis, F.J. Deep learning: New computational modelling techniques for genomics. Nat. Rev. Genet. 2019, 20, 389–403. [Google Scholar] [CrossRef]
  18. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  19. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef]
  20. Girshick, R. Fast R-CNN. In Proceedings of the CVPR 2015, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  21. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  22. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the CVPR 2016, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  23. Delgado, N.L.; Usuyama, N.; Hall, A.K.; Hazen, R.J.; Ma, M.; Sahu, S.; Lundin, J. Fast and accurate medication identification. Npj Digit. Med. 2019, 2, 10. [Google Scholar] [CrossRef]
  24. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  25. Wong, Y.F.; Ng, H.T.; Leung, K.Y.; Chan, K.Y.; Chan, S.Y.; Loy, C.C. Development of fine-grained pill identification algorithm using deep convolutional network. J. Biomed. Inf. 2017, 74, 130–136. [Google Scholar] [CrossRef]
  26. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the CVPR 2015, Boston, MA, USA, 8–10 June 2015; pp. 1–9. [Google Scholar]
  27. Maji, S.; Bose, S. CBIR using features derived by deep learning. ACM/IMS Trans. Data Sci. 2021, 2, 1–24. [Google Scholar] [CrossRef]
  28. He, K.; Gkioxari, G.; Dollar, P.; Girshick, R. Mask R-CNN. In Proceedings of the ICCV 2017, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  29. Yao, J.; Qi, J.; Zhang, J.; Shao, H.; Yang, J.; Li, X. A real-time detection algorithm for kiwifruit defects based on YOLOv5. Electronics 2021, 10, 1711. [Google Scholar] [CrossRef]
  30. Zhao, J.; Zhang, X.; Yan, J.; Qiu, X.; Yao, X.; Tian, Y.; Zhu, Y.; Cao, W. A wheat spike detection method in UAV images based on improved YOLOv5. Remote Sens. 2021, 13, 3095. [Google Scholar] [CrossRef]
  31. Darma, I.W.A.S.; Suciati, N.; Siahaan, D. A Performance Comparison of Balinese Carving Motif Detection and Recognition using YOLOv5 and Mask R-CNN. In Proceedings of the ICICoS 2021, Semarang, Indonesia, 24–25 November 2021; pp. 52–57. [Google Scholar]
  32. Zhou, F.; Zhao, H.; Nie, Z. Safety helmet detection based on YOLOv5. In Proceedings of the ICPECA 2021, Shenyang, China, 22–24 January 2021; pp. 6–11. [Google Scholar]
  33. Wang, X.Y.; Wu, J.F.; Yang, H.Y. Robust image retrieval based on color histogram of local feature regions. Multimed. Tools Appl. 2010, 49, 323–345. [Google Scholar] [CrossRef]
  34. Lee, Y.B.; Park, U.; Jain, A.K.; Lee, S.W. Pill-ID: Matching and retrieval of drug pill images. Pattern Recognit. Lett. 2012, 33, 904–910. [Google Scholar] [CrossRef]
  35. Guo, P.; Stanley, R.J.; Cole, J.G.; Hagerty, J.; Stoecker, W.V. Color Feature-based Pillbox Image Color Recognition. In Proceedings of the VISIGRAPP 2017, Porto, Portugal, 27 February–1 March 2017; pp. 188–194. [Google Scholar]
  36. Alcantarilla, P.F.; Bartoli, A.; Davison, A.J. KAZE features. In Proceedings of the ECCV 2012, Florence, Italy, 7–13 October 2012; pp. 214–227. [Google Scholar]
  37. Bay, H.; Tuytelaars, T.; Gool, L.V. SURF: Speeded up Robust Features. In Proceedings of the ECCV 2006, Graz, Austria, 7–13 May 2006; pp. 404–417. [Google Scholar]
  38. Viswanathan, D.B. Features from accelerated segment test (fast). In Proceedings of the WIAMIS 2009, London, UK, 6–8 May 2009; pp. 6–8. [Google Scholar]
  39. Lowe, D.G. Distinctive image features from scale-invariant key points. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  40. Szeliski, R. Computer Vision: Algorithms and Applications Second Edition; Springer Nature: Berlin/Heidelberg, Germany, 2022. [Google Scholar]
  41. Harris, C.; Stephens, M. A combined corner and edge detector. In Proceedings of the Alvey Vision Conference 1988, Manchester, UK, 31 August–2 September 1988; pp. 147–151. [Google Scholar]
  42. Kwon, H.J.; Kim, H.G.; Lee, S.H. Pill Detection Model for Medicine Inspection Based on Deep Learning. Chemosensors 2022, 10, 4. [Google Scholar] [CrossRef]
  43. Holtkötter, J.; Amaral, R.; Almeida, R.; Jácome, C.; Cardoso, R.; Pereira, A.; Pereira, M.; Chon, K.H.; Fonseca, J.A. Development and Validation of a Digital Image Processing–Based Pill Detection Tool for an Oral Medication Self-Monitoring System. Sensors 2022, 22, 2958. [Google Scholar] [CrossRef]
  44. Nguyen, A.D.; Pham, H.H.; Trung, H.T.; Nguyen, Q.V.H.; Truong, T.N.; Nguyen, P.L. Highly Accurate and Explainable Multi-Pill Detection Framework with Graph Neural Network–Assisted Multimodal Data Fusion. arXiv 2023, arXiv:2303.09782. [Google Scholar] [CrossRef]
  45. Chang, W.J.; Chen, L.B.; Hsu, C.H.; Lin, C.P.; Yang, T.C. A Deep Learning–Based Intelligent Medicine Recognition System for Chronic Patients. IEEE Access 2019, 7, 44441–44458. [Google Scholar] [CrossRef]
  46. Ting, H.W.; Chung, S.L.; Chen, C.F.; Chiu, H.Y.; Hsieh, Y.W. A drug identification model developed using deep learning technologies: Experience of a medical center in Taiwan. BMC Health Serv. Res. 2019, 20, 1–9. [Google Scholar] [CrossRef]
  47. Nguyen, A.D.; Nguyen, T.D.; Pham, H.H.; Nguyen, T.H.; Nguyen, P.L. Image-based Contextual Pill Recognition with Medical Knowledge Graph Assistance. In Proceedings of the ACIIDS 2022, Ho Chi Minh City, Vietnam, 28–30 November 2022; pp. 354–369. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.