You are currently viewing a new version of our website. To view the old version click .
Journal of Imaging
  • Article
  • Open Access

29 March 2023

A Genetic Algorithm Based One Class Support Vector Machine Model for Arabic Skilled Forgery Signature Verification

,
,
and
1
Centre of Artificial Intelligence, Faculty of Information Sciences and Technology, University Kebangsaan Malaysia, Bangi 50300, Malaysia
2
Information Technology Center, Iraqi Commission for Computers and Informatics, Baghdad 10009, Iraq
3
Institute of Graduate Studies and Research, University of Alexandria, Alexandria 21526, Egypt
4
Information Technology Research and Development Center (ITRDC), University of Kufa, Najaf 540011, Iraq
This article belongs to the Topic Computer Vision and Image Processing

Abstract

Recently, signature verification systems have been widely adopted for verifying individuals based on their handwritten signatures, especially in forensic and commercial transactions. Generally, feature extraction and classification tremendously impact the accuracy of system authentication. Feature extraction is challenging for signature verification systems due to the diverse forms of signatures and sample circumstances. Current signature verification techniques demonstrate promising results in identifying genuine and forged signatures. However, the overall performance of skilled forgery detection remains rigid to deliver high contentment. Furthermore, most of the current signature verification techniques demand a large number of learning samples to increase verification accuracy. This is the primary disadvantage of using deep learning, as the figure of signature samples is mainly restricted to the functional application of the signature verification system. In addition, the system inputs are scanned signatures that comprise noisy pixels, a complicated background, blurriness, and contrast decay. The main challenge has been attaining a balance between noise and data loss, since some essential information is lost during preprocessing, probably influencing the subsequent stages of the system. This paper tackles the aforementioned issues by presenting four main steps: preprocessing, multifeature fusion, discriminant feature selection using a genetic algorithm based on one class support vector machine (OCSVM-GA), and a one-class learning strategy to address imbalanced signature data in the practical application of a signature verification system. The suggested method employs three databases of signatures: SID-Arabic handwritten signatures, CEDAR, and UTSIG. Experimental results depict that the proposed approach outperforms current systems in terms of false acceptance rate (FAR), false rejection rate (FRR), and equal error rate (EER).

1. Introduction

A signature is one of the most important human attributes. It is often used as proof of identity on legal documents like bank checks, credit cards, and wills. An effective automatic system can handle many fraud issues and other daily crimes. There are two different kinds of signature verification scenarios: online and offline. An online signature verification system uses tablets, PDAs, iPads, and smartphones to evaluate the signature image. The system has a dynamic nature, operating on features such as writing, orientation, pen tip positions, momentum, velocity, pressure, etc. [1,2].
An offline verification validates signatures by employing an optical detector to collect signatures on paper. This approach contains static data such as inclination, boundary, signature length and altitude, baseline, pressure, and size [3]. Offline verification is more intricate than online verification due to the absence of dynamic parameter information. Moreover, in an OSV system, signatures are obtained from various devices, so the resolution of training and testing samples is not the same, resulting in intraclass variation [4]. The raw signature may contain additional pixels known as noises or may not be in perfect working order, necessitating preprocessing.
Additionally, variations in the original signer’s signature are related to document orientation, signature degeneration, illness, illegible signatures, pen width, age, etc. As a result, preprocessing is a fundamental stage to improve the input data or isolate raw data samples into a standard format appropriate for the feature extraction stage. Nevertheless, it is crucial to balance noise removal and data loss, since a certain amount of relevant information may be lost during preprocessing, impairing the accuracy of subsequent system stages [5]. Therefore, proposing an algorithm that can effectively remove noise while maintaining relevant information is crucial.
Generally, the handwritten signature verification system determines whether the query signature is genuine or forged [2]. There are three forms of forgery: random forgery, simple (unskilled), and skilled forgery. In random forging, the forger uses a signature without knowing the original user’s name or signature. This signature shape is distinct from the genuine signature. The simple forger knows the user’s name, but is uninformed of the signature’s pattern. In this case, the forms of the original and forged signatures may not be similar.
On the other hand, the skilled forger learns the signature form and professionally mimics the signature with practice. This form of imitation is more challenging to detect since it is comparable to an authentic signature [6]. Figure 1 shows the samples of signatures. The most challenging aspect of the signature verification process is the substantial intra-class variance across signatures from the same individual, in contrast to low intra-class between forgery and genuine. Second, no comprehensive system can recognize all scripts.
Figure 1. Sample of signatures.
According to the literature, signature verification systems have been the subject of substantial study in several languages, including English, Hindi, Bangla, and Chinese. However, studies on Arabic signatures are still limited, and improvement is slow because the language is more difficult to analyze than others. Arabic script is characterized by its cursive form, changing letter shapes, joining and non-joining characters, delayed strokes, and ligatures [7,8].
Recently, researchers have focused on two major processes: feature extraction and verification methods. Feature extraction methods depended on handcraft, such as statistical, geometric, structural, and automatic deep learning. In several cases, researchers have combined multiple methods to improve performance. Fusion is the process of integrating multiple patterns of components into a single matrix using a fusion approach, such as score-level fusion and high-priority index feature fusion [9].
The problem with earlier approaches was that they typically employed a combination of features without considering correlations and discriminants, such that the generated feature vector could not recognize a skilled forgery. In addition, fusion techniques were used to increase system complexity and computational time. Moreover, the current deep learning approaches are defective in the signature verification domain because deep training models necessitate large data samples and effort in selecting images suitable for learning construction [8]. Deep learning also necessitates substantial computer resources, such as expensive GPUs.
This research aims to propose an offline Arabic signature verification system that can recognize skilled forgery and genuine signatures at a highly accurate rate and low FAR, FRR, and EER. First, efficient preprocessing techniques such as image cropping, denoising, binarization, and removing stray isolated pixels are recommended to decrease noise while maintaining essential data.
The contributions of this study include the following:
  • Efficient preprocessing techniques are recommended to decrease noise while maintaining essential data.
  • Hybrid feature types have been proposed to solve the low inter-class variability between authentic and skilled forgery and the high intra-class variability in each individual’s signature.
  • The early serial concatenation fusion approach (ESCF) integrates multiscale information without prejudice complication.
  • Propose GA_OCSVM to improve feature selection and tackle the potential correlation between fused features
  • Settle the problem of unbalanced and restricted forgery samples by using one-class classification.
The extensive computing time and storage capacity are unnecessary for the proposed approach.
This paper has the following organization: Section 2 presents the related studies. The framework for the proposed signature verification system is given in Section 3. The experimental results and comparisons of the proposed methodology to previous research are presented in Section 4. The summary of the proposed work is presented in Section 5. The conclusion and recommendations for further work are provided in Section 6.

3. Materials and Methods

The proposed model comprises five phases: preprocessing approaches, feature extraction, feature fusion, feature selection, and classification, as shown in Figure 2.
Figure 2. Flowchart of the proposed model.

3.1. Preprocessing Phase

A review of the problems related to offline signatures may include noisy pixels or equipment that may not be in perfect working order. As a result, several preprocessing methods are presented to provide an improved image that can be utilized for subsequent phases without losing data.

3.1.1. Image Conversion

The first step in the proposed approach is to convert an RGB image to a grayscale image, which is required to decrease system complexity and processing time because grayscale images are simpler to modify than RGB images.

3.1.2. Noise Reduction

The scanner or the paper backdrop might produce noise in a scanned image; the image may become fuzzy due to insufficient illumination and stained regions, such as dots and speckles. The image filtering technique improves the image by converting irrelevant brightness information into valuable data, and is easily understandable and concentrated on machine interactions. The median filtering (MF) strategy is utilized in this work to remove noise from the signature image. The MF technique [40] is a statistically based nonlinear method for reducing image noise. Applying a linear low-pass filter is the preferred approach for smoothing, which is the appropriate method in a static signature. The MF has the following two key advantages:
MF retains sharp edges, whereas low-pass linear filtering softens the edges.
MF is quite effective in smoothing down a noise spike.
MF retains the pertinent information of the image and changes the original grey value of each pixel to the median gray value of the area of the neighborhood. This filter reduces visual noise without causing edge blurring. The median is calculated by sorting the pixel values of the neighborhood window and substituting the considered pixel with the middle (median) value. The formula for the MF image D ( x , y ) of image I ( m , n ) is represented as (1).
D x , y = m e d i a n ( m , n ) R x y I m , n
where m , n center around the processed pixel ( x , y ).

3.1.3. Binarization

This process transforms a grayscale image into a binary image. Image binarization is the earliest step of image processing and analysis. Pixels in an image are separated into two different areas, black and white. The main goal of image binarization is to be able to describe the difference between text in the foreground and text in the background. The thresholding technique is the simplest type of binarization. In thresholding, pixels are identified as foreground or background by comparing them to the maximum threshold value. However, determining the optimal threshold value for such signature text is challenging. Inaccurate estimation of the threshold value leads to the erroneous classification of pixels as foreground or background, which affects binarization results and the accuracy of signature authentication.
In this research, the backdrop of an image is estimated using the grayscale morphological method [41]. The contrast of the image text regions is boosted using the approximate background data. A recognition threshold value for image sections is determined by analyzing the histogram of the contrast image. In image processing, morphology can be applied to two types of pixel sets: objects and structural elements (SEs). Objects are described as collections of foreground pixels. SEs are created using both foreground and background pixels.
The size of the SEs is first determined and calculated using the histogram of the distance between consecutive edges. The morphological processing techniques include dilation and erosion. SEs generate both dilation and erosion by interacting with a collection of exciting pixels in an image. The SEs have a morphology and an origin. A B denotes dilation, which is the collection of all shifts satisfying the condition in Equation (2):
A B = z B ^ z A
where A is a set of foreground pixels, B is SEs, B ^ is the reflection of the structuring B about its origin, followed by a shift by z, and z’s are foreground values (one’s). The erosion represented by the symbol A B is defined as Equation (3):
A B = z B z A

3.1.4. Image Segmentation

Segmentation is used to extract the signature region from an image. This procedure decreases the processing time by deleting the excess pixels of the image. In this work, an automated segmentation technique is computed using the histogram; the region is automatically segmented based on pixel values. The histogram depicts the total amount of black-and-white pixels [42]. The distribution is horizontally and vertically separated. The white pixel’s highest point is utilized as a trimming reference. The image is divided in half to simplify the recovery of the starting and ending points.
Consequently, two points denote the beginning and end of cutting originating from the highest point. The horizontal histogram determines the starting and ending positions of the horizontal trim. Additionally, vertical cutting uses the vertical histogram to calculate the beginning and ending locations. Equations (4)–(6) generate histograms:
H X = y = 1 n b i m x , y
H Y = x = 1 m b i m x , y
i m c r o p = o r i g i n i m a g e X m a x   <   X   <   X min     a n d   Y m a x   <   Y   <   Y m i n 0 o t h e r w i s e
b i m denotes the binary image, while m and n indicate the matrix b i m s rows and columns, respectively.

3.1.5. Stray Isolated Pixel Elimination

In some signatures, extra points caused by ink flowing unrelated to the signature may affect the original signature area. Consequently, the MATLAB function eliminates any connected components (objects) with less than 50 pixels from the binary image B ( x , y ) . This procedure is known as an area opening, as shown in Equation (7).
B n e w x , y = bwareaopen   ( B x , y

3.1.6. Skeletonization and Thinning

Thinning is an iterative process that results in skeleton production. This procedure minimizes the number of character characteristics to aid feature extraction and classification by erasing the width fluctuations of the pen. Applying a specific morphological operation to the binary image B, a fast parallel thinning method (FPT) removes inside pixels to leave an outline of the signature [43]. The FPT approach extracts the skeleton from an image by removing all contour points except those relevant to the skeleton. As illustrated in Figure 3, each point p ( i , j ) has eight neighbors.
Figure 3. Designations of the pixels in a 3 × 3 window.
Each iteration is separated into two subiterations to preserve the structure of the skeleton. See Algorithm 1 and Figure 4.
Algorithm 1: FPT
1: A (P1) is the number of (01) patterns in the ordered set P2, P3, P4, …., P8, P9 that are the eight neighbors of P1
2: B (P1) is the number of nonzero neighbors of P1
3 : B ( P 1 ) = i = 2 9 P i
4: Iteration 1: P1 = 0
                If   2 B (P1) ≤ 6
                If A (P1) = 1
                    If   P 2 × P 4 × P 6 = 0
                   If   P 4 × P 6 × P 8 = 0
                             Else
                                  P1 = 1
                                     A (P1) = 2
5: Iteration 2:
                           P2 × P6 × P8 = 0, P2 × P4 × P8 = 0
                   Keep the rest points
6: End
Figure 4. Counting the 01 patterns in the ordered set P2,…, P9.
Skeletonization is achieved by removing specific foreground pixels from a binary image through image thinning. Consequently, a collection of tiny arcs and curves portrays the signature pattern. The complement of the image is adjusted to make the signature bright and the background dark to skeletonize the original image. See Figure 5 and Figure 6.
Figure 5. Locations of points that satisfy the conditions.
Figure 6. Preprocessing steps (a) grayscale image, (b) image denoising, (c) segmentation, (d) isolated removal, and (e) thinning and skeletonization.

3.2. Hybrid Feature Extraction

Feature extraction is a crucial step in the verification of a signature. The proposed Hybrid Statistical Feature Extraction (HSFE) technique extracts highly informative features by combining multiple types of features using three statistical approaches: interest point features, global and local texture features, and curvelet transformation features.

3.2.1. Texture Feature

In image processing, the texture is described as a function of the spatial variation of the brightness intensity of the pixel. Image processing is the primary term to define objects or concepts in a given image. Texture analysis is critical in computer vision applications such as object recognition, surface defect detection, pattern recognition, and medical image analysis. This paper combines two statistical methods of edge direction matrices (EDMs) [44] and local binary pattern (LBP) [45] to extract texture attributes.
LBP features are also known as the texture operator for a grayscale image, which helps to characterize the spatial structure of the input image texture. Once the central pixel value is obtained, the pattern code can be computed by comparing these values to its neighborhoods. It can be expressed as Equation (1).
L B P N , R = n = 0 N 1 s ( I n I g ) 2 N
s ( x ) = 1 , x 0 0 , x < 0
where I g denotes the gray value of the center pixel, I n represents the gray values of the circularly symmetrical neighborhood, and N denotes the total number of spaced pixels on a circle of radius R. The final texture feature employed in texture analysis is the histogram of the operator outputs (i.e., pattern labels) accumulated over a texture sample. The operator for grayscale and rotation-invariance texture description is shown in Equations (9) and (10).
L B P N , R = n = 0 N 1 s I n I g i f U L P B N , R 2 N + 1 o t h e r w i s e
where
U L P B N , R = | s I N 1 I g s I 0 I g | + n = 1 N 1 | s I n I g s I n 1 I g |
However, LBP cannot provide information about shape; that is, the spatial relationships of pixels in an image. As a result, LBP is combined with EDMS. The global features are the features that result from the shape of a signature contour [45]. EDMS is a feature extraction approach that detects the texture of a binary image I ( x , y ) based on edge-to-neighbor pixel relationships. Eight adjoining kernel matrices were applied, and each pixel was linked to two neighboring pixels. A connection was established between the edge pixel E ( x , y ) and its neighboring pixels, as illustrated in Figure 7a. The eight pixels were used to change the surrounding values into the position values, as shown in Figure 7b.
Figure 7. (a) Two neighboring edge pixels, (b) EDMs principal.
This approach is presented from two perspectives: first-order relationship (FOR) identification and second-order relationship (SOR) identification. Each cell in the FOR matrix has a location between 0 and 315 degrees, depending on the pixel neighborhood association. The relationship between the pixel values can be determined by computing the occurrence of the FOR values while considering the edge image of each pixel concerning two other pixels.
The relationships are sorted according to their priority by ordering the values in FOR in descending order. Subsequently, the highest-order relationships are selected, and the others are disregarded. The acquired relationships are computed and saved in the SOR cell. Algorithms 2 and 3 provide critical statistical features, including data attributes and distribution descriptions.
Algorithm 2: FOR
1:  for each pixel in (E (x,y))
2: If p (x,y) = 0    {Black pixel at center}    Then
Increase the frequency of occurrences at FOR (2,2) by 1
3: If p (x + 1) = 0              {Black pixel at 0°}          Then,
Increase the frequency of occurrences at FOR (2,3) by 1
4:    If p (x + 1, y − 1) = 0    {Black pixel at 45°}        Then
Increase the frequency of occurrences at FOR (3,1) by 1
5:        If p ((x, y − 1)) = 0                {Black pixel at 90°} Then
Increase the frequency of occurrences at FOR (2,1) by 1
6:        If p ((x − 1, y − 1)) = 0            {Black pixel at 135°} Then
Increase the frequency of occurrences at FOR (1,1) by 1
7:     If p (x, y − 1) = 0            {Black pixel at 180°} Then
Increase the frequency of occurrences at FOR (2,3) by 1
8: If p (x − 1, y + 1) = 0          {Black pixel at 225°}  Then
Increase the frequency of occurrences at FOR (3,1) by 1
9:    If p (x, y + 1) = 0            {Black pixel at 270°}  Then
Increase the frequency of occurrences at FOR (2,1) by 1
10:    If p (x + 1, y + 1) = 0            {Black pixel at 315°}  Then
                         Increase the frequency of occurrences at FOR (1,1) by 1
      End
Algorithm 3: SOR
1:   Sort R1 = FOR (x, y)↓
2:   For each pixel in (E (x, y))
3: If E (x, y) = Black                  Then
                R2 = Relationships of neighborhood two pixels in E (x, y))
4:      Compare (R1, R2)
5:                 Connected cell in SOR = SOR + 1,
End

3.2.2. Interest Point Features

This work uses the speeded up robust feature (SURF) to identify an image’s interesting points. SURF is a resilient representation approach invariant to translation, rotation, and scaling. This descriptor is used to find the similarity between different interesting points. The entry of an integral image ( x , y ) at a location (x,y)T is used to represent the sum of all pixels in the input image I ( x , y ) within a rectangular region formed by the origin and (x,y)T. See Equation (11).
( x , y ) = i = 0 i x j = 0 j y I i , j
Additionally, the Hessian matrix is used to identify blob-like formations at regions where the determinant is optimal. The Hessian matrix H (p, σ) at point p = (x,y)T and scale σ = 3 is shown as Equation (12):
H p , σ = L x x p , σ L x y p , σ L x y p , σ L y y p , σ
where Lxx (p, σ) is the convolution of the second-order derivative of the Gaussian g ( x , σ ) x with image ( x , y ) at point p and is similar to L x y (p, σ) and Lyy (p, σ).

3.2.3. Curvelet Transformation (CT)

CT is a multiscale pyramid with several orientations and placements at each length and is needle-shaped at a small scale. CT was produced in recent years to address the inherent limits of conventional multiscale representations, which describe curve-like edges with a limited number of coefficients compared to wavelets far from optimal [46]. The CT technique captures the curved edge of characters in an Arabic script. The CT is mathematically described as Equations (13) and (14):
CT = R 0 f x W a , b , θ x d x
W a , b , θ x = W a R θ x b
where W a , b , θ x are the wavelet coefficients; a is the number of levels in the wavelet pyramid ( a = 4); b = [3 4 4 5] represents location scalar parameters; θ is an orientation parameter θ 0 , 2 π ; R θ = cos θ sin θ sin θ cos θ and is the rotation matrix with angle θ.

3.3. Feature Fusion

The precision of signature classification can be improved by extracting appropriate features. A method for fusing hybrid features is proposed to solve the restriction of a single feature extraction technique, as shown in Figure 8. Feature fusion combines several feature vectors to generate the final feature vector, which involves complementing each other’s advantages to obtain more robust and accurate outcomes [47]. The ESCF technique converts the feature matrix into a feature vector that describes the signature and can reduce error rates. ESCF is simple to implement, does not cause the loss of information, and has no impact on computational efficiency.
Figure 8. A strategy of feature fusion.
Let A and B be two feature spaces specified on pattern space Ω. For an arbitrary sample ξ ∈ Ω, the associated feature vectors are α ∈ A with n-dimensional features and β ∈ B with m-dimensional features; the Serial Fused feature of ξ is defined as γ = α β with dimension ( m + n ). The mathematical description of the fusing process is based on Equation (15).
F ( v ) = C T 1 × N E D M 1 × N S U R F 1 × N L B P 1 x N
where F(v) is the final fused vector of 1 × s u m ( N ) dimensions for all samples.

3.4. Feature Selection

Feature selection has been a productive area of research in intelligent algorithms and machine learning, which is undoubtedly essential. Feature selection eliminates attributes that may negatively affect the performance of classifiers, such as irrelevant, redundant, or less informative features. As indicated in the preceding section, a simple concatenated fusion technique combines various statistical features to generate an additional dimension that can identify skilled forgeries and genuine signatures with high accuracy.
The problem with combining characteristics without considering correlation and discrimination is that the resulting feature vector cannot detect a skilled forgery. Furthermore, fused features from multiple approaches may provide high-dimension features that could influence the verification process. As a result, a feature selection approach is necessary to minimize the number of features and remove data correlations.
GA has achieved success in many applications. GA can handle more complicated problems than neural networks and specializes in identifying an appropriate feature for a given class. However, automating the design of such fitness functions is still an open challenge. Adopting simple and effective fitness functions is a critical issue for GA.
In this work, GA is used with one class support vector machine (OC-SVM) classifier to discover the genes with the highest predictive performance. Meanwhile, OC-SVM is employed for the classification. This proposal is one of the valuable contributions to reducing the issue of complexity and extending search spaces. Figure 9 shows the flowchart of the GA-OCSVM.
Figure 9. Flowchart of feature selection.
The procedure starts by randomly generating an initial population.
  • The initial population size is created and set to 10.
  • Calculate and assign a score of the fitness value to each member of the current population. These values are regarded as the raw fitness scores. The fitness function of each individual is determined by evaluating the OC-SVM using a training set. As a result, the fitness function containing classification precision is utilized in this study, as described in Equation (16).
    f i t t n e s s f = M a x ( a c c u r a c y f )
    where accuracy ( f ) is the accuracy of the classifier for the subset selection of features expressed by f .
  • Select members, known as parents, according to their expectations. Some individuals in the present population with maximum fitness levels are selected as elite (the subset with the best classification precision). These elite members are transmitted to the following population.
  • Generates offspring from the selected parents. Offspring are produced by combining the vector entries of two parents (crossover). A uniform crossover with a crossover rate of 0.8 is employed.
  • Low-frequency offspring introduce variety into a single-parent population (mutation). A uniform mutation method is selected with a mutation rate of 0.2.
The roulette wheel method is applied to randomly cross and mutate the Chromosome, which keeps the selective pressure in the center rather than at the extremes. In roulette wheel selection, the circular wheel is divided into n pies, where n equals the number of individuals. Based on their fitness level, each individual receives an appropriate circular piece. The wheel is rotated at a defined circumferential point. The region of the wheel just forward of the setpoint is referred to as the parent. The same method is followed for the next generations. The probability Pi of the individual is defined as Equation (17).
P i = f i S
where S = i = 1 n f i , n is the size of the population, and f i is the fitness of individual i .
  • Individuals with higher fitness are more likely to be selected for reproduction.
The 110 features are selected as optimal features, a collection of discriminant characteristics. These distinguishing characteristics are fed into the classifier for verification.

3.5. One-Class Classification

One-class classification (OCC) is used to solve the issue of an imbalanced database of signatures in the real world, where the authentic signature is only generated, and the forged signature is absent, which indicates that the original signer is incapable of forging the signature. This research employs OC-SVM to address these issues. The OC-SVM can successfully deal with the positive samples in the training set. The appropriate distance used by the radial basis function kernel (RBF kernel) must be specified to train the OC-SVM. OC-SVM is developed in two phases. First, one-class information (normal class) trains a classifier to distinguish genuine instances. The classifier rejects the samples belonging to unknown classes and classifies them as forgeries, as shown in Figure 10. A hypersphere with the shortest radius is constructed around the positive class data, which encloses approximately every point in the dataset. According to the parameter RBF γ, the hypersphere is defined by Equation (18).
K x i , x = e γ d x , x i
where d x , x i is the distance between the original images x and the target samples xi (or positive), and γ = 0.07 is the deviation parameter of the kernel function. The OC-SVM decision function is shown in Equation (19).
f x = s i g n ( i = 1 N α i K x i , x ρ ) ,
0 α i 1 v
where N is the number of training instances, ρ is the distance of the hypersphere from the origin, α i denotes the Lagrange multiplier for each distance, and v = 0.01 represents the trade-off between maximizing the data points (encompassed by the hypersphere) and reducing the hypersphere’s distance from the origin. If the decision value of the sample is more significant than zero, we conclude that the target is a positive class; otherwise, it is a negative class.
Figure 10. One-class support vector machine.

4. Experimental Results and Analysis

The suggested model was evaluated on three databases: SID Arabic signatures, CEDAR, and UTSig [48,49,50]. The model was constructed in MATLAB®2021a on an Intel® CoreTM i5-8300H CPU @2.30 GHz.

4.1. Experiments and Evaluation of Preprocessing

The proposed preprocessing strategies’ performance is appraised using image quality performance metrics such as mean square error (MSE) and peak signal-to-noise ratio (PSNR) in Equations (20) and (21). The assessment aims to illustrate the efficiency of image enhancement procedures on captured signature images.
The PSNR ratio is a quality measurement between the original image I and the enhanced image j. The high value of PSNR demonstrates the significant quality of the processed image. However, as mentioned above, more than this ratio is needed in the offline system because critical data may be lost during preprocessing. As a result, MSE is used to verify the image quality without losing key features. If the value of MSE is close to zero, the image quality is accepted. Conversely, the image loses its main attributes.
PSNR ( I , J ) = 10 log 10 R 2 MSE ( I , J )
MSE ( I , J ) = 1 N * M a = 1 N b = 1 m ( I ( a , b ) J ( a , b ) ) 2
M*N signifies the image size, and I(a, b) and J(a, b) denote the original and processed image pixel intensities, respectively. R is the maximum allowable pixel value. Six experiments were performed to show the impact and significance of each preprocessing method on the image, as shown in Table 2, Table 3 and Table 4.
Table 2. Results of preprocessing on the SID database.
Table 3. Results of preprocessing on the UTSIG database.
Table 4. Results of preprocessing on the CEDAR database.
Based on the MSE and PSNR results. The findings demonstrate that after applying the proposed preprocessing, the output images maintain the exact representation of the original images but with fewer pixels. It can be evidenced that the proposed preprocessing methods achieve an equilibrium between noise reduction and image data preservation.

4.2. Experiments and Evaluation of Verification

In this step, the performance of the proposed model is comprehensively evaluated. The verification is evaluated in two experiments: (a) features selection and extracted features without combining preprocessing; and (b) integration preprocessing with feature extraction and selection. FAR, FRR, and EER measurements were used to assess the model. FAR and FRR are two types of error measurements used to evaluate the performance of biometric systems. FRR is the percentage of authorized users whose request is incorrectly rejected. FAR is the percentage of unauthorized users who are mistakenly accepted, and EER specifies the point where the FRR and FAR are equal and stated as flowing equations.
FAR = n o .   o f   f o r g e d   s i g n a t u r e s   c l a s s i f i e d   a s   g e n u i n e T o t a l   n o :   f o r g e d   s i g n a t u r e s × 100
FRR = n o .   o f g e n u i n e   s i g n a t u r e s   c l a s s i f i e d a s   f o r g e d T o t a l   n o :   g e n u i n   s i g n a t u r e s × 100
EER = FRR + FAR 2
The proposed model was only trained on genuine samples with no forgeries to simulate the verification system in the real world, as shown in Table 5. The ratio validation of the model is 0.2. with random selected.
Table 5. Division of database samples for modeling assessment.
  • In the first experiment, the model performance was examined without preprocessing steps. This experiment aims to show the impact of preprocessing on verification accuracy. As shown in Table 6, the unsatisfactory verification results confirm that each preprocessing step significantly enhances image quality; this is what the second experiment proved.
    Table 6. Results of verification without preprocessing.
  • The second experiment included all stages of the proposed model, including preprocessing, hybrid feature extraction, feature fusion, feature selection, and verification. The training phase was separately performed using three sets of genuine (G) samples. Table 7 displays the verification results on the SID Arabic database.
    Table 7. Results of the proposed model on the SID database.
The suggested approach was also evaluated on CEDAR (English signature) and UTSIG (Farsi signature) to demonstrate its comprehensive performance. The model attained superb results on these databases, as shown in Table 8 and Table 9.
Table 8. Results of the proposed model on the CEDAR database.
Table 9. Results of the proposed model on the UTSIG database.
To compare the proposed algorithm results with existing signature verification techniques, Table 10, Table 11 and Table 12 show the results of the proposed model with state-of-the-art methods. From the results, it is clear that the performance of the presented model is good in terms of FAR and FRR using one-class training. The average error rate has been reduced to 10% in the SID database, 2–10% in the UTSIG database, and 3–7% in the CEDAR database.
Table 10. Comparison of the results of the proposed approach and the state-of-the-art methods using the SID database.
Table 11. Comparison of the results of the proposed approach and the state-of-the-art methods using the UTSIG database.
Table 12. Comparison of the results of the proposed approach and the state-of-the-art methods using the CEDAR database.

4.3. Discussion

The proposed approach obtained superior FAR, FRR, and ERR on the three databases, particularly for skilled forgeries, which is the essential contribution of this study. Each stage of the model contributed to the increased precision. The preprocessing steps enhanced the verification results because all uninformative data and noise were removed. Moreover, the verification system’s supremacy is due to fused hybrid features and discrimination feature selection. The proposed feature extraction is advantageous because it combines multiple features to address the low intraclass difference between skilled forgery and genuine signatures, and the high intraclass difference between original signatures to the same writer. This combination maximizes the merit of each approach by complementing the advantages of other techniques, hence improving verification capabilities.
EDMs is a global textural descriptor used to analyze the entire image. Although EDMs were adequate for simple forgeries, they could not achieve high precision for skilled forgeries. For skilled forgers, LBP was more effective and accurate than EDMs. LBP is a local texture descriptor that describes a small part of the image and extracts more information. Some images may have local details that LBP could not figure out. In order to increase the detection percentage of skilled forgery, the SURF descriptor was used to add more distinct local features. SURF can detect and describe the interesting feature of the image. The key points in the picture include characteristics such as corners, edges, spots, and so on. The consistency of the key points can be helpful for performance. SURF outperforms SIFT in terms of performance and computational complexity.
Furthermore, the curvy lines in the Arabic characters were captured using a CT; CT accurately represents curved discontinuities. In addition, the feature selection strategy plays a crucial role in improving accuracy by removing insignificant characteristics. It also tackled the problem of correlation that may result from the feature fusion process, as illustrated in Figure 11.
Figure 11. The accuracy in the plot curve.
Overall, the findings of the proposed methodology prove that the proposed method performs considerably better than recent signature verification approaches.
The complexity of the proposed model was assessed in this work using computing time. Each signature was processed in 0.01063 s, 0.01982 s, and 0.01544 s for the SID, CEDAR, and UTSIG databases.

5. Summary of the Scientific Work

The Arabic OSV system was presented using six stages: preprocessing, hybrid feature extraction and fusion, GA-OCSVM-based optimum feature selection, and OCC. This research suggests a multi-step process for preprocessing images, starting with image binarization, and moving on to denoising, segmenting, isolating, thinning the signature, and skeletonizing. Experiments yielded efficient results with high PSNR and low MSE. The suggested approaches significantly impacted verification processing time and accuracy.
Though the proposed method elicits and fuses four different statistical techniques, the ESCF fusion strategy has remained feasible regarding complexity. The best features were then selected using GA.
Additionally, OCC was employed to address the need for forgery signature samples in practical applications in the real world. The proposed model was implemented using the three databases mentioned earlier.

6. Conclusions

This paper proposed a signature verification model with four primary phases: preprocessing and hybrid feature extraction, followed by feature fusion. Finally, features selection and verification. The algorithm’s output was constructed with genuine and forged signature samples from three standard databases: SID-Signatures, CEDAR, and UTSIG. The suggested approaches significantly impacted the verification processing time and accuracy.
The proposed method combined four different statistical techniques. The best GA-based features were then identified for classification. Additionally, the proposed model employed OC-SVM to address the restriction of the current Arabic OVS regarding skilled forgery. The results revealed that the proposed system outperformed existing techniques. It improved the FAR by 10% on the SID-Arabic signature database without increasing the computation time. The experiment yielded 0.037 FRR, 0.039 FAR_skilled, 0.063 FAR simple, and 0.044 EER.
Moreover, the model was superior in enhancing the EER values of UTSig and CEDAR databases, which achieved 0.074 and 0.048, respectively. The FRR value could be enhanced by adding structural features in the future. Additionally, the accuracy of feature selection can be strengthened by improving crossover and mutation.

Author Contributions

A.A.A., M.F.N. and S.M.D.; methodology, A.A.A.; software, A.A.A.; validation, A.A.A., M.F.N. and Z.A.A.A.; formal analysis, A.A.A., M.F.N., S.M.D. and Z.A.A.A.; investigation, A.A.A. and M.F.N.; resources, A.A.A.; writing—original draft preparation, A.A.A., M.F.N. and S.M.D.; writing—review and editing, M.F.N., S.M.D. and Z.A.A.A.; visualization, Z.A.A.A., M.F.N. and S.M.D.; supervision, A.A.A., M.F.N., S.M.D. and Z.A.A.A.; project administration, M.F.N.; funding acquisition. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Universiti Kebangsaan Malaysia (UKM), grant number GUP-2020-065.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors thank the CAIT research group at the Faculty of Information Science and Technology, University Kebangsaan Malaysia.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Taherzadeh, G.; Karimi, R.; Ghobadi, A.; Beh, H.M. Evaluation of Online Signature Verification Features. In Proceedings of the 13th International Conference on Advanced Communication Technology (ICACT2011), Gangwon-Do, Republic of Korea, 13–16 February 2011; pp. 772–777. [Google Scholar]
  2. Ahmed, K.; El-Henawy, I.M.; Rashad, M.Z.; Nomir, O. On-Line Signature Verification Based on PCA Feature Reduction and Statistical Analysis. In Proceedings of the 2010 International Conference on Computer Engineering & Systems, Cairo, Egypt, 30 November–2 December 2010; pp. 3–8. [Google Scholar]
  3. Pal, S.; Pal, U.; Blumenstein, M. Off-Line English and Chinese Signature Identification Using Foreground and Background Features. In Proceedings of the 2012 International Joint Conference on Neural Networks (IJCNN), Brisbane, Australia, 10–15 June 2012; pp. 1–7. [Google Scholar]
  4. Diaz, M.; Ferrer, M.A.; Impedovo, D.; Malik, M.I.; Pirlo, G.; Plamondon, R. A Perspective Analysis of Handwritten Signature Technology. ACM Comput. Surv. 2019, 51, 1–39. [Google Scholar] [CrossRef]
  5. Poddar, J.; Parikh, V.; Bharti, S.K. Offline Signature Recognition and Forgery Detection Using Deep Learning. Procedia Comput. Sci. 2020, 170, 610–617. [Google Scholar] [CrossRef]
  6. Al-Omari, Y.M.; Abdullah, S.N.H.S.; Omar, K. State-of-the-Art in Offline Signature Verification System. In Proceedings of the 2011 International Conference on Pattern Analysis and Intelligence Robotics, Kuala Lumpur, Malaysia, 28–29 June 2011; Volume 1, pp. 59–64. [Google Scholar]
  7. Haddad, A.; Riadi, L. Online Automatic Arabic Handwritten Signature and Manuscript Recognition. In Proceedings of the 2017 International Conference on Engineering & MIS (ICEMIS), Monastir, Tunisia, 8–10 May 2017; pp. 1–7. [Google Scholar]
  8. Wilson-Nunn, D.; Lyons, T.; Papavasiliou, A.; Ni, H. A Path Signature Approach to Online Arabic Handwriting Recognition. In Proceedings of the 2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR), London, UK, 12–14 March 2018; pp. 135–139. [Google Scholar]
  9. Majid, A.; Khan, M.A.; Yasmin, M.; Rehman, A.; Yousafzai, A.; Tariq, U. Classification of Stomach Infections: A Paradigm of Convolutional Neural Network along with Classical Features Fusion and Selection. Microsc. Res. Tech. 2020, 83, 562–576. [Google Scholar] [CrossRef] [PubMed]
  10. Smejkal, V.; Kodl, J. Dynamic Biometric Signature—An Effective Alternative for Electronic Authentication. Adv. Technol. Innov. 2018, 3, 166–178. [Google Scholar]
  11. Levy, A.; Nassi, B.; Elovici, Y.; Shmueli, E. Handwritten Signature Verification Using Wrist-Worn Devices. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018, 2, 1–26. [Google Scholar] [CrossRef]
  12. Rehman, A.; Naz, S.; Razzak, M.I. Writer Identification Using Machine Learning Approaches: A Comprehensive Review. Multimed. Tools Appl. 2019, 78, 10889–10931. [Google Scholar] [CrossRef]
  13. Morocho, D.; Morales, A.; Fierrez, J.; Vera-Rodriguez, R. Towards Human-Assisted Signature Recognition: Improving Biometric Systems through Attribute-Based Recognition. In Proceedings of the 2016 IEEE International Conference on Identity, Security and Behavior Analysis (ISBA), Sendai, Japan, 29 February–2 March 2016; pp. 1–6. [Google Scholar]
  14. Stauffer, M.; Maergner, P.; Fischer, A.; Ingold, R.; Riesen, K. Offline Signature Verification Using Structural Dynamic Time Warping. In Proceedings of the 2019 International Conference on Document Analysis and Recognition (ICDAR), Sydney, Australia, 20–25 September 2019; pp. 1117–1124. [Google Scholar] [CrossRef]
  15. Basu, A.; Shioya, H.; Park, C. Statistical Inference: The Minimum Distance Approach; CRC: Boca Raton, FL, USA, 2019; ISBN 1420099663. [Google Scholar]
  16. Kaur, H.; Kansa, E.R. Distance Based Online Signature Verification with Enhanced Security. Int. J. Eng. Dev. Res. 2017, 5, 1703–1710. [Google Scholar]
  17. Al-juboori, S.S. Signature Verification Based on Moments Technique. Ibn AL-Haitham J. Pure Appl. Sci. 2013, 26, 385–395. [Google Scholar]
  18. Suthaharan, S. Support Vector Machine. In Machine Learning Models and Algorithms for Big Data Classification; Springer: Boston, MA, USA, 2016; pp. 207–235. [Google Scholar]
  19. Kumar, A.; Bhatia, K. Offline Handwritten Signature Verification Using Decision Tree; Springer: Singapore, 2023; pp. 305–313. [Google Scholar]
  20. Chu, J.; Zhang, W.; Zheng, Y. Signature Veri Cation by Multi-Size Assembled-Attention with the Backbone of Swin-Transformer. Neural Process. Lett. 2023. preprint. [Google Scholar] [CrossRef]
  21. Kumari, K.; Rana, S. Towards Improving Offline Signature Verification Based Authentication Using Machine Learning Classifiers. Int. J. Innov. Technol. Explor. Eng. 2019, 8, 3393–3401. [Google Scholar] [CrossRef]
  22. Hafemann, L.G.; Sabourin, R.; Oliveira, L.S. Analyzing Features Learned for Offline Signature Verification Using Deep CNNs. arXiv 2016, arXiv:1607.04573. [Google Scholar]
  23. Hafemann, L.G.; Sabourin, R.; Oliveira, L.S. Writer-Independent Feature Learning for Offline Signature Verification Using Deep Convolutional Neural Networks. In Proceedings of the 2016 International Joint Conference on Neural Networks (IJCNN 2016), Vancouver, BC, Canada, 24–29 July 2016; pp. 2576–2583. [Google Scholar]
  24. Chugh, A.; Jain, C. Kohonen Networks for Offline Signature Verification. Comput. Sci. 2017, 4, 18–23. [Google Scholar]
  25. Akhundjanov, U.Y.; Starovoitov, V.V. Static Signature Verification Based on Machine Learning; Big Data and Advanced Analytics: Minsk, Belarus, 2022; pp. 11–12. [Google Scholar]
  26. Singh, A.; Viriri, S. Online Signature Verification Using Deep Descriptors. In Proceedings of the 2020 Conference on Information Communications Technology and Society (ICTAS), Durban, South Africa, 11–12 March 2020; pp. 1–6. [Google Scholar]
  27. Stauffer, M.; Maergner, P.; Fischer, A.; Riesen, K. A Survey of State of the Art Methods Employed in the Offline Signature Verification Process. In New Trends in Business Information Systems and Technology; Dornberger, R., Ed.; Springer: Berlin/Heidelberg, Germany, 2021; Volume 294, pp. 17–30. ISBN 9783030483319. [Google Scholar]
  28. Stauffer, M.; Maergner, P.; Fischer, A.; Riesen, K. Graph Embedding for Offline Handwritten Signature Verification. In Proceedings of the 2019 3rd International Conference on Biometric Engineering and Applications, Stockholm, Sweden, 29–31 May 2019; pp. 69–76. [Google Scholar]
  29. Zulkarnain, Z.; Mohd Rahim, M.S.; Othman, N.Z.S. Feature Selection Method for Offline Signature Verification. J. Technol. 2015, 75, 79–84. [Google Scholar] [CrossRef]
  30. Guru, D.S.; Manjunatha, K.S.; Manjunath, S.; Somashekara, M.T. Interval Valued Symbolic Representation of Writer Dependent Features for Online Signature Verification. Expert Syst. Appl. 2017, 80, 232–243. [Google Scholar] [CrossRef]
  31. Kar, B.; Mukherjee, A.; Dutta, P.K. Stroke Point Warping-Based Reference Selection and Verification of Online Signature. IEEE Trans. Instrum. Meas. 2017, 67, 2–11. [Google Scholar] [CrossRef]
  32. Sekhar, V.C.; Mukherjee, P.; Guru, D.S.; Pulabaigari, V. Online Signature Verification Based on Writer Specific Feature Selection and Fuzzy Similarity Measure. arXiv 2019, arXiv:1905.08574. [Google Scholar]
  33. Souza, V.L.F.; Oliveira, A.L.I.; Cruz, R.M.O.; Sabourin, R. Improving BPSO-Based Feature Selection Applied to Offline WI Handwritten Signature Verification through Overfitting Control. In Proceedings of the GECCO’2020: Proceedings of the 2020 Genetic and Evolutionary Computation Conference Companion, Cancun, Mexico, 8–12 July 2020; pp. 69–70. [Google Scholar] [CrossRef]
  34. Manjunatha, K.S.; Guru, D.S.; Annapurna, H. Interval-Valued Writer-Dependent Global Features for Off-Line Signature Verification. In Lecture Notes in Computer Science, Proceedings of the Mining Intelligence and 7 Exploration: 5th International Conference, MIKE 2017, Hyderabad, India, 13–15 December 2017; Springer: Berlin/Heidelberg, Germany, 2017; Volume 10682, pp. 133–143. ISBN 9783319719276. [Google Scholar]
  35. Hafemann, L.G.; Sabourin, R.; Oliveira, L.S. Meta-Learning for Fast Classifier Adaptation to New Users of Signature Verification Systems. IEEE Trans. Inf. Forensics Secur. 2020, 15, 1735–1745. [Google Scholar] [CrossRef]
  36. Batool, F.E.; Attique, M.; Sharif, M.; Javed, K.; Nazir, M.; Abbasi, A.A.; Iqbal, Z.; Riaz, N. Offline Signature Verification System: A Novel Technique of Fusion of GLCM and Geometric Features Using SVM. Multimed. Tools Appl. 2020, v, 1–20. [Google Scholar] [CrossRef]
  37. Ghosh, R. A Recurrent Neural Network Based Deep Learning Model for Offline Signature Verification and Recognition System. Expert Syst. Appl. 2021, 168, 114249. [Google Scholar] [CrossRef]
  38. Agrawal, P.; Chaudhary, D.; Madaan, V.; Zabrovskiy, A.; Prodan, R.; Kimovski, D.; Timmerer, C. Automated Bank Cheque Verification Using Image Processing and Deep Learning Methods. Multimed. Tools Appl. 2021, 80, 5319–5350. [Google Scholar] [CrossRef]
  39. Sharma, N.; Gupta, S.; Mehta, P.; Cheng, X.; Shankar, A.; Singh, P.; Nayak, S.R. Offline Signature Verification Using Deep Neural Network with Application to Computer Vision. J. Electron. Imaging 2022, 31, 041210. [Google Scholar] [CrossRef]
  40. Gandhi, T.; Bhattacharyya, S.; De, S.; Konar, D.; Dey, S. Advanced Machine Vision Paradigms for Medical Image Analysis. In Hybrid Computational Intelligence for Pattern Analysis and Understanding; Academic Press: Amsterdam, The Netherland, 2020. [Google Scholar]
  41. Gonzalez, R.C.; Woods, R.E. Digital Image Processing, 4th ed.; Pearson: New York, NY, USA, 2018; ISBN 9780133356724. [Google Scholar]
  42. Martiana, K.E.; Barakbah, A.R.; Akmilis, S.S.; Hermawan, A.A. Auto Cropping on Iris Image for Iridology Using Histogram Analysis. In Proceedings of the 2016 International Conference on Knowledge Creation and Intelligent Computing (KCIC), Manado, Indonesia, 15–17 November 2016; pp. 42–46. [Google Scholar] [CrossRef]
  43. Sathesh, A.; Babikir Adam, E.E. Hybrid Parallel Image Processing Algorithm for Binary Images with Image Thinning Technique. J. Artif. Intell. Capsul. Netw. 2021, 3, 243–258. [Google Scholar] [CrossRef]
  44. Talab, M.A.; Abdullah, S.N.H.S.; Razalan, M.H.A. Edge Direction Matrixes-Based Local Binary Patterns Descriptor for Invariant Pattern Recognition. In Proceedings of the 2013 International Conference on Soft Computing and Pattern Recognition (SoCPaR), Hanoi, Vietnam, 15–18 December 2013; pp. 13–18. [Google Scholar]
  45. Cajote, R.D.; Guevara, R.C.L. Combining Local and Global Features for Offline Handwriting Recognition. Cajote 2006, 26, 21–32. [Google Scholar]
  46. Starck, J.L.; Candès, E.J.; Donoho, D.L. The Curvelet Transform for Image Denoising. IEEE Trans. Image Process. 2002, 11, 670–684. [Google Scholar] [CrossRef] [PubMed]
  47. Sudha, D.; Ramakrishna, M. Comparative Study of Features Fusion Techniques. In Proceedings of the 2017 International Conference on Recent Advances in Electronics and Communication Technology (ICRAECT), Bangalore, India, 16–17 March 2017; pp. 235–239. [Google Scholar] [CrossRef]
  48. Abdelghani, I.A.B.; Amara, N.E.B. SID Signature Database: A Tunisian Off-Line Handwritten Signature Database. In Lecture Notes in Computer Science, Proceedings of the New Trends in Image Analysis and Processing, ICIAP 2013 Workshops, Naples, Italy, 11 September 2013; Springer: Berlin/Heidelberg, Germany, 2013; Volume 8158, pp. 131–139. ISBN 9783642411892. [Google Scholar]
  49. Soleimani, A.; Fouladi, K.; Araabi, B.N. UTSig: A Persian Offline Signature Dataset. IET Biom. 2017, 6, 1–8. [Google Scholar] [CrossRef]
  50. Singh, S.; Hewitt, M. Cursive Digit and Character Recognition in CEDAR Database. In Proceedings of the 15th International Conference on Pattern Recognition (ICPR-2000), Barcelona, Spain, 3–7 September 2000; Volume 2, pp. 569–572. [Google Scholar]
  51. Abdelghani, I.A.B.; Amara, N.E. Ben Planar Multi-Classifier Modelling-NN/SVM: Application to Off-Line Handwritten Signature Verification. In Advances in Intelligent Systems and Computing; Springer: Berlin/Heidelberg, Germany, 2015; Volume 369, pp. 87–97. ISBN 9783319197128. [Google Scholar]
  52. Abroug, I.; Amara, N.E. Ben Off-Line Signature Verification Systems: Recent Advances. In Proceedings of the International Image Processing, Applications and Systems Conference, Kuala Lumpur, Malaysia, 19–21 October 2015; pp. 1–6. [Google Scholar]
  53. Mersa, O.; Etaati, F.; Masoudnia, S.; Araabi, B.N. Learning Representations from Persian Handwriting for Offline Signature Verification, a Deep Transfer Learning Approach. In Proceedings of the 2019 4th International Conference on Pattern Recognition and Image Analysis (IPRIA), Tehran, Iran, 6–7 March 2019; pp. 268–273. [Google Scholar]
  54. Jagtap, A.B.; Hegadi, R.S.; Santosh, K.C. Feature Learning for Offline Handwritten Signature Verification Using Convolutional Neural Network. Int. J. Technol. Hum. Interact. 2019, 15, 54–62. [Google Scholar] [CrossRef]
  55. Soleimani, A.; Araabi, B.N.; Fouladi, K. Deep Multitask Metric Learning for Offline Signature Verification. Pattern Recognit. Lett. 2016, 80, 84–90. [Google Scholar] [CrossRef]
  56. Narwade, P.N.; Sawant, R.R.; Bonde, S.V. Offline Handwritten Signature Verification Using Cylindrical Shape Context. 3D Res. 2018, 9, 48. [Google Scholar] [CrossRef]
  57. Banerjee, D.; Chatterjee, B.; Bhowal, P.; Bhattacharyya, T.; Malakar, S.; Sarkar, R. A New Wrapper Feature Selection Method for Language-Invariant Offline Signature Verification. Expert Syst. Appl. 2021, 186, 115756. [Google Scholar] [CrossRef]
  58. Arab, N.; Nemmour, H.; Chibani, Y. New Local Difference Feature for Off-Line Handwritten Signature Verification. In Proceedings of the 2019 International Conference on Advanced Electrical Engineering (ICAEE), Algiers, Algeria, 19–21 November 2019; pp. 1–5. [Google Scholar]
  59. Bharathi, R.K.; Shekar, B.H. Off-Line Signature Verification Based on Chain Code Histogram and Support Vector Machine. In Proceedings of the 2013 International Conference on Advances in Computing, Communications and Informatics (ICACCI), Mysore, India, 22–25 August 2013; pp. 2063–2068. [Google Scholar] [CrossRef]
  60. Bhunia, A.K.; Alaei, A.; Roy, P.P. Signature Verification Approach Using Fusion of Hybrid Texture Features. Neural Comput. Appl. 2019, 31, 8737–8748. [Google Scholar] [CrossRef]
  61. Sharif, M.; Khan, M.A.; Faisal, M.; Yasmin, M.; Fernandes, S.L. A Framework for Offline Signature Verification System: Best Features Selection Approach. Pattern Recognit. Lett. 2020, 139, 50–59. [Google Scholar] [CrossRef]
  62. Shariatmadari, S.; Emadi, S.; Akbari, Y. Nonlinear Dynamics Tools for Offline Signature Verification Using One-Class Gaussian Process. Int. J. Pattern Recognit. Artif. Intell. 2020, 34, 2053001. [Google Scholar] [CrossRef]
  63. Jagtap, A.B.; Sawat, D.D.; Hegadi, R.S.; Hegadi, R.S. Verification of Genuine and Forged Offline Signatures Using Siamese Neural Network (SNN). Multimed. Tools Appl. 2020, 79, 35109–35123. [Google Scholar] [CrossRef]
  64. Qiu, S.; Fei, F.; Cui, Y. Offline Signature Authentication Algorithm Based on the Fuzzy Set. Math. Probl. Eng. 2021, 2021, 5554341. [Google Scholar] [CrossRef]
  65. Wencheng, C.; Xiaopeng, G.; Hong, S.; Limin, Z. Offline Chinese Signature Verification Based on AlexNet. In Proceedings of the Advanced Hybrid Information Processing, First International Conference, ADHIP 2017, Harbin, China, 17–18 July 2017; Volume 219, pp. 33–37. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.