1. Introduction
Associative memory (AM) is a knowledge-based system inspired by the human brain’s ability to store and recall information by association [
1,
2]. Apart from a large storage capacity, an ideal AM model should exhibit a certain tolerance to noise. In other words, we expect to retrieve a stored item not only from presenting the original stimulus but also from a similar input stimulus [
1]. We speak of auto-associative memory if stimulus and response coincide. For instance, our memory acts as an auto-associative model when we recognize a friend wearing sunglasses or a scarf. In other words, we obtain the desired output (recognize a friend) from a partial or noise input (their occluded face). We speak of a hetero-associative memory model if at least one stimulus differs from its corresponding response.
Several associative memory models have been introduced in the literature, and their applications range from optimization [
3,
4] and prediction [
5,
6,
7] to image processing and analysis [
8,
9,
10]. Associative memory models have also been applied for pattern classification [
11,
12,
13,
14], including face recognition [
15]. Moreover, the interest in associative memory models increased significantly in the last few years due to their relationship with the attention mechanism used in transformer models [
16,
17,
18].
An AM model designed for storing and recalling fuzzy sets on a finite universe of discourse is called fuzzy associative memory (FAM) [
19]. On the one hand, fuzzy sets can be interpreted as elements from a complete lattice [
20]. On the other hand, mathematical morphology can be viewed as a theory on mappings between complete lattices [
21]. Thus, many important FAM models from the literature belong to the broad class of fuzzy morphological associative memories (FMAMs) [
5,
22]. Briefly, FMAMs are implemented by fuzzy morphological neural networks. A morphological neural network is equipped with neurons that perform an elementary operation from mathematical morphology, possibly followed by a non-linear activation function [
23]. The class of FMAMs includes, for example, the max-minimum and max-product FAMs of Kosko [
19], the max-min FAM of Junbo et al. [
24], the max-min FAM with a threshold of Liu [
25], the fuzzy logical bidirectional associative memories of Belohlavek [
26], and the implicative fuzzy associative memories (IFAMs) of Sussner and Valle [
9].
The max-
C and min-
D auto-associative fuzzy morphological memories (AFMMs), synthesized using fuzzy learning by adjunction (FLA), are fuzzy versions of the matrix-based auto-associative morphological memories (AMMs) proposed by Ritter et al. [
5,
22,
27]. The main features of max-
C and min-
D AFMMs are
unlimited absolute storage capacity,
one-step convergence when employed with feedback,
excellent tolerance to either erosive or dilative noise.
On the downside, matrix-based AFMMs with FLA have many spurious memories [
5]. A spurious memory is an item that is unintentionally stored in the memory [
1]. Furthermore, the information stored on an AFMM with FLA is distributed on a synaptic weight matrix. Consequently, these auto-associative fuzzy memories consume lots of computational resources when designed for storing and recalling large items [
5,
28].
Many auto-associative fuzzy memory models have been proposed in the literature to improve the noise tolerance or reduce the computational cost of AFMMs with FLA. For example, to increase the noise tolerance of the IFAMs, Bui et al. introduced so-called content-association associative memory (ACAM) [
29]. Using a fuzzy preorder relation, Perfilieva and Vajgl proposed a novel theoretical justification for IFAMs [
30]. They also introduced a fast algorithm for data retrieval based on an IFAM model with a binary fuzzy preorder [
31]. Moreover, Vajgl reduced the computational cost of an IFAM by replacing its synaptic weight matrix with a sparse matrix [
32]. Quantale-based associative memories (QAMs) generalize several lattice-based auto-associative memories. They have been effectively applied for storing and recalling large color images [
33]. Li et al. increased the storage capacity of fuzzy associative memories using piecewise linear transformations [
34]. Sussner and Schuster proposed interval-valued fuzzy morphological associative memories (IV-FMAMs) designed for storing and retrieving interval-valued fuzzy sets [
6]. The novel IV-FMAMs have been effectively applied for time-series prediction.
Apart from the distributed models like the FMAMs with FLA and their variations, non-distributed associative memory models have received considerable attention partly due to their low computational effort and extraordinary successes in pattern recognition and image restoration tasks. Examples of non-distributed associative memories include models based on Hamming distance [
35] and kernels [
15] as well as subsethood and similarity measures [
11,
12,
14]. In the context of non-distributed models, we introduced max-plus and min-plus projection auto-associative morphological memories (max-plus and min-plus PAMMs), which can be viewed as non-distributed versions of the auto-associative morphological memories of Ritter et al. [
36]. Max-plus and min-plus PAMMs have fewer spurious memories than their corresponding distributed models. Thus, they are more robust to dilative or erosive noise than the original auto-associative morphological memories. Computational experiments revealed that PAMMs and their compositions are competitive with other methods from the literature on classification tasks [
36].
In our conference paper [
37], we introduced max-
C projection auto-associative fuzzy morphological memories (max-
C PAFMMs) as an alternative to AFMMs, building on the success of max-plus and min-plus PAMMs. Max-
C PAFMMs were further discussed in [
38], where some results concerning their implementation and storage capacity were given without proof. In a few words, a max-
C PAFMM projects the input into the family of all max-
C combinations of the stored items. Furthermore, we developed the dual version of max-
C PAFMMs, the class of min-
D PAFMMs which projects the input into the set of all min-
D combinations of the stored items, in conference paper [
39]. Although we addressed some theoretical results concerning both max-
C and min-
D PAFMM models in [
39], the theoretical part of this paper can be viewed as an extended version of our previous papers [
37,
38,
39]. In particular, this paper provides the mathematical background for developing the new PAFMM models, which are obtained by enriching a complete lattice with a residuated binary operation with a left identity. Also, we show in this paper that the most robust max-
C PAFMM for inputs corrupted by dilative noise is based on Zadeh’s inclusion measure. The resulting model is referred to as Zadeh’s max-
C PAFMM. Accordingly, the dual of Zadeh’s max-C PAFMM is the min-
D PAFMM most robust to erosive noise. Finally, inspired by the work of Urcid and Ritter [
40], the frail tolerance of the max-
C and min-
D PAFMMs for mixed noise can be improved significantly by masking the noise contained in the input [
41]. Although some preliminary experiments can be found in conference paper [
41], in this paper, we provide conclusive computational experiments concerning the application of Zadeh’s max-
C PAFMM for face recognition.
The paper is organized as follows. The mathematical background, including some basic concepts on lattice theory and fuzzy logic, is presented in the next section.
Section 2 also introduces an algebraic structure called residuated complete lattice-ordered groupoid with left identity (R-clogli), which provides the mathematical background for developing PAFMM models.
Section 3 briefly reviews max-
C and min-
D AFMMs with FLA. Max-C and min-
D PAFMMs are addressed subsequently in
Section 4. Zadeh’s PAFMMs and the noise masking strategy are discussed in
Section 5 and
Section 6, respectively. The performance of Zadeh’s max-
C PAFMM for face recognition tasks is addressed in
Section 7. The paper finishes with some concluding remarks in
Section 8 and
Appendix A containing proofs of the main theorems.
2. Some Basic Concepts on Fuzzy Systems
This section reviews the mathematical background necessary for developing the new max-
C and min-
D projection associative memories, including lattice theory, residuated lattices, and related concepts. Readers familiar with complete lattice and mathematical morphology may skip
Section 2.1 or read it later. Similarly, readers familiar with fuzzy logic operations may skip
Section 2.3.
2.1. Complete Lattice and Mathematical Morphology
A non-empty partially ordered set
is called a complete lattice, denoted by
, if every subset
has an infimum and a supremum
[
42]. The infimum and the supremum of
are denoted by
and
, respectively. The least and the greatest elements of complete lattice
are denoted, respectively, by
and
. When
is a finite subset of
, we write the infimum and the supremum of
X as
and
, respectively.
The unit interval
is an example of a complete lattice with the usual order. Cartesian product
of complete lattice
is also a complete lattice with the component-wise ordering defined as follows for
and
:
Mathematical morphology is a non-linear theory widely used for image processing and analysis [
21,
43,
44,
45]. The elementary operations from mathematical morphology are dilations and erosions. Dilations and erosions are operators that commute with the supremum and infimum operations, respectively [
21,
44]. Formally, given complete lattices
and
, operators
and
represent, respectively, a dilation and an erosion if
for all
and
. We recall that dilation
and erosion
satisfy
and
[
43].
One key concept of mathematical morphology on complete lattices is the notion of adjunction [
43]. Adjunctions arise naturally on complete lattices and are closely related to Galois connection and residuation theory [
42,
46,
47,
48]. We consider operators
and
between complete lattices
and
. We say that
is an adjunction between
and
if
Adjunctions can be used to define the elementary operations from mathematical morphology. In fact, if
forms an adjunction, then
is an erosion, and
is a dilation. Conversely, given dilation
, residual operator
is unique erosion
such that
is an adjunction [
43,
47]. Dually, given erosion
, residual operator
is unique dilation
such that
is an adjunction.
2.2. Residuated Complete Lattice-Ordered Groupoids with Left Identity
This section reviews mathematical structures obtained by enriching a complete lattice with binary operations. Let us begin by introducing the following definition:
Definition 1 (Residuated complete lattice-ordered groupoid with left identity). A residuated complete lattice-ordered groupoid with left identity (R-clogli), denoted by , is an algebra such that
- 1.
is a complete lattice.
- 2.
is a groupoid (or magma) where the binary operation has a left identity, that is, there exists such that for all .
- 3.
Operations and ∖ satisfy the following adjunction relationship:Similarly, a dual R-clogli denoted by is an algebra such that is a complete lattice equipped with a binary operation that has a left identity, and and satisfyWe speak of an associative R-clogli and an associative dual R-clogli if binary operations and are associative.
From the adjunction relationship, we conclude that
is a dilation while
is an erosion in the first argument. Specifically, given fixed element
a in
, operators
defined by
and
for all
are a dilation and an erosion, respectively. As a consequence, we have
Moreover, the following identities hold for all
:
In a word, ∖ and
are the residuals of
and
, respectively.
The following briefly addresses the relationship between R-clogli and other mathematical structures from the literature.
First, an R-clogli is a residuated complete lattice-ordered dilative monoid, R-clodim for short, if the binary operation is associative and has a two-sided identity. Sussner introduced the notion of R-clodim as an appropriate mathematical background for
-fuzzy mathematical morphology [
49], and it is closely related to Maragos’ complete lattice-ordered double monoid [
50].
Residuated lattices, introduced by Ward and Dilworth [
51] and widely taken as an appropriate mathematical background for fuzzy logic [
52,
53], are also R-cloglis. In fact, a complete residuated lattice, denoted by
, is an algebra such that
is a complete lattice with least element and greatest element .
Operation is commutative, associative, and for all .
Operations and → are adjoint, that is, if and only if .
Therefore, a complete residuated lattice is equivalent to an R-clogli in which is associative and commutative, and the greatest element of is its identity element.
Finally, an R-clogli in which
is associative and performs a dilation in both arguments (not only in the first) is a quantale. Quantales, introduced by Mulvey to provide a constructive formulation for the logic of quantum mechanics [
54], have been used to develop quantale-based auto-associative memories (QAMs). QAMs include the original auto-associative morphological memories and auto-associative fuzzy implicative memories as particular instances [
33].
2.3. Fuzzy Sets and Fuzzy Logic Operations
The associative memories considered in this paper are based on fuzzy set theory and operations from fuzzy logic. This section briefly reviews the most essential concepts of fuzzy systems. The reader is invited to consult [
55,
56,
57,
58,
59] for a detailed review of fuzzy logic and fuzzy set theory.
Fuzzy set
A on a universe of discourse
X is defined by
where
is the membership function of fuzzy set
A. The family of all fuzzy sets on
X is denoted by
. When the universe of discourse
is finite, fuzzy set
A can be identified with vector
, where
for all
[
19]. In this paper, we focus only on fuzzy sets defined in finite universes of discourse. Moreover, we identify them with vectors on hypercube
.
Definition 2 (Fuzzy conjunction and disjunction). Increasing mapping is a fuzzy conjunction if it satisfies and . A fuzzy disjunction is increasing operator such that and .
Definition 3 (Fuzzy implication and co-implication). A fuzzy implication is mapping decreasing in the first argument and increasing in the second argument that satisfies identities and . A fuzzy co-implication is operator decreasing in the first argument and increasing in the second argument such that and .
We note that fuzzy logic connectives are generalizations of classical connectives. For the development of fuzzy morphological associative memories, we focus our attention on fuzzy conjunctions, which have a left identity and satisfy the following adjunction relationship:
with respect to binary operation
called the residual of
C. From the mathematical point of view, we assume that
is an R-clogli. We remark that
R, the residual of a fuzzy conjunction
C, is a fuzzy implication if and only if
implies
. Indeed, we suppose there exists
such that
. From the adjunction relationship, we have
, and thus
R is not a fuzzy implication.
In a similar fashion, we focus on fuzzy disjunctions which have a left identity and satisfy
for binary operation
called the residual of
D. Throughout the paper, we assume that
is a dual R-clogli. We remark that residual
S, derived from
D using (
12), is a fuzzy co-implication if and only if
implies
.
Examples of R-clogli and dual R-clogli include unit interval equipped with the following:
Minimum fuzzy conjunction and Gödel’s implication .
Lukasiewicz’s fuzzy conjunction and fuzzy implication .
Gaines’ fuzzy conjunction and fuzzy implication defined, respectively, by
Maximum fuzzy disjunction
and Gödel’s fuzzy co-implication
Lukasiewicz’s disjunction and co-implication .
Gaines’ fuzzy disjunction and fuzzy co-implication defined as follows:
We note that and are neither commutative nor have a two-sided identity. Thus, these fuzzy logical operators do not yield a complete residuated lattice nor a residuated complete lattice-ordered dilative monoid (R-clodim). Notwithstanding, algebra is an R-clogli while is a dual R-clogli.
Apart from the adjunction relationship, fuzzy logical operators can be connected through a strong fuzzy negation. A strong fuzzy negation is a nonincreasing mapping such that , , and for all . Standard fuzzy negation is a strong fuzzy negation.
A fuzzy logic operator
can be connected to a fuzzy logic operator
by means of a strong fuzzy negation
as follows:
In this case, we say that pair
is dual with respect to
. For example, the pairs of fuzzy conjunction and fuzzy disjunction,
,
, and
, are duals with respect to standard fuzzy negation
. Pairs
,
, and
of fuzzy implication and fuzzy co-implication are also dual with respect to standard fuzzy negation. The commutative diagram shown in
Figure 1 establishes the relationship between adjunction and negation [
21,
60].
Fuzzy logic operators can be combined with either maximum or minimum operations to yield matrix products. For instance, the max-
C and the min-
D matrix products of
by
, denoted, respectively, by
and
, are defined by the following equations for all
and
:
In analogy to the concept of linear combination, we say that
is a max-
C combination of the vectors belonging to finite set
if
where
for all
. Similarly, a min-
D combination of the vectors of
is given by
where
, for all
. The sets of all max-
C combinations and min-
D combinations of
are denoted, respectively, by
and
The sets of max-
C and min-
D combinations play a major role in the projection auto-associative fuzzy morphological memories (PAFMMs) presented in
Section 4. However, before introducing PAFMMs, let us briefly review the fuzzy auto-associative morphological memories, defined using fuzzy logical connectives and adjunctions.
3. Auto-Associative Fuzzy Morphological Memories
Let us briefly review the auto-associative fuzzy morphological memories (AFMM). The reader interested in a detailed account of this subject is invited to consult [
5,
22]. To simplify the exposition, we assume that
and
, where
C is a fuzzy conjunction and
D is a fuzzy disjunction, are, respectively, an R-clogli and a dual R-clogli.
As far as we know, most AFMMs are implemented by a single-layer network defined in terms of either the max-
C or the min-
D matrix products established by (
17) [
5]. Formally, max-
C and min-
D auto-associative fuzzy morphological memories (AFMMs) are mappings
defined, respectively, by the following equations:
where
are called the synaptic weight matrices. Examples of AFMMs include the auto-associative version of the max-minimum and max-product fuzzy associative memories of Kosko [
19], the max-min fuzzy associative memories with a threshold [
25], and the implicative fuzzy associative memories [
9].
We point out that using a strong fuzzy negation
, we can derive from a max-
C AFMM
another AFMM called the negation of
and denoted by
. Formally, the negation of
is defined by equation
where the strong fuzzy negation
is applied in a component-wise manner. It is not hard to show that the negation of a max-
C AFMM
is a min-
D AFMM
, and vice-versa, where fuzzy conjunction
C and fuzzy disjunction
D are dual with respect to strong fuzzy negation
, i.e., they satisfy (
16) [
22].
Let us now turn our attention to a recording recipe called fuzzy learning by adjunction (FLA), which can be effectively used for the storage of vectors on AFMM
and
defined by (
22) [
22]. Given
, called the fundamental memory set, FLA determines matrix
of a max-
C AFMM and matrix
of the min-
D AFMM by means of the following equations for all
:
where
R and
S are the residuals of
C and
D, respectively.
The following proposition reveals that a min-
D AFMM
and a max-
C AFMM
, both synthesized using FLA, project input
into the set of their fixed points if
C and
D are both associative [
5]. Furthermore, Proposition 1 shows that output
of a min-
D AFMM with FLA is the greatest fixed-point less than or equal to input
. Analogously, a max-
C AFMM with FLA yields the least fixed-point greater than or equal to the input [
5].
Proposition 1 (Valle and Sussner [
5]).
We let and be an associative R-clogli and an associative dual R-clogli, respectively. The output of the min-D AFMM defined by (22) with FLA given by (24) satisfieswhere denotes the set of all fixed points of which depends on and includes the fundamental memory set . Dually, the output of the max-C AFMM with FLA satisfieswhere denotes the set of all fixed points of , which also depends on and contains fundamental memory set . As a consequence of Proposition 1, AFMMs with FLA present the following properties: they can store as many vectors as desired; they have a large number of spurious memories; an AFMM exhibits tolerance to either dilative noise or erosive noise, but it is susceptible to mixed (dilative+erosive) noise. We recall that distorted version
of fundamental memory
underwent a dilative change if
. Dually, we say that
underwent an erosive change if
[
27].
Example 1. We consider fundamental memory setUsing Gödel’s co-implication in (24), synaptic weight matrix of the min- AFMM with FLA isNow, we consider input fuzzy setWe note that is a dilated version of fundamental memory because . The output of the min- AFMM with FLA iswhere “” denotes the min- product defined in terms of fuzzy disjunction . According to Proposition 1, output is a fixed point of that does not belong to fundamental memory set . Thus, it is a spurious memory of . Similarly, we can use FLA to store fundamental set into the min-D AFMMs and obtained by considering, respectively, Lukasiewicz and Gaines fuzzy disjunctions. Upon presentation of input vector given by (29), the min-D AFMMs and yield, respectively,andsuch that the min- AFMM , the auto-associative memories and failed to produce the desired output, . 4. Max- and Min- Projection Auto-Associative Fuzzy Morphological Memories
As distributed matrix-based auto-associative memories, a great deal of computer memory is consumed by min-D and max-C AFMMs if length n of the stored vectors is considerable. Furthermore, from Proposition 1, their tolerance to either dilative or erosive noise is degraded as the number of fixed points increases.
Inspired by the feature that min-
D and max-
C AFMMs with FLA project the input vector into the set of their fixed point, we can improve the noise tolerance of these memory models by reducing their set of fixed points. Accordingly, we recently introduced the max-
C projection auto-associative fuzzy memories (max-
C PAFMMs) by replacing in (
25) set
by set
of all max-
C combinations of vectors of
[
37,
38]. Formally, we let
be an R-clogli. Given set
of fundamental memories, a max-
C PAFMM
is defined by
where set
is defined in (
20). A dual model, referred to as min-
D PAFMM, is obtained by replacing
by set
of all min-
D combinations of the fundamental memories in (
26). Specifically, we let
be a dual R-clogli. A min-
D PAFMM
satisfies
where set
is given in (
21). The following theorem is a straightforward consequence of these definitions.
Theorem 1. We let and be an R-clogli and a dual R-clogli, respectively. The max-C and min-D PAFMMs given, respectively, by (33) and (34) satisfy inequalities for any input vector . Furthermore, and for all . As a consequence of Theorem 1, a max-
C PAFMM and a min-
D PAFMM are, respectively, opening and closing form fuzzy mathematical morphologies [
61]. Like min-
D AFMM, max-
C PAFMM exhibits only tolerance to dilative noise. Also, it is susceptible to either erosive or mixed noise. In fact, fundamental memory
cannot be retrieved by a max-
C PAFMM from an input such that
. Similarly, like max-
C AFMM, min-
D PAFMM
exhibits tolerance to erosive noise, but it is not robust to dilative or mixed noise.
Let us now address the absolute storage capacity of max-C and min-D PAFMMs. The following theorem shows that all fundamental memories are fixed points of and . Thus, max-C and min-D PAFMMs exhibit unlimited absolute storage capacity.
Theorem 2. We let and be, respectively, an R-clogli and a dual R-clogli and consider a fundamental memory set . Max-C PAFMM given by (33) satisfies for all . Dually, for all , where S denotes min-D PAFMM given by (34). The following theorem, which is a straightforward consequence of the adjunction relationships given by (
11) and (
12), provides effective formulas for the implementation of max-
C and min-
D PAFMMs.
Theorem 3. We let and be an R-clogli and a dual R-clogli, respectively. Also, we consider a fundamental memory set and let be an arbitrary input. Max-C PAFMM given by (33) satisfiesSimilarly, the output of min-D PAFMM can be computed by Remark 1. Theorem 3 above gives a formula for coefficient that is used to define the output of max-C PAFMM. In some sense, coefficient corresponds to the degree of inclusion of fundamental memory in input fuzzy set . Specifically, if residual R of fuzzy conjunction C is a fuzzy implication, then we havewhere denotes the Bandler–Kohout fuzzy inclusion measure [62]. Regarding computational effort, PAFMMs are generally less expensive than their corresponding AFMMs because they are non-distributive memory models, which means that they do not require the storage of a synaptic weight matrix of size
. Additionally, they involve fewer floating-point operations than their corresponding min-
D and max-
C AFMMs if
. To illustrate this remark, we consider a fundamental memory set
, where
for all
with
. On the one hand, to synthesize the synaptic weight matrix of a min-
D AFMM
, we perform
evaluation of residual
S and
comparisons. In addition, the resulting synaptic weight matrix consumes
of the memory space. In the recall phase, the min-
D AFMM
requires
evaluations of a fuzzy disjunction and
comparisons. On the other hand, to compute parameters
’s of a max-
C PAFMM
, we perform
evaluations of a residual operator and
comparisons. The subsequent step of the max-
C PAFMM
requires
evaluations of a fuzzy conjunction and
comparisons. Lastly, it consumes
of memory space to store the fundamental memories. Similar remarks hold for a max-
C AFMM and a min-
D PAFMM.
Table 1 summarizes the computational effort in the recall phase of AFMMs and PAFMMs. Here, fuzzy operations refer to evaluations of fuzzy conjunctions or disjunctions and their residual operators.
Finally, different from min-D and max-C AFMMs, max-C and min-D PAFMMs are not dual models with respect to a strong fuzzy negation. The following theorem shows that the negation of a min-D PAFMM is a max-C PAFMM designed to store the negation of fundamental memories, and vice-versa.
Theorem 4. We let and be, respectively, an R-clogli and a dual R-clogli where pairs and are dual operators with respect to strong fuzzy negation η. Given fundamental memory set , we define by setting , for all and . Also, we let and be, respectively, max-C and the min-D PAFMMs designed for the storage of and define their negation as follows for every :The negation of is the max-C PAFMM designed for the storage of , that is,Analogously, negation of is the min-D PAFMM given by It follows from Theorem 4 that negations and fail to store fundamental memory set .
Example 2. We consider fundamental memory set given by (27). We let and be the minimum fuzzy conjunction and the fuzzy implication of Gödel, respectively. We synthesize the max-C PAFMM designed for the storage of using adjunction pair . From Theorem 2, equation holds for . Given input vector defined by (29), we obtain from (35) coefficientsThus, the output of the max-C PAFMM isWe note that failed to retrieve fundamental memory . Analogously, we can store fundamental memory set into the max-C PAFMM using Lukasiewicz fuzzy conjunction and implication. Upon presentation of vector , the max-C PAFMM producesLike the min-D AFMM , memory failed to recall fundamental memory . Nevertheless, the outputs of the max-C PAFMMs and are more similar to the desired vector than the min-D AFMMs , , and (see Example 1). Quantitatively, Table 2 shows the normalized mean squared error (NMSE) between the recalled vector and the desired output . We recall that the NMSE between and is given byThis simple example confirms that a max-C PAFMM can exhibit a better tolerance with respect to dilative noise than its corresponding min-D AFMM.
Table 2.
Normalized mean squared error.
Table 2.
Normalized mean squared error.
• | x | | | | | | |
---|
| 0.33 | 0.32 | 0.05 | 0.33 | 0.01 | 0.05 | 0.00 |
Let us conclude the section by emphasizing that we can only ensure optimal absolute storage capacity if C has a left identity.
Example 3. We consider the “compensatory and” fuzzy conjunction defined byFuzzy conjunction does not have a left identity. Moreover, the fuzzy implication that forms an adjunction with isNow, we let be the max- PAFMM designed for the storage of fundamental memory set given by (27). Upon presentation of fundamental memory as input, we obtain from (35) coefficientsThus, the output vector of the max-C PAFMM isIn a similar fashion, using fundamental memories and as input, we obtain from outputsWe note that inequality holds for . However, fundamental memories , , and are not fixed points of the max-C PAFMM . 5. Zadeh Max- PAFMM and Its Dual Model
The noise tolerance of an auto-associative memory is usually closely related to its spurious memories. In general, the more spurious memories an auto-associative memory has, the less tolerant to noise it is. We recall that a spurious memory is a fixed point of an auto-associative memory that does not belong to the fundamental memory set [
1]. For example, in Example 2, fixed point
is a spurious memory of the max-
C PAFMM
. Indeed, the set of fixed points, max-
C PAFMM, corresponds to the set of all max-
C combinations of the fundamental memories. Therefore, the noise tolerance of a max-
C PAFMM increases as family
becomes smaller.
Interestingly, by considering the fuzzy conjunction of Gaines
in (
20), we can significantly reduce
, where
is a fundamental memory set
. Indeed, from Theorem 3, the output of the max-
C PAFMM based on Gaines’ fuzzy conjunction is given by
where
Here,
denotes the fuzzy inclusion measure of Zadeh defined as follows for all
:
In other words, coefficients
are determined using Zadeh’s fuzzy inclusion measure
. Hence, this max-
C PAFMM is called Zadeh’s max-
C PAFMM and is denoted by
.
From (
52), coefficient
is either zero or one. Moreover,
if and only if
for all
. Also, we have
and
for all
. Therefore, for any input
, the output of Zadeh’s max-
C PAFMM is alternatively given by equation
where
is the set of indexes
such that
is less than or equal to input
, i.e.,
. Here, we have
if
, where
is a vector of zeros.
In a similar manner, from (
36), the dual of Zadeh’s max-
C PAFMM is the min-
D PAFMM defined by
Here,
and
denote the fuzzy disjunction and the fuzzy co-implication of Gaines, respectively. Alternatively, the output of Zadeh’s min-
D PAFMM is given by
where
is the set of indexes
such that
. Here, we have
if
, where
is a vector of ones.
We note from (
53) and (
56) that no arithmetic operation is performed during the recall phase of Zadeh’s max-C PAFMM and its dual model; they only perform comparisons! Thus, both
and
are computationally cheap and fast associative memory models. In addition, Zadeh’s max-
C PAFMM is exceptionally robust to dilative noise, while its dual model
exhibits excellent tolerance to erosive noise. The following theorem addresses the noise tolerance of these memory models.
Theorem 5. We consider fundamental memory set . Identity holds true if there exists a unique such that . Furthermore, if there exists a unique such that then .
Example 4. We consider fundamental memory set given by (27) and input fuzzy set defined by (29). Clearly, , , and . Thus, the set of indexes defined by (54) is . From (53), the output of Zadeh’s max-C PAFMM is We note that the max-C PAFMM perfectly recalled the original fundamental memory. As a consequence, the NMSE is zero. From Table 2, the max-C PAFMM of Zadeh yielded the best NMSE, followed by the max-C PAFMMs , and . Let us conclude this section by remarking that Zadeh’s max-
C PAFMM also belongs to the class of
-fuzzy associative memories (
-FAMs) proposed by Esmi et al. [
11].
Remark 2. An auto-associative Θ-FAM is defined as follows: We consider fundamental memory set and let be operators such that for all . Given input and weight vector , Θ-FAM yields where is the following set of indexes:Now, the max-C PAFMM of Zadeh is obtained by considering and , for all . Specifically, in this case, coincides with the set of index defined by (53). 6. Noise-Masking Strategy for PAFMMs
Unfortunately, a max-
C PAFMM cannot retrieve fundamental memory
from input
. Hence, one of the major weaknesses of max-
C PAFMMs is their limited tolerance to erosive or mixed noise, which restricts their application to real-world problems. Similar remarks hold for min-
D PAFMMs, according to the duality principle. However, the noise tolerance of a PAFMM can be significantly enhanced by masking the noise contained in a corrupted input [
40]. Specifically, the noise-masking strategy aims to transform an input degraded by mixed noise into a vector corrupted by either dilative or erosive noise. Inspired by the works of Urcid and Ritter [
40], we present a noise-masking strategy for PAFMMs (which has been previously discussed in conference paper [
41]). To simplify the presentation, we focus only on max-
C PAFMMs.
Suppose we have a max-C PAFMM, denoted by , which was synthesized using fundamental memory set . Let be a version of fundamental memory corrupted by mixed noise. In this case, represents the masked input vector. This masked vector contains only dilative noise, meaning inequality holds. Since the max-C PAFMM is robust to dilative noise, we can expect it to be able to retrieve the original fuzzy set when presented with masked vector .
Nevertheless, to mask the noise, we need to know beforehand which fundamental memory is corrupted. To overcome this practical shortcoming, Urcid and Ritter proposed comparing masked vector to original input and each fundamental memory , . The comparison is based on a meaningful measure, such as normalized mean squared error (NMSE).
According to our previous study [
41], we suggest using a fuzzy similarity measure to determine the masked vector. Specifically, a fuzzy similarity measure is mapping
, which yields the degree of similarity between
and
[
63,
64,
65,
66,
67]. By using a fuzzy similarity measure, we can obtain masked vector
by computing the maximum similarity between input
and fundamental memory
. In mathematical terms, we have
where
is an index such that
In summary, using noise masking to retrieve vectors through a max-
C PAFMM
results in an auto-associative fuzzy morphological memory
defined by
where
is an index that satisfies (
59). Similarly, we can use the technique of noise masking for the recall of vectors using a min-
D PAFMM
. Formally, we denote by
the auto-associative fuzzy morphological memory given by
where
is an index that satisfies (
59) and
is a min-
D PAFMM.
Finally, we mentioned in the previous section that the Zadeh max-
C PAFMM
and its dual model do not perform floating point arithmetic operations. However, some arithmetic operations may be required for computing the masked input fuzzy set. For example, if we use the Hamming similarity measure
defined by
then
performs
floating point operations during the retrieval phase.
Example 5. We consider fundamental memory set given by (27) but let the input fuzzy set beWe point out that is obtained by introducing mixed noise into fundamental memory . Specifically, we have . Because a max-C PAFMM exhibits only tolerance to dilative noise, it is not able to retrieve fundamental memory . For example, Zadeh’s max-C PAFMM yields , which is none of the fundamental memories. In other words, Zadeh’s max-C PAFMM fails to retrieve the desired fundamental memory . However, the Hamming similarity measure given by (62) results inHence, satisfies (59). Furthermore, using the noise masking strategy, the max-C PAFMM yieldsIn conclusion, the max-C PAFMM with the noise masking strategy perfectly recalls original fundamental memory . In light of this example, we only consider fuzzy morphological associative memories with the noise masking strategy in the following section. 7. Computational Experiments
Inspired by the auto-associative memory-based classifiers described in [
13,
15], we propose the following auto-associative memory-based classifier for face images. Suppose we have a training dataset with
different face images from an individual
i, for
. Each face image is encoded into a column-vector
, where
and
. We address below two approaches to encode face images into
. For now, we let
denote an auto-associative memory designed for the storage of the fundamental memory set
composed of all training images from individual
. Given an unknown face image, we also encode it into a column-vector
using the same procedure as the training images. Then, we present
as input to the auto-associative memories
’s. Finally, we assign the unknown face image to the first individual
such that
where
denotes a fuzzy similarity measure. In other words,
belongs to an individual such that the recalled vector is the most similar to the input.
In our experiments, we used in (
66) the Hamming similarity measure defined by (
62). Furthermore, a face image was encoded into a column-vector
using either one of the following two approaches:
A pre-processing approach given by the sequence of
MATLAB-style commands:
rgb2gray (This command is applied only if the input is a color face image in the RGB color space),
im2double,
imresize, and
reshape. We point out that we resized the images according to the dimensions used by Feng et al. [
68].
A convolutional neural network as feature extractor followed by data transformation. Specifically, we used the
python package
face_recognition (Available at
https://github.com/ageitgey/face_recognition, accessed on 11 October 2023), which is based on
Dlib library [
69]. Package
face_recognition includes a pre-trained ResNet network designed to extract features for face recognition. The ResNet feature extractor maps a face image into vector
. We obtained
by applying the following transformation where
and
denote the mean and standard deviation of the
ith component of all 128-dimensional training vectors:
Apart from literature models, we only consider the min-
AFMM and Zadeh’s max-
C PAFMM. We recall that the AFMMs based on Lukasiewicz connectives outperformed many other AFMMs in experiments concerning the retrieval of gray-scale images [
9]. Furthermore, the min-
AFMM can be obtained from the morphological auto-associative memory of Ritter et al. using thresholds [
9,
27]. As pointed out in
Section 5, Zadeh’s PAFMM is expected to exhibit tolerance to dilative noise larger than the other max-
C PAFMMs. Thus, we synthesized the classifiers based on the min-
D AFMM
and Zadeh’s max-
C PAFMM
, both equipped with the noise masking strategy described by (
60) and (
62). These classifiers, combined with the first encoding strategy listed above, are denoted, respectively, by Resized+
and Resized+
. Similarly, we refer to ResNet+
and ResNet+
, the fuzzy associative memory-based classifiers combined with the second approach listed above.
Performances of the
- and
-based classifiers have been compared with the following approaches from the literature: sparse representation classifier (SRC) [
70], linear regression-based classifier (LRC) [
71], collaborative representation-based classifier (CRC) [
72], fast superimposed sparse parameter (FSSP(1)) classifier [
68], and the deep ResNet network available at the
python face_recognition package. We point out that face recognition methods SRC, LRC, CRC, and FSSP(1) are included for comparison purposes because they use subspace projection methods, which somewhat resemble the proposed PAFMM classifier. In contrast, the ResNet is among the state-of-the-art deep neural networks for image classification tasks. According to
Dlib package developers, ResNet achieved a 99.38% accuracy on the standard “Labeled Faces in the Wild” benchmark, a performance comparable to that of other state-of-the-art approaches. Despite the many other face recognition methods in the literature, these are representative approaches for highlighting the potential application of the proposed projection auto-associative fuzzy morphological memories.
7.1. Face Recognition with Expressions and Pose
Face recognition has been an active research topic in pattern recognition and computer vision due to its applications in human–computer interaction, security, access control, and others [
15,
68,
72]. To evaluate the performance of the
-based classifier, we conducted experiments using three standard face databases, namely the Georgia Tech Face Database (GT) [
73], the AT&T Face Database [
74], and the AR Face Image Database [
75]. These face databases incorporate pose, illumination, and gesture alterations.
The Georgia Tech (GT) Face Database contains face images of 50 individuals taken in two or three sessions at the Center for Signal and Image Processing at the Georgia Institute of Technology [
73]. These images, up to 15 per individual, show frontal and tilted faces with different facial expressions, lighting conditions, and scales. In this paper, we used the cropped images in the GT dataset.
Figure 2 presents 15 facial images of one individual from the GT database. As pointed out previously, the color images from the cropped GT database were converted into gray-scale face images and resized to
pixels before being presented to the classifiers SRC, LRC, CRC, FSSP(1), Resized+
, and Resized+
.
The AT&T database, formerly known as the ORL database of faces, has 10 different images for each of 40 distinct individuals [
74]. All face images are in up-right and frontal positions. The ten images of an individual are shown in
Figure 3 for illustrative purposes. For classifiers SRC, LRC, CRC, FSSP(1), Resized+
, and Resized+
, the face images of the AT&T database were resized to
pixels.
The AR face image database contains over 4000 facial images from 126 individuals [
75]. For each individual, 26 images were taken in two sessions, separated by two weeks. The face images feature different facial expressions, illumination changes, and occlusions. Our experiments used a subset of the cropped AR face image database with face images of 100 individuals. Furthermore, we only considered each individual’s eight facial images with different expressions (normal, smile, anger, and scream). The eight face images of one individual of the AR database are shown in
Figure 4. Finally, we point out that the images in the AR database were converted to gray-scale images and resized to
pixels for classifiers SRC, LRC, CRC, FSSP(1), Resized+
, and Resized+
.
Figure 2.
Images of one individual from the GT face database.
Figure 2.
Images of one individual from the GT face database.
Figure 3.
Images of one individual from the AT&T face database.
Figure 3.
Images of one individual from the AT&T face database.
Figure 4.
Some images of one individual from the AR face database.
Figure 4.
Some images of one individual from the AR face database.
For GT and the AT&T face databases, we followed the “first
N” scheme adopted by Feng et al. [
68]. Each person’s first
N face images were used as the training set. The remaining face images of each individual were used for testing. Number
N varied according to the computational experiments described in [
68]. As for the AR face image database, we also followed the same evaluation protocol described in [
68]: three facial expressions were used for training (e.g., normal, smile, and anger), while the remainder were used for testing (e.g., scream).
Table 3,
Table 4 and
Table 5 list the recognition rates (RRs) yielded by the classifiers. These tables also provide the average recognition rate (ARR) of a given scenario. For a visual interpretation of the overall performance of the classifiers,
Figure 5a shows the box plot comprising the normalized recognition rates listed in
Table 3,
Table 4 and
Table 5. The normalized recognition rates were obtained by subtracting and dividing the values in
Table 3,
Table 4 and
Table 5, respectively, by column-wise mean and standard deviation. Furthermore,
Figure 5b shows the Hasse diagram obtained from the outcome of the Wilcoxson signed-ranks test comparing any two classifiers with a confidence level at 95% using all the recognition rates listed in
Table 3,
Table 4 and
Table 5 [
76,
77,
78]. Specifically, two classifiers were connected by an edge if the test rejected the null hypothesis that the two classifiers perform equally well against the alternative hypothesis that the recognition rates of the classifier at the top are significantly larger than the recognition rates of the classifier on the bottom of the edge. In other words, the method at the top outperformed the method at the bottom of an edge. Also, we refrained from including the edges that can be derived from transitivity. For example, from
Figure 5b, we derived data signifying that
+
is above ResNet, and ResNet is above FSSP(1). Thus, we deduced that the
+
-based classifier outperformed FSSP(1) in these experiments.
In conclusion,
Figure 5 shows that the ResNet+
and ResNet+
-based classifiers outperformed all other classifiers, including the ResNet and FSSP(1) models, for recognition of uncorrupted face images. We recall, however, that Zadeh’s PAFMM
is computationally cheaper than AFMM
. Let us now evaluate the performance of the classifiers in the presence of noise.
7.2. Face Recognition in the Presence of Noise
In many practical situations, captured images are susceptible to different levels of noise and blurring effects. According to Gonzalez and Woods [
79], the principal noise sources in digital images arise during image acquisition or transmission. The performance of imaging sensors is affected by various factors, such as environmental conditions and the quality of the sensing elements. For instance, Gaussian noise arises in an image due to electronic circuit noise and sensor noise factors, while transmission errors cause salt and pepper noise. A blurred image may arise when the camera is out of focus, or relative motion exists between the camera and objects in the scene.
Figure 6 displays undistorted and corrupted versions of an image from the AT&T face database. The noise images were obtained by introducing salt and pepper noise with probability
, Gaussian noise with a mean zero, variance
, and a horizontal motion of nine pixels (blurred images).
In order to simulate real-world conditions, we evaluated the performance of the classifiers when the training images were not distorted, but some noise corrupted test images. Specifically, test images were corrupted by the following kinds of noise:
Salt and pepper noise with probability ;
Gaussian noise with mean 0 and variance ;
Horizontal motion whose number of pixels varied from 1 to 20.
In each scenario, the first five images of each individual of the AT&T face database were used for training. The remaining images corrupted by some noise were used for testing.
Figure 7 shows the average recognition rates (ARRs) produced by the eight classifiers in 30 experiments for each noise intensity. Furthermore,
Figure 8 shows the Hasse diagram of the outcome of the Wilcoxson signed-ranks test comparing any two classifiers with a confidence level of 95%. In contrast to the previous experiment with undistorted face images, classifiers ResNet, ResNet+
, and ResNet+
exhibited the worst recognition rates for corrupted input images. Moreover, we concluded the following from
Figure 7 and
Figure 8:
FSSP(1) and Resized+ were, in general, the best classifiers for the recognition of images corrupted by salt and pepper noise.
FSSP(1) and SRC yielded the largest recognition rates in the presence of Gaussian noise.
Resized+ outperformed all the other classifiers for recognizing blurred input images.
In general, Resized+ and FSSP(1) were the most robust classifiers for corrupted input images.
7.3. Computational Complexity
Let us conclude this section by analyzing the computational complexity of the classifiers considered. To this end, we let c denote the number of individuals, k be the number of training images per individual, and n be either the number of pixels of the resized face image or for vectors encoded by ResNet.
First of all, the following inequalities rank the computational complexity of classifiers SRC, LRC, CRC, and FSSP(1) [
68]:
Let us now compare the computational complexity of LRC and Resized+
classifiers. On the one hand, LRC is dominated by the solution of
c least square problems with
k unknowns and
n equations. Therefore, the computational complexity of the LRC method is
On the other hand, due to the noise-masking strategy, the computational complexity of the Resized+
classifier is
Thus, the Resized+
classifier is computationally cheaper than LRC. According to
Table 1, the Resized+
classifier is also computationally cheaper than the Resized+
classifier. Therefore, the Resized-
classifier is the cheapest among the classifiers based on resized versions of the face images.
The computational effort to encode a face image into a vector of length by a pre-trained neural network is fixed. Discarding the encoding phase, the computational complexity of the ResNet+ classifier depends only on comparisons between the encoded input and the encoded trained vectors, which results in . In contrast, the computational effort of ResNet+ is , which is certainly more expensive than that of both ResNet and ResNet+.
In conclusion, -based classifiers are competitive with state-of-the-art approaches because they exhibit a graceful balance between accuracy and computational cost.
8. Concluding Remarks
In this paper, we investigated max-
C and min-
D projection auto-associative fuzzy memories (max-
C and min-
D PAFMMs), which are defined on an algebraic structure called residuated complete lattice-ordered groupoid with left identity (R-clogli) and its dual version. Briefly, PAFMMs are non-distributed versions of the well-known max-
C and min-
D auto-associative fuzzy morphological memories (max-
C and min-
D AFMMs) [
5,
22]. Specifically, a PAFMM projects the input vector into the family of all max-
C (or min-
D) combinations of the stored items. Moreover, max-
C and min-
D PAFMMs are more robust to dilative or erosive noise than the AFMMs. In addition, PAFMMs are computationally cheaper than AFMMs if the number of stored items
k is less than their length
n.
Apart from a detailed discussion on PAFMM models, we focused on the particular model referred to as Zadeh’s max-
C PAFMM because it is obtained by considering Zadeh’s fuzzy inclusion measure. Zadeh’s max-
C and min-
D PAFMMs are the most robust PAFMMs to either dilative or erosive noise. On the downside, they are susceptible to mixed noise. In order to improve the noise tolerance of Zadeh’s PAFMMs to mixed noise, we proposed a variation of the noise masking strategy of Urcid and Ritter using a fuzzy similarity measure [
40].
Finally, experimental results using three famous face databases confirmed the potential application of Zadeh’s max-
C PAFMM for face recognition. Specifically, using Wilcoxson’s signed-rank test, we concluded that ResNet+
, which is based on ResNet encoding and Zadeh’s max-
C PAFMM classifier, outperformed important classifiers from the literature including ResNet [
80], LRC [
71], CRC [
72], and FSSP(1) [
68] classifiers for the recognition of undistorted face images. Furthermore, the experiments revealed that the Resized+
classifier performs as well as the FSSP(1) [
68] method but requires much less computational resources.
In the future, we intend to investigate further applications of max-C and min-D PAFMMs. In particular, we plan to study further the combinations of deep neural networks and these fuzzy associative memories. We also plan to generalize Zadeh’s PAFMMs to more general complete lattices.