Abstract
Infinite cyclic groups created by various objects belong to the class to the class basic algebraic structures. In this paper, we construct the infinite cyclic group of differential neurons which are modifications of artificial neurons in analogy to linear ordinary differential operators of the n-th order. We also describe some of their basic properties.
MSC:
06F15; 68T07; 11F25
1. Introduction
In our paper, we study artificial, or formal, neurons. Recall that these are the building blocks of mathematically modeled neural networks, e.g., [1]. The design and functionality of artificial neurons are derived from observations of biological neural networks. Our investigation belongs to the theory which is developed and applied in various directions contained in many publications, cf. [2,3,4,5,6]. The bodies of artificial neurons compute the sum of the weighted inputs and bias and “process” this sum with a transfer function, cf. [1,2,3,4,5,6,7,8,9,10].
In the next step, the information is passed via outputs (output functions). Thus, artificial neural networks have the structure similar to that of weighted directed graphs with artificial neurons being their nodes and connections between neuron inputs and outputs being directed edges with weights. Recall that in the framework of artificial neural networks there are networks of simple neurons called perceptrons. The basic concept (perceptron) was introduced by Rosenblatt in 1958. Perceptrons compute single outputs (the output function) from multiple real-valued inputs by forming a linear combination according to input weights, and then possibly putting the output through some nonlinear activation functions. Mathematically, this can be written as
where denotes the vector of time dependent weight functions, is the vector of time dependent (or time varying) input functions, b is the bias and is the activation function. The use of time varying functions as weights and inputs is a certain generalization of the classical concept of artificial neurons from the work of Warren McCulloch and Walter Pitts (1943); see also [1,2,3,4,5,6,7,8,9,10] and references mentioned therein.
2. Differential Neurons and Their Output Functions
In accordance with our previous papers [1,7,8,9], we regard the above mentioned artificial neurons such that inputs and weights will be functions of argument t belonging into a linearly ordered (tempus) set T with the least element As the index set we use the interval of real numbers , where denotes the set of all real numbers. So, denote by W the set of all non-negative functions forming a subsemiring of the ring of all real functions of one real variable . Denote by for , and the mapping
which will be called the artificial neuron with the bias in fact the output function of the corresponding neuron. By we denote the collection of all such artificial neurons.
Neurons are usually denoted by capital letters or . However, we use also notation , where is the vector of weights.
We suppose, for the sake of simplicity, that transfer functions (activation functions) , (or f) are the same for all neurons from the collection or that this function is the identity function
Now, similarly as in the case of the collection of linear differential operators, we will construct a cyclic group of artificial neurons, extending their monoid, cf. [1].
Denote by the so called Kronecker delta, , i.e., and , whenever
Suppose , , , , . Let be a such an integer that We define
where
and, of course, the neuron is defined as mapping Further, for a pair of neurons from we put
if , , and , and with the same bias.
Remark 1.
There exists a link between formal neurons and linear differential operators of the n-th order. This link is important for our future considerations. Recall the expression of formal neuron with inner potential where is the vector of inputs, is the vector of weights. Using the bias b of the considered neuron and the transfer function σ we can expressed the output as
Now consider a fundamental function where is an open interval; inputs are derived from the function as follows:
Further the bias As weights we use continuous functions ,
Then formula
is a description of the action of the neuron which will be called a formal (artificial) differential neuron. This approach allows to use solution spaces of corresponding linear differential equations.
3. Products and Powers of Differential Neurons
Suppose are fixed vectors of continuous functions and be the bias for any polynomial . We consider a differential neuron by the action
with the identity activation function According to the formula, we can calculate the output function of the differential neuron
Firstly, we describe the product of neurons , i.e., outputs of neurons
The vector of weights of the neuron is of the form where
Then the neuron is defined using its output function
In a greater detail:
Application of the above product onto the case of differential neurons: Suppose , are neurons with output functions
where , Denote
Then the output function of the neuron has the form
Now, using the above formula we can express output functions of powers , (for and (the neutral element-unit) of the infinite cyclic group The output function of the differential neuron is of the form
In the paper [1] the following theorem is proved:
Theorem 1.
Consider a differential neuron with the vector of time variable weights and the vector of inputs with polynomial , , and , The output function of the above mentioned neuron is of the form
with the bias Suppose Then the output function of the differential neuron has the form
Now, we discuss a certain type of subgroup which appears in all groups. The following text up to Proposition 2 incl. contains well-known facts, which are overtaken from the monography [11] (Chapter 2, §2,4).
Take any group G and any element Consider all powers of Define (the neutral element), and for define to be the product of k factors of (A little more properly, is defined inductively by declaring For define
Recall briefly some well-known classical facts.
Definition 1.
Let a be an element of a group The set of powers of a is a subgroup of called the cyclic subgroup generated by If there is an element such that one says that G is a cyclic group. We say that a is a generator of the cyclic group.
There are two possibilities for one possibility is that all the powers are distinct, in which case, of course, the subgroup is infinite; if this is so, we say that a has infinite order.
The other possibility is that two powers of a coincide, but this is not our case.
Definition 2.
The order of the cyclic subgroup generated by a is called the order of If the order of a is finite, then it is the least positive integer n such that
Proposition 1.
Let a be an element of a group
- (a)
- If a has infinite order then is isomorphic to
- (b)
- If a has finite order then is isomorphic to the group of n-th roots of
Proposition 2.
- (a)
- Any non-trivial subgroup of is cyclic and isomorphic to
- (b)
- Let be a finite cyclic group. Any subgroup of G is also cyclic.
For a construction of a cyclic group of artificial differential neurons we need to extend the cyclic monoid of differential neurons obtained in the paper [1] by negative powers of differential neurons, in particular to describe their output functions, so we need to construct negative powers of differential neurons which belong to the basic contribution of this paper. We suppose the existence of such inverse elements, i.e., negative powers of the generated element of the considered group.
In general, for the construction of the negative power with it seems to be a suitable way of a using of this equality:
where on the right hand side is given an arbitrary general differential neuron with the vector of time variable weight functions, with the vector of inputs
with a polynomial , , and , The neuron has the output function
with the bias However, we will construct the proof using mathematical induction—similarly as in [1]—the proof of the Theorem 1, which seems to be a more convenient way. So we are going to prove the following theorem.
Theorem 2.
Suppose the existence of an inverse elements (i.e., negative powers of the generated element of the considered group). Let be a differential neuron with the vector of time variable weights and with the vector of inputs with a polynomial , , and , , i.e., the output function of the neuron is of the form
with the bias Suppose Then the output function of the differential neuron has the form
or
Proof.
Consider the equality
where the output function of the neuron (the identity element of the monoid from [1]) is of the form
Let be the output function of the neuron with the bias and
be the output function of the neuron Since and for any we have
Moreover, which implies that the bias Thus, the output function is of the form
Using of Equation (16) we obtain after some simple calculation the expression:
This function is in a fact the output function of the neuron .
Now, for we obtain
which is in fact the Expression (24).
We have
which is the Equality (25) written for instead for The other negative powers can be also obtained from example we have. □
Using output functions of corresponding differential neurons we verify a validity of equalities
certifying that the neuron is the neutral element also for negative powers of the neuron
Denote by the output function of the neuron
Since the output function of the neuron (the unit element) has the form
we have
which is in fact the output function of the differential neuron In a similar way we can verify the second equality.
Remark 2.
In paper [12] there is defined a concept of a general n-hyperstructure as there follows:
Let be an arbitrary positive integer and be a system of non-empty sets. By a general n-hyperstructure we mean the pair
where is a mapping assigning to any n-tuple a non-empty subset Here means the power set of M without the empty set
Similarly as above, with this hyperoperation there is associated a mapping of power sets
defined by
This construction is also based on an idea of Nezhad and Hashemi for
At the end of this section we give this example:
Let be an open interval, be the ring (with respect to the usual addition and multiplication of functions) of all real functions with continuous derivatives up to the order including. Now, as in suppositions of Theorems 1 and 2, we consider a differential neuron with the vector of time variable weights and the vector of inputs with the polynomial , , and , The output function of the mentioned neuron is of the form
with the bias and In accordance with [13], we put
As above, we put whenever and Defining
for any n-tuple we obtain that
is a general n-hyperstructure for the polynomial
It is to be noted, that the used concept of investigated neurons is in a certain sense motivated by ordinary differential operators forming of left-hand sides of corresponding differential equations, see, e.g., [13,14].
Therefore, the construction of differential neurons consists of a certain modification of the concept of an artificial neuron which is investigated in a certain formal analogy to linear differential operators as mentioned above. Using the obtained cyclic group of differential neurons, we will construct a certain other hyperstructure of differential neurons. The mentioned relationship is in [8] described by the construction of a homomorphism.
It is to be noted that a hypergroup is a multistructure , where H is a non-empty set and is a mapping which is associative, i.e.,
for any triad where for , and Further, the reproduction axiom
for any element is satisfied.
The above definition of a hypergroup is in the sense of F. Marty.
Let be an open interval (bounded or unbounded) of real numbers, be the ring (with respect to usual addition and multiplication of functions) of all real functions with continuous derivatives up to the order including. We write instead of . For a positive integer we denote by the set of all linear homogeneous differential equations of the n-th order with continuous real coefficients on J,, i.e.,
(cf. [14,15,16]), where , for any (this is not an essential restriction). Denote the above defined linear operator defined by
and put
Further and stands for the Kronecker . For any we denote by the set of all linear differential operators of the n-th order , where for any , , (i.e., for each ). Using the vector notation we can write , i.e., a scalar product.
We define a binary operation and a binary relation on the set in this way:
For arbitrary pair we put where
and whenewer Evidently, is an ordered set.
In paper [14] there is presented the sketch of the proof of the following lemma:
Lemma 1.
The triad is an ordered (non-commutative) group.
4. Groups and Hypergroups of Artificial Neurons
As it is mentioned in the dissertation [2] neurons are the atoms of neural computation. Out of those simple computational units all neural networks are build up. For a pair of neurons from we put if and and with the same bias. Evidently is an ordered set. A relationship (compatibility) of the binary operation “·” and the ordering on is given by this assertion analogical to the above one. In paper [1] there is established that the structure is a non-commutative group.
Lemma 2.
The triad (algebraic structure with an ordering) is a non-commutative ordered group.
Sketch of the proof is presented in [8]. Denoting
we get the following assertion, the proof of which with necessary concepts is contained in [1].
Proposition 3.
Let Then for any positive integer and for any integer m such that the semigroup is an invariant subgroup of the group
If then a certain relationship between groups is contained in the following proposition:
Proposition 4.
Let and are integers such that Define a mapping by this rule: For an arbitrary neuron where we put with the action:
Then the mapping is a homomorphism of the group into the group
Consider and denote Denote There holds
where
Here and Then The neutral element is also mapped onto the neutral element of the group thus the mapping is a group homomorphism.
Now, using the construction described in Lemma 2, we obtain the final transpozition hypergroup (called also non-commutative join space). Denote by the power set of consisting of all nonempty subsets of the last set and define a binary hyperoperation
by the rule
for all pairs More in detail if , then if Then we have that is a non-commutative hypergroup. The above defined invariant (termed also normal) subgroup of the group is the carried set of a subhypergroup of the hypergroup and it has certain significant properties.
Using certain generalization of methods from [8] we obtain after investigation of constructed structures this result:
Let Then for any positive integer and for any integer m such that the hypergroup , where
is a transpozition hypergroup (i.e., a non-commutative join space) such that is its subhypergroup, which is
- -
- Invertible (i.e., implies and implies for all pairs of neurons
- -
- Closed (i.e., for all pairs
- -
- Reflexive (i.e., for any neuron and
- -
- Normal (i.e., for any neuron
Remark 3.
We can define a certain transformation function which mappes the output function into the output function . This function denoting by also determines the transformation of powers of corresponding differential neurons: In more detail, let us describe output functions and mentioned transformation function
Transformation function of the output function into the output function which determines the transformation of powers of corresponding differential neurons.
So,
Denoting
we can write
5. Conclusions
We have constructed the infinite cyclic group of differential neurons which is isomorphic to the cyclic group possessing the neuron as the identity element of Thus,
It is to be noted that the above constructed cyclic (infinite) group of artificial differential neurons can be also used for the construction of certain hyperstructures formed by such neurons [17,18,19,20]. So the above presented approach enables an additional elaboration of the hyperstructure theory ([8,9,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32]) in connection with time varying weights and with vectors of differentiable input functions.
The construction of the considered infinite cyclic group of differential neurons can be onto other its isomorphic images under the using other weights and inputs. After those constructions there is possible to create abelian finitely or infinitely generated groups of artificial differential neurons and to investigate their direct products or sums. Using a suitable ordering these considerations involve to obtain neural networks with prescribed structures.
Author Contributions
Investigation, J.C., B.S.; writing—original draft preparation, J.C., B.S.; writing—review and editing, J.C., B.S., J.V. All authors have read and agreed to the published version of the manuscript.
Funding
J.C. was supported by the FEKT-S-17-4225 grant of Brno University of Technology and J.V. was supported by the FEKT-S-20-6225 grant of Brno University of Technology.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Not applicable.
Acknowledgments
The authors would like to express their thanks to Dario Fasino and Domenico Freni.
Conflicts of Interest
The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.
References
- Chvalina, J.; Novák, M.; Smetana, B. Construction on an Infinite Cyclic Monoid of Differential Neurons. In Mathematics, Information Technologies and Applied Sciences 2021 Post-Conference Proceedings of Extended Versions of Selected Papers, Proceedings of the MITAV 2021, Brno, Czech Republic, 17–18 June 2021; Baštinec, J., Hrubý, M., Eds.; University of Defence: Brno, Czech Republic, 2021; pp. 1–10. [Google Scholar]
- Koskela, T. Neural Network Methods in Analysing and Modelling Time Varying Processes; Report B, 35; Helsinki University of Technology Laboratory of Computational Engineering Publications, Department of Electrical and Communications: Helsinki, Finland, 2003. [Google Scholar]
- Buchholz, S. A Theory of Neural Computation with Clifford-Algebras; Technical Report Number 0504; Christian-Albrechts-Universität zu Kiel, Institut für Informatik und Praktische Mathematik: Kiel, Germany, 2005. [Google Scholar]
- Tučková, J.; Šebesta, V. Data Mining Approach for Prosody Modelling by ANN in Text-to-Speech Synthesis. In Proceedings of the IASTED Inernational Conference on Artificial Intelligence and Applications—AIA 2001, Marbella, Spain, 4–7 September 2001; Hamza, M.H., Ed.; ACTA Press: Marbella, Spain, 2001; pp. 164–166. [Google Scholar]
- Volná, E. Neuronové Sítě 1. [Neural Networks], 2nd ed.; Ostravská Univerzita: Ostrava, Czech Republic, 2008. [Google Scholar]
- Waldron, M.B. Time varying neural networks. In Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, New Orleans, LA, USA, 4–7 November 1988. [Google Scholar]
- Chvalina, J.; Smetana, B. Models of Iterated Artificial Neurons. In Proceedings of the 18th Conference on Aplied Mathematics Aplimat 2019, Bratislava, Slovakia, 5–7 February 2019; pp. 203–212. [Google Scholar]
- Chvalina, J.; Smetana, B. Groups and Hypergroups of Artificial Neurons. In Proceedings of the 17th Conference on Aplied Mathematics Aplimat 2018, Bratislava, Slovakia, 6–8 February 2018; pp. 232–243. [Google Scholar]
- Chvalina, J.; Smetana, B. Solvability of certain groups of time varying artificial neurons. Ital. J. Pure Appl. Math. 2021, 45, 80–94. [Google Scholar]
- Pollock, D.; Waldron, M.B. Phase dependent output in a time varying neural net. In Proceedings of the Annual Conference on EMBS, Seattle, WA, USA, 9–12 November 1989; pp. 2054–2055. [Google Scholar]
- Goodman, F.M. Algebra: Abstract and Concrete; Prentice Hall: London, UK, 1998. [Google Scholar]
- Chvalina, J.; Hošková-Mayerová, Š.; Dehghan Nezhad, A. General actions of hypergroups and some applications. Analele Stiint. Univ. Ovidius Constanta 2013, 21, 59–82. [Google Scholar]
- Chvalina, J.; Chvalinová, L. Action of centralizer hypergroups of n-th order linear differential operators on rings on smooth functions. J. Appl. Math. 2008, 1, 45–53. [Google Scholar]
- Chvalina, J.; Chvalinová, L. Modelling of join spaces by n-th order linear ordinary differential operators. In Proceedings of the 4th International Conference APLIMAT 2005, Bratislava, Slovakia, 1–4 February 2005; pp. 279–284. [Google Scholar]
- Chvalina, J.; Novák, M.; Smetana, B.; Staněk, D. Sequences of Groups, Hypergroups and Automata of Linear Ordinary Differential Operators. Mathematics 2021, 9, 319. [Google Scholar] [CrossRef]
- Chvalina, J.; Novák, M.; Staněk, D. Sequences of groups and hypergroups of linear ordinary differential operators. Ital. J. Pure Appl. Math. 2019. accepted. [Google Scholar]
- Novák, M. n-ary hyperstructures constructed from binary quasi-ordered semigroups. Analele Stiint. Univ. Ovidius Constanta Ser. Mat. 2014, 22, 147–168. [Google Scholar] [CrossRef] [Green Version]
- Novák, M. On EL-semihypergroups. Eur. J. Comb. 2015, 44, 274–286. [Google Scholar] [CrossRef]
- Novák, M. Some basic properties of EL-hyperstructures. Eur. J. Comb. 2013, 34, 446–459. [Google Scholar] [CrossRef]
- Cristea, I.; Novák, M.; Křehlík, Š. A class of hyperlattices induced by quasi-ordered semigroups. In Proceedings of the 16th Conference on Aplied Mathematics Aplimat 2017, Bratislava, Slovakia, 31 January–2 February 2017; pp. 1124–1135. [Google Scholar]
- Corsini, P. Prolegomena of Hypergroup Theory; Aviani Editore Tricesimo: Udine, Italy, 1993. [Google Scholar]
- Corsini, P.; Leoreanu, V. Applications of Hyperstructure Theory; Kluwer: Dordrecht, The Netherlands; Boston, MA, USA; London, UK, 2003. [Google Scholar]
- Cristea, I. Several aspects on the hypergroups associated with n-ary relations. Analele Stiint. Univ. Ovidius Constanta 2009, 17, 99–110. [Google Scholar]
- Cristea, I.; Ştefănescu, M. Binary relations and reduced hypergroups. Discrete Math. 2008, 308, 3537–3544. [Google Scholar] [CrossRef] [Green Version]
- Cristea, I.; Ştefănescu, M. Hypergroups and n-ary relations. Eur. J. Combin. 2010, 31, 780–789. [Google Scholar] [CrossRef] [Green Version]
- Leoreanu-Fotea, V.; Ciurea, C.D. On a P-hypergroup. J. Basic Sci. 2008, 4, 75–79. [Google Scholar]
- Račková, P. Hypergroups of symmetric matrices. In Proceedings of the 10th International Congress of Algebraic Hyperstructures and Applications, (AHA), Brno, Czech Republic, 3–9 September 2008; pp. 267–272. [Google Scholar]
- Vougiouklis, T. Hyperstructures and their Representations; Hadronic Press: Palm Harbor, FL, USA, 1994. [Google Scholar]
- Novák, M.; Cristea, I. Composition in EL–hyperstructures. Hacet. J. Math. Stat. 2019, 48, 45–58. [Google Scholar] [CrossRef]
- Vougiouklis, T. Cyclicity in a special class of hypergroups. Acta Univ. Carol. Math. Phys. 1981, 22, 3–6. [Google Scholar]
- Chvalina, J.; Svoboda, Z. Sandwich semigroups of solutions of certain functional equations and hyperstructures determined by sandwiches of functions. J. Appl. Math. 2009, 2, 35–43. [Google Scholar]
- Borzooei, R.A.; Varasteh, H.R.; Hasankhani, A. F-Multiautomata on Join Spaces Induced by Differential Operators. Appl. Math. 2014, 5, 1386–1391. [Google Scholar] [CrossRef] [Green Version]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).