On the Duality of Regular and Local Functions

In this paper, we relate Poisson’s summation formula to Heisenberg’s uncertainty 1 principle. They both express Fourier dualities within the space of tempered distributions and 2 these dualities are furthermore the inverses of one another. While Poisson’s summation 3 formula expresses a duality between discretization and periodization, Heisenberg’s 4 uncertainty principle expresses a duality between regularization and localization. We define 5 regularization and localization on generalized functions and show that the Fourier transform 6 of regular functions are local functions and, vice versa, the Fourier transform of local 7 functions are regular functions. 8 9


Introduction
Regularization is a popular trick in applied mathematics; see [1] for example.It is the technique "to approximate functions by more differentiable ones" [2].Moreover, its terminology coincides with the terminology used in generalized function spaces.They contain two kinds of functions: "regular functions" and "generalized functions".While regular functions are functions in the ordinary functions sense which are smooth (infinitely differentiable) in the ordinary functions sense, all other functions become smooth (infinitely differentiable) in the "generalized functions sense" [3].In this way, all functions are smooth.Localization, in contrast to that, is another popular technique.It allows, for example, to integrate functions which could not be integrated otherwise, if we think of "locally integrable" functions or if we think of the Short-Time Fourier Transform (STFT), capable of analyzing infinitely extended signals.
Although regularization (Figure 1) and localization (Figure 2) appear to be quite different, a connection between these operations is no surprise.It is rather ubiquitous in the literature.However, the theorem below (Figure 3) appears in a wider sense.It even holds within the space of tempered distributions and is directly related to Heisenberg's uncertainty principle.It is moreover the inverse of an already known discretization-periodization duality (Poisson's Summation Formula).While discretization and periodization map functions from smoothness towards discreteness in both time and frequency domain, the operations investigated in this study, regularization and localization, map functions from discreteness towards smoothness in both time and frequency domain (Figure 4 and Figure 5).In this way, we have a full range of tools (discretization, periodization, regularization, localization) in order to describe any transition from smoothness towards discreteness and also back from discreteness towards smoothness in a mathematically rigorous and formally correct way.
This paper is organized as follows.Section 2 describes the overall motivation for this study and Section 3 explains the rough proof idea.Section 4 provides an introduction to the notations used and previous results.Section 5 presents a justification for Section 6 where regularization and localization Mathematics 2017, 5, 41 2 of 14 within the space of tempered distributions are defined.Section 7 provides symbolic calculation rules based on these definitions, needed to prove the theorem in Section 8. Section 9 connects these results to results in a previous study and Section 10, finally, concludes this study and provides an outlook.

Motivation
In contrast to more practical applications of regularizations and localizations, the actual intention of this study is to provide universally valid calculation rules for regularizations and localizations together with the scope of their validity within a preferably most general setting.In this way, regularizations can be expressed in terms of localizations and, vice versa, localizations can be expressed in terms of regularizations, i.e., poorly converging formulas can be replaced by better converging ones.Another important application of our rules is knowing the scope of their validity.Knowing not only a formula but also its validity scope is especially important in applied mathematics.It is simply not very reasonable to implement a formula that will never converge.

Generalized Functions
The setting of generalized functions, also called distributions, is today a most modern, very general setting in Fourier analysis [3][4][5][6][7][8][9][10][11].It meets the requirements of Fourier analysis in both theory and practice.It allows, for example, to Fourier transform functions which are not even Lebesgue integrable, such as the Dirac impulse δ or even constants such as the function that is constantly 1.
In this study, we prove simple symbolic calculation rules within S , the space of tempered distributions, for handling both, regularizations and localizations, and their interactions.The space of tempered distributions S is known to be an ideal setting in Fourier analysis, mainly due to three facts.First, the Fourier transform is understood in a much wider sense than usually (on Lebesgue square-integrable functions); secondly, the Fourier transform is still an automorphism in S (as on Lebesgue square-integrable functions), i.e., the Fourier transform of any element in S is again an element in S ; thirdly, all functions are smooth, i.e., all functions can be derived infinitely often.Smoothness is especially required to treat Fourier series (and the Fourier transform as their limit).They consist of complex exponentials which are functions that can be derived infinitely often.Furthermore, recall that Lebesgue square-integrable functions (and their Fourier transforms accordingly) must decrease to zero as their arguments increase, otherwise they would not be Lebesgue integrable.This very strong restriction has, in fact, never been accepted by those who are working with Fourier transforms practically.It actually led to the acceptance of S as a standard space in Fourier analysis throughout many scientific communities; beside mathematics and engineering, also in physics, especially in quantum mechanics [12][13][14][15].In S , in contrast to traditional approaches, we may even allow functions to grow up to the order of polynomials as their arguments increase and they can still be sampled and/or Fourier transformed.

Symbolic Calculation
The overall goal in this and in our previous study [16] is to compile a number of rules including their validity scope and therewith to fill a toolbox that will enable us to symbolically calculate with operations of discretization (sampling), periodization (replicating), regularization (smoothing) and localization (windowing).Performing such calculations on a symbolic level has several advantages.We are now able to rigorously prove higher-level theorems in S ; we may replace more complicated formulas by simpler ones (or by better converging ones) prior to implementing them in software and, lastly, we may even implement these rules in symbolic calculation environments such as in Wolfram Mathematica [17] or in Python SymPy [18] in order to exclude human errors in formula derivations.Some applications of our rules [19] are, for example, extending the scope of validity of Poisson's Summation Formula to a validity on tempered distributions [16], proving that four differently defined Fourier transforms (more precisely, the Integral Fourier Transform, Discrete-Time Fourier Transform (DTFT), Discrete Fourier Transform (DFT) and the Integral Fourier Transform for periodic functions) coincide in the distributional sense [20][21][22] or deriving a digital radar image from the continuous case (the landscape being imaged) via operations of localization (windowing), discretization (sampling), convolution (defocusing) and deconvolution (focusing) [23].

Idea
The proof technique in generalized function spaces may appear a bit unusual to those who are practically oriented.For any equality f = g in generalized function spaces X where f , g ∈ X , it must be shown that both, f and g, yield the same result if applied to arbitrarily chosen test functions ϕ ∈ X taken from their test function space X.A very short introduction to this technique is given in the section Distribution Theory in our previous paper [16].
However, the argumentation in this paper is much simpler.We do not need to prove any results on test functions because it has already been done by other authors [24][25][26][27].Instead, we simply argue with subspaces of S .We implicitly use the fact that functions in subspaces in S inherit all function properties from the spaces where they are embedded.This idea is one of the fundamental tricks in distribution theory.A full range of distribution space embeddings can be found in [24], p. 170.However, only a few of these spaces are needed here (see section Notation and Figure 4 in [16]).We need the space of slowly growing regular functions O M ⊂ S (where "regular" means smooth functions in the ordinary functions sense), the space of rapidly decreasing generalized functions O C ⊂ S (where "generalized" means that they are smooth but not necessarily "regular") and the Schwartz space S ⊂ S which is the space of regular, rapidly decreasing functions, i.e., functions in O M ∩ O C = S are "regular" and "local" (rapidly decreasing).They inherit "regular" from O M and "local" from O C .Another important fact is that S contains the space D ⊂ S which is the space of compactly supported smooth functions.So, in the special case of having some ϕ ∈ D, the property of ϕ being "local" turns into ϕ being "finite", i.e., ϕ is zero outside some given interval.Accordingly, ϕ • f for some f ∈ S will be zero outside the given interval.In this case, we talk about "finitization" as a special form of "localization".A concatenation of discretization and finitization is of particular importance, especially for turning integrals into finite sums.
Moreover, we make intensive use of the following dualities of the Fourier transform: F (S) = S and hence F (S ) = S as well as F (O M ) = O C and F (O C ) = O M .The first duality means that the Fourier transform maintains the function property of being "regular and local", the second duality means that "tempered" functions (slowly growing functions, i.e., functions that do not grow faster than polynomials) are again "tempered" after Fourier transforming them and the latter duality (Theorem 3, p. 424 in [26]) means that the Fourier transform of slowly growing ("regular") functions are rapidly decreasing ("local") functions and, vice versa, the Fourier transform of rapidly decreasing ("local") functions are slowly growing ("regular") functions.This simple circumstance resulted in a validity statement for the Poisson Summation Formula in S , see Theorem 1 in [16]-the widest comprehension of the Poisson Summation Formula today to the best of this author's knowledge.
Another issue in generalized function spaces, hence also in S , is the fact that arbitrary generalized functions cannot be multiplied with each other.Accordingly, arbitrary generalized functions cannot be convolved with each other.We evade this problem in S by using Laurent Schwartz' [24] subspace of multiplication operators O M and his space of convolution operators O C as introduced above.These spaces provide a secure footing for multiplications and convolutions in S , see Theorem 3, p. 424 in [26].Recall that O stands for Opérateurs (operators), M for Multiplication, C for Convolution, an apostrophe as X indicates spaces of "generalized" functions, no apostrophe as X indicates spaces of "regular" functions.Elements from X can always be applied to elements of X.

Preliminaries
Let δ kT be the Dirac impulse shifted by k ∈ Z n units of T ∈ R n + = {t ∈ R n : 0 < t ν < ∞, 1 ν n}, kT being component-wise multiplication, within the space S ≡ S (R n ) of tempered distributions (generalized functions that do not grow faster than polynomials) and let δ kT be the Dirac comb.Then, δ kT ∈ S and III T ∈ S for any T ∈ R n + are tempered distributions [8,24,25].We shortly write δ instead of δ kT if k = 0.The Fourier transform F in S is defined as usual and such that F1 = δ and F δ = 1 where 1 is the function being constantly one [3][4][5][6][7]9,11,24].The Dirac comb is moreover known for its excellent discretization (sampling) and periodization (replicating) properties [4,9,11,28].While multiplication III T • f in S samples a function f ∈ S , the corresponding convolution product III T * f periodizes f in S .
The following two lemmas summarize the demands that must be put on f ∈ S such that f can be sampled and/or periodized in S .Recall that smoothness, i.e., infinite differentiability, is not a demand.It is a given fact for all functions in generalized function spaces.Also recall that O M is the space of multiplication operators in S and O C is the space of convolution operators in S according to Laurent Schwartz' theory of distributions [14,16,[24][25][26][27][29][30][31].
Proof.Any uniform discretization (sampling) in S corresponds to forming the product where III T ∈ S is the Dirac comb.Furthermore, III T is not a regular function, i.e., III T ∈ S \ O M .However, for any multiplication product in S , it is required that at least one of the two factors is in O M .Otherwise, the existence of this product cannot be secured.Hence, f ∈ O M ⊂ S or another reasonable definition of this multiplication product would be required.Vice versa, if f ∈ O M then III T • f exists due to S • O M ⊂ S (Theorem 3, p. 424 in [26]).
One may observe that even "if and only if" holds in Lemma 1 if the operation • is not redefined (in the sense that at least one of the two factors converges in any sense).An equivalent statement is the following lemma.

Lemma 2 (Periodization
Proof.Any periodization in S corresponds to forming the convolution product where III T ∈ S is the Dirac comb.Furthermore, III T is of no rapid descent, i.e., III T ∈ S \ O C .However, for any convolution product in S , it is required that at least one of the two factors is in O C .Otherwise, the existence of this product cannot be secured.Hence, f ∈ O C ⊂ S or another reasonable definition of this convolution product would be required.Vice versa, if f ∈ O C , then III T * f exists due to S * O C ⊂ S (Theorem 3, p. 424 in [26]).
In a previous study [16], we used these insights in order to define operations of discretization ⊥⊥⊥ T and periodization T in S .While discretization is an operation Starting from these two definitions, we proved that hold in S , both being expressions of Poisson's Summation Formula.We shortly write ⊥⊥⊥ and instead of ⊥⊥⊥ T and T if T ν = 1 for all 1 ν n.Moreover, recall that these rules are a consequence of the Fourier duality for any α ∈ O M , g ∈ O C and f ∈ S which is, according to Laurent Schwartz' theory of generalized functions, the widest possible comprehension of both multiplication and convolution within the space of tempered distributions [24,26,30].It lies at the very heart of S .Many calculation rules in S , including Equations ( 1), ( 2), ( 7), ( 8), ( 11), ( 12) and Lemmas 1, 2, 3, 4 rely on it.

Feasibilities
The following two lemmas provide justifications for the way we will define regularizations and localizations in S below.They will allow us to invert discretizations and periodizations in S .Lemma 3 (Regularization).Let ϕ ∈ S.Then, for any f ∈ S , ϕ * f can be sampled.
Proof.This is a consequence of the fact that S * S ⊂ O M [7,24,26,30] and Lemma 1.
An equivalent statement is the following lemma.It is interesting to observe that ϕ * and ϕ • stretch and compress f ∈ S , respectively.Moreover, this property is independent of the actual choice of ϕ ∈ S. It can therefore be attributed to the operations of convolution and multiplication themselves.

Definitions
"One of the main applications of convolution is the regularization of a distribution" [26] or the regularization of ordinary functions which are not infinitely differentiable in the conventional functions sense.Its actual importance lies in the fact that it is the reversal of discretization.
Regularization (see Figure 1) is usually understood as the approximation of generalized functions via approximate identities [2,[5][6][7]26,[32][33][34].In this paper, however, we extend this idea by allowing any ϕ ∈ S and also any f ∈ S , i.e., even ordinary functions (either smooth or not smooth) can be regularized.This approach naturally includes the special case of choosing approximate identities without unnecessarily restricting our theorem below.
Lemmas 3 and 4 above justify the following two definitions.
Definition 1 (Regularization).Let ϕ ∈ S.Then, for any tempered distribution f ∈ S we define another tempered distribution by which is common to all tempered distributions but add the regularity of ϕ to f .It follows that regularized functions can always be sampled according to Lemma 1.
Regularizations of type (5), where f as a member in some distribution space X ′ is singular (i.e., there is no locally integrable function representing f ) and ϕ as a member in its corresponding space of test functions X is used for its regularization, are usually called "singular convolution" [34,35] and with f replaced by a sequence f ǫ converging towards f they become so-called "discrete singular convolutions (DSCs)", a standard technique today for the regularization of singular distributions.
Away from the generalized functions literature, we furthermore encounter regularizations in terms of "smoothings", "interpolations", "zero-paddings" or "approximations" because they are not only applied to generalized functions, they are also applied to ordinary functions, usually to obtain better "regularity" properties for functions, i.e. better differentiability.Regularity is also a topic discussed in [51], for example.It is closely related to localization.
Definition 6 (Localization).Let ϕ ∈ S. Then for any tempered distribution f ∈ S ′ we define another tempered distribution by  Regular functions (according to this definition) are functions in the ordinary functions sense which are smooth (infinitely differentiable) in the ordinary functions sense, a function property that is of immense value in many branches of mathematics.Regular functions belong to O M because S * S ⊂ O M , see e.g., [7,24,30].They maintain the 'tempered' property, i.e., they do not grow faster than polynomials, which is common to all tempered distributions but add the regularity of ϕ to f .It follows that regularized functions can always be sampled according to Lemma 1.
Regularizations of type (5), where f as a member in some distribution space X is singular (i.e., there is no locally integrable function representing f ) and ϕ as a member in its corresponding space of test functions X is used for its regularization, are usually called "singular convolution" [35,36] and with f replaced by a sequence f converging towards f , they become so-called "discrete singular convolutions (DSCs)", a standard technique today for the regularization of singular distributions.

Calculation Rules
"One of the basic principles in classical Fourier analysis is the impossibility to find a function f being arbitrarily well localized together with its Fourier transform Ff " [65].This, in particular, can easily be seen if one tries to localize the function that is constantly 1.
Proof.According to (4), δ ∈ O C ′ can be convolved with ϕ ∈ S ⊂ S ′ and, equivalently, 1 ∈ O M can be multiplied with ψ ∈ S ⊂ S ′ , hence holds in S ′ .The second formula is shown in an analogous manner.
It is moreover interesting to observe that in analogy to the Dirac comb identity [15] △ △ △δ ≡ III ≡ ⊥ ⊥ ⊥1 Local functions (according to this definition) belong to O C because S • S ⊂ O C [24,26].They add the 'rapid descent' property of Schwartz functions ϕ ∈ S to f ∈ S .It follows that localized functions can always be periodized according to Lemma 2.

Calculation Rules
"One of the basic principles in classical Fourier analysis is the impossibility to find a function f that is arbitrarily well localized together with its Fourier transform F f " [66].This, in particular, can easily be seen if one tries to localize the function that is constantly 1.

Lemma 5 (Localization Balance). Let ϕ ∈ S and let
In (8), we see that by localizing 1, we delocalize δ, i.e., 1 and its Fourier transform δ cannot be both arbitrarily well localized.This phenomenon is known as Heisenberg's uncertainty principle [6,7,11,[66][67][68]. Vice versa, in (7), we see that by regularizing δ we increasingly deregularize 1.The entity ∩ ϕ δ is also known as an "approximate identity" of δ, usually denoted as δ where is a parameter describing the proximity to δ (see e.g., [26] p. 316, p. 401 or [32] p. 5).Convolving any f ∈ S with δ , it creates an approximate identity f of f which is a function in the ordinary sense being infinitely differentiable.
Proof.According to (4), δ ∈ O C can be convolved with ϕ ∈ S ⊂ S and, equivalently, 1 ∈ O M can be multiplied with φ ∈ S ⊂ S , hence holds in S .The second formula is shown in an analogous manner.
It is also interesting to observe that in analogy to the Dirac comb identity [16] δ ≡ III ≡ ⊥⊥⊥1 the following identity, let's say a "localization balance" holds a balance in S if ϕ ≡ φ is satisfied for ϕ ∈ S, which obviously is the best achievable compromise in localizing 1 and thereby delocalizing δ.It is true for the Gaussian Ω(t) ≡ e −πt 2 and herewith expresses Hardy's uncertainty principle [69].However, it is also true for the Hyperbolic secant Ω(t) ≡ 2/(e t + e −t ), see e.g., [11], and for every fourth Hermite function H, i.e., all H satisfying F H ≡ H.A connection between Gaussians and Hyperbolic Secants is that both belong to a class of "Pólya frequency functions" [70,71].Gaussians, Hyperbolic Secants and Hermite functions are treated in [72,73] for example.Hermite polynomials are also known for their very important role in distribution theory [13,35,36,43].Hyperbolic Secants may also replace Gaussians in Gabor systems; see e.g., Janssen and Strohmer [74].A link between Gaussians and Hyperbolic Secants is also known in soliton physics where the "initial Gaussian beam reshapes to a squared hyperbolic secant profile" [75].Studying fixpoints Ω of the Fourier transform in S is therefore a worthwhile goal.
Another calculation rule that we need to prove for the theorem below is the following.It holds in analogy to already shown properties of discretizations and periodizations [16].
Proof.We may allow that at most one of the operands in ϕ * g * f is not an element in O C .This is indeed true as ϕ ∈ S ⊂ O C , g ∈ O C and f is an arbitrary element in S .It follows that ϕ * g * f exists in S and, hence, operands may be interchanged arbitrarily.Using O C * S ⊂ S twice and ( 5), we obtain g * ( in S .The other half of this equation results from the fact that the roles of f and g can be exchanged due to commutativity.The second formula is then shown in an analogous manner.

A Regularization-Localization Duality
The interaction between regularizations and localizations is ubiquitous in the literature today, for example as "regularization" and multiplication with smooth "cutoff functions" in Hörmander [2], as "two components of the approximation procedure" in S , see Strichartz [8], or as "approximation by cutting and regularizing" in Trèves [27], p. 302, or as "cutting out" one period of f and applying "(quasi-)interpolation" [63] or in terms of "zero-padding in the time domain corresponds to interpolation in the frequency domain" [76].Detailed studies of the interaction of both regularizations and localizations, can be found for example in [49,53,54,77] and in engineering literature, we encounter these interactions in terms of the interplay between "windowing" on one hand and "interpolation" on the other.An equivalent is the so-called "zero-padding" technique (e.g. in [76,[78][79][80][81][82][83]) found in engineering textbooks as a way to implement interpolations.It corresponds to the regularization of a discrete function by embedding it into a higher-dimensional space where it is smooth.However, we may summarize this regularization-localization duality in the following way.
Theorem 1 (Regularization vs. Localization).Let ϕ ∈ S, f ∈ S and let φ := F ϕ. Then So, this duality asserts that regularizing a function means to localize its Fourier transform and, vice versa, localizing a function means to regularize its Fourier transform.It is the one-to-one counterpart of a discretization-periodization duality in S , given in (1) and (2).
Proof.Formally, according to the calculation rules shown above, the following equalities hold in S .We start using f = δ * f where δ ∈ O C ⊂ S is the identity element with respect to convolution in S .Then, we apply Equations ( 10), ( 4), (7) and (9), in this order.Finally, with F f = g ∈ S , we use g = 1 • g where 1 ∈ O M ⊂ S is the identity element with respect to multiplication in S .The second formula is now shown in an analogous manner.
So, this duality asserts that regularizing a function means to localize its Fourier transform and, vice 253 versa, localizing a function means to regularize its Fourier transform.It is the one-to-one counterpart of 254 a discretization-periodization duality in S ′ , given in ( 1) and ( 2).

255
Proof.Formally, according to the calculation rules shown above the following equalities hold in S ′ .We start using f = δ * f where δ ∈ O C ′ ⊂ S ′ is the identity element with respect to convolution 256 in S ′ .Then we apply Equations ( 10), ( 4), ( 7) and (9), in this order.Finally, with Ff = g ∈ S ′ we use 259  An immediate consequence of the theorem is that f and its Fourier transform F f cannot both be arbitrarily well localized, a fact that is known as Heisenberg's uncertainty principle.Also note that F loc , as shown in Figure 3, is the Short-Time Fourier Transform (STFT) with window function ϕ ∈ S and it is the Gabor transform if ϕ is a Gaussian.Consequently, the result of Gabor transforms will be smooth, i.e., they cannot be discrete for example.In contrast, its Fourier dual, the Fourier transform of regular functions F reg , as shown in Figure 3, corresponds first to regularizing functions before Fourier transforming them.Consequently, the result of such transforms will be local, i.e., they cannot be periodic for example.
Obviously, by looking at these interactions, one may think of discrete functions as the 'opposite' of regular functions and, equivalently, one may think of periodic functions as the 'opposite' of local functions.This is examined more closely in the next section.

Four Subspaces
Let O M be the complement of regular functions O M in S .It is the space of all ordinary or generalized functions in S which are not infinitely differentiable in the ordinary functions sense.Furthermore, let O C be the complement of local functions O C in S .It consists of all ordinary or generalized functions in S which either do not fade to zero as |t| increases (periodic functions for example) or they fall to zero but too slowly (with polynomial decay rather than with exponential decay).
Then, the following diagram (Figure 4) holds in S .This diagram can also be expressed in a slightly more "human readable" form (Figure 5) by recalling that O M are "regular functions" and O M are "generalized functions" (in the sense that they are not "regular") and O C are "local functions" and O C are "global functions" (in the sense that they are not "local").This diagram can moreover be expressed a little bit more "human readable" by recalling that O M are "regular functions" and ∁ O M are "generalized functions" (in the sense that they are not "regular") and O C ′ are "local functions" and ∁ O C ′ are "global functions" (in the sense that they are not "local").Apparently, the Schwartz space S ≡ O M ∩ O C ′ , the "smooth world", in some sense, is the 'opposite' of ∁ O M ∩ ∁ O C ′ , the "discrete world".One may also note that no additional information is used yet beside pure operator definitions.There is also no statement yet on the reversibility of our operations ⊥ ⊥ ⊥ and △ △ △ and ∩ ϕ and ⊓ ϕ in S ′ .Such inversions will be treated from a more quantitative point of view in a follow-on study.

Conclusions and Outlook
It is shown that in analogy to a discretization-periodization duality in S ′ there is also a regularization-localization duality in S ′ .Proving these dualities even follows the same pattern.In addition, the two dualities are inverses of each other in the sense that the first one maps towards Apparently, the Schwartz space S ≡ O M ∩ O C , the "smooth world", in some sense, is the 'opposite' of O M ∩ O C , the "discrete world".One may also note that no additional information is used yet beside pure operator definitions.There is also no statement yet on the reversibility of our operations ⊥⊥⊥ and and ∩ ϕ and φ in S .Such inversions will be treated from a more quantitative point of view in a follow-on study.

Conclusions and Outlook
It is shown that in analogy to a discretization-periodization duality in S , there is also a regularization-localization duality in S .Proving these dualities even follows the same pattern.In addition, the two dualities are inverse of each other in the sense that the first one maps towards discreteness and the latter one maps towards smoothness.In total, we derived several calculation rules that are suitable to symbolically calculate with operations of discretization, periodization, regularization and localization in order to describe transitions from smoothness towards discreteness, even finite discreteness, and back from discreteness towards smoothness in a mathematically rigorous and formally correct way.We may replace, for example, discretizations by periodizations or regularizations by localizations whenever it is advantageous to do so.Our rules can, for example, be used to derive higher-level theorems in S .They can also be implemented in symbolic calculation environments such as Wolfram Mathematica or Python SymPy and thereby become a very useful toolbox for algorithm design.

Lemma 4 (
Localization).Let ϕ ∈ S. Then for any f ∈ S , ϕ • f can be periodized.Proof.It follows from the fact that S • S ⊂ O C [24,30], which is the Fourier dual F (S * S ) = F (O M ) of S * S ⊂ O M , and Lemma 2.

)
which is a generalized function of rapid descent in O C ′ ⊂ S ′ .The operation ⊓ ϕ is called localization or restriction of f by means of ϕ.It is a linear continuous operation ⊓ ϕ : S ′ → O C ′ , f → ⊓ ϕf .If ϕ ∈ D ⊂ S, it is also called finitization.The result of ⊓ ϕ is called local function of f in S ′ .

Figure 1 .
Figure 1.A regularization of generalized function f yields regular function ∩ ϕ f .

)Figure 2 .
Figure 2. The localization of generalized function f yields local function ⊓ ϕf .Local functions (according to this definition) belong to O C ′ because S ′ • S ⊂ O C ′ [23,25].They add the 'rapid descent' property of Schwartz functions ϕ ∈ S to f ∈ S ′ .It follows that localized functions can always be periodized according to Lemma 2.

Figure 2 .
Figure 2. A localization of generalized function f yields local function ϕ f .

257 g = 1 •
g where 1 ∈ O M ⊂ S ′ is the identity element with respect to multiplication in S ′ .The second 258 formula is now shown in an analogous manner.

Figure 5 .
Figure 5.The same diagram as above, drawn in another fashion.

Figure 5 .
Figure 5.The same diagram as above, drawn in another fashion.
)which is a regular, slowly growing function in O M ⊂ S .The operation ∩ ϕ is called regularization, approximation, interpolation or smoothing of f by means of ϕ.It is a linear continuous operation ∩ ϕ : S → O M , f → ∩ ϕ f .The result of ∩ ϕ is called regular function of f in S .The regularization of generalized function f yields regular function ∩ ϕf .Regular functions (according to this definition) are functions in the ordinary functions sense which are smooth (infinitely differentiable) in the ordinary functions sense, a function property that is of immense value in many branches of mathematics.Regular functions belong to O M because S * S ′ ⊂ O M , see e.g.