1. Introduction
The goal of this paper is to provide an intuitive explanation of an important theory in the field of high-frequency applied electromagnetics, known as optimal sampling (or minimum redundancy sampling) of an electromagnetic field [
1,
2,
3]. The discussion is restricted to observation curves, while the case of observation planes is only briefly outlined.
The theory was initially developed in the context of antennas, with particular reference to near-field measurements. A key practical challenge in near-field measurements is to minimize the number of spatial samples required to achieve a target accuracy [
4]. Reducing the sample count can translate directly into shorter test times, lower operating costs, and higher throughput, without compromising measurement fidelity. In this context, the optimal sampling strategies introduced in [
1,
2,
3] offer a rigorous and effective solution [
5]. These methods rely on a robust mathematical framework to identify where, and at what density, to sample in order to meet a prescribed accuracy with the fewest possible measurements. This field of application has been subject of a large literature, with a large number of scientific papers, extensively reported in [
6].
However, the efficiency and effectiveness of the optimal sampling quickly extended its application far beyond antenna measurement, and nowadays applications of the optimal sampling theory cover almost all high-frequency electromagnetic disciplines. A review of the applications of the theory is reported in [
7].
Before entering into the details of the theory, it is essential to clarify what is meant by optimality of the sampling technique.
Loosely speaking, an electromagnetic problem can be recast as a process that goes from some a priori information on the radiating system to a posteriori information that reduces the set of possible solutions, ideally to a single element of the search space. The number of measurements represents the information gain and depends on the initial set, i.e., on the available a priori information. Adding more a priori information reduces the search space, thereby requiring less “information gain”. The main challenge is to identify suitable and reliable a priori information about the electromagnetic problem and to develop effective algorithms capable of exploiting such information.
The procedure proposed in [
1,
3] exploits simple, readily available a priori information about the source (e.g., position, shape) and the observation domain (e.g., shape, regularity characteristics) to shrink the feasible set of configurations, and a fast and highly efficient interpolation method based on Shannon–Whittaker–Kotelnikov series, using sinc-function bases. The method’s strength lies in the practicality of these assumptions—which makes the approach easy to deploy—in their effectiveness at reducing the number of required measurements to a value close to the practical minimum, and in the highly efficient interpolation formula. It is understood, however, that different a priori models induce different search spaces and may permit even fewer measurements; for example, in sparse-source settings the measurement burden can drop by orders of magnitude [
8]. This method exploits the sparsity of the source as a priori information, thereby drastically reducing the solution search space, but with a higher computational cost, requiring a convex minimization.
As previously noted, this paper presents an intuitive, practice-oriented explanation of the principle at the basis of this sampling technique. This explanation, rather than dwelling on full mathematical details, emphasizes a physical interpretation of the underlying theory.
Section 2 of the paper discusses the theoretical lower bound on the number of measurements in field representation by linear interpolation using a functional approach grounded in the results of [
2].
Section 3 presents the sampling representation of the field along a curve and summarizes the main results of optimal sampling theory without delving into mathematical details, for which interested readers are referred to [
1,
2,
3].
Section 4 develops an intuitive approach to sampling on curves. Some general indications regarding possible extensions to 2D surfaces are reported. Finally,
Section 5 concludes the paper.
2. Estimation of the Lower Bound of the Number of Samples in NF Measurements
The aim of this Section is to identify a lower bound for the number of measurements required to represent the field using linear representations of a source whose geometry and position in the space is known.
In order to identify the optimal representation in the mean–square–error sense (i.e., in the
norm), let us model the antenna under test (AUT) as a time-harmonic electromagnetic source placed in a bounded domain
D (see
Figure 1). The field, observed on the domain
(assumed not to intersect
D), is given by [
1].
where
is the source current density,
is the free-space dyadic Green’s function [
4], and “·” denotes the dyadic–vector product.
In the following discussion we suppose that the currents and the fields belong to separable Hilbert spaces equipped with the norm. The set of admissible currents is supposed to be bounded with unit radius.
The radiation operator can be diagonalized using the Hilbert–Schmidt expansion [
9,
10]. In particular, there exist non-negative singular values
, an orthonormal set of left singular functions
, and an orthonormal set of right singular functions
such that
where
and
denotes the
inner product.
A fundamental result of approximation theory states that truncating the series to
terms, with
equal to the number of singular values above
plus one, guarantees an approximation error not exceeding
. This number
is called the Number of Degrees of Freedom of the field at
level of approximation,
:
This number is also the minimum number of measurements required to represent the field within the required approximation using linear interpolation functionals. Summarizing, the Hilbert–Schmidt decomposition allows one to obtain both the optimal basis functions to represent the field on the measurement domain (i.e., the left singular functions) and the number of basis functions required to represent the field within a given approximation, which equals the number of singular values greater than .
In the asymptotic regime (
), where
is the free-space wavenumber and
a is a characteristic dimension of the radiating structure (e.g., its maximum length in the case of a linear source, or the radius in the case of a sphere), the decay of the singular values beyond the knee becomes extremely fast, and the number of terms in the series becomes essentially insensitive to the choice of
. This allows one to introduce the concept of the Number of Degrees of Freedom (NDF) independently of the specific value of
[
2].
In practical applications, the parameter is selected so that, with very high probability, the random noise affecting the measurements does not exceed . In general, throughout the paper we will assume that the asymptotic regime has been reached and we will use accordingly, resorting to only when strictly necessary.
It must be also noted that in real scenarios the effective number of basis functions
required to approximate the field with accuracy
in the Fresnel or Fraunhofer region is typically only slightly larger than the asymptotic number of degrees of freedom
[
1]. Analytical upper bounds on the representation error are provided in [
2].
In the following discussion, we analyze a single component of the current and the corresponding component of the electric field, thereby reducing the vector problem to a scalar one. This simplifies the mathematical expressions without losing any conceptually important detail.
3. Sampling the Field
In this section, the discussion will be focused on an observation curve, on which we have defined a curvilinear abscissa
s. Consequently, the points on the observation curve are defined by the position vector
(see
Figure 1). For the sake of notational simplicity, we will often indicate the position of the observation point simply with its abscissa
s. Furthermore, in all the subsequent discussion, we will assume that the observation curve lies outside the reactive zone of the radiating system.
Representation (
3) allows one to interpolate the field using
parameters. However, such parameters cannot be directly measured at single points in space, since they require computing the inner product between the field on the domain of observation and the singular functions.
A more useful series from an application point of view is the Shannon–Whittaker–Kotelnikov series, using sinc-function bases. Projecting the field onto the Paley–Wiener space, we obtain the
band-limited representation of the field:
where
,
,
, and
is the spatial bandwidth.
For practical applications, it is required to truncate the series to, for example, terms. However the field on the observation curve is not a band-limited function. This does not prevent representing the field within any prescribed accuracy—since the set of band-limited functions is dense in the set of fields on the observation curve—but it can greatly increase the number of basis functions required, often yielding a largely redundant representation.
To reduce the redundancy of the series, instead of the field itself, the authors of [
1,
2,
3] introduced an optimal sampling strategy, that takes advantage of some a priori information on the position and size of the radiating source to decrease the number of samples required to interpolate the field. Loosely speaking, the idea is to change the function to be projected onto the Paley–Wiener space. Instead of
, a “reduced” field is considered, obtained by extracting a suitable phase function
and introducing a suitable parameterization
of the observation curve:
This mapping is used to embed geometrical information into the sampling process.
The corresponding sampling series is
where
and
.
With a suitable choice of the functions
and
, the quantity
can be represented with far fewer samples than (
6) for a given maximum approximation error. Since reconstructing the field by inverting (
5) clearly requires the same number of samples as
, representation of the field using (
4) is more efficient than one obtained directly from field samples.
The functions
and
, which depend on the source and observation geometries, are obtained via an optimization procedure [
3]. This procedure determines the function
that minimizes the field’s local bandwidth—i.e., the bandwidth over a short segment around each observation point
s. It also finds the optimal parameterization
that keeps constant the local bandwidth, thereby avoiding unnecessary oversampling along portions of the curve. In short, instead of relying on the field’s global bandwidth, the method exploits local spatial spectral properties to achieve an accurate representation with the minimum total number of samples.
The theory provides analytical expressions for the main source geometries, which are very useful in practical applications. For example, with reference to the practically relevant case of a linear source and a linear observation curve, the functions are explicitly reported in [
11] (see
Figure 2), giving
where
and
denote the distances between the so-called extremal points of the source and the observation point
. In general, the extremal points are the points on the source surface where the lines originating from
are tangent to the surface.
In the present example, involving a linear source, they reduce to the endpoints of the segment, i.e., the points
. The mathematical details are reported in [
3].
In [
1] it is demonstrated that it is possible to approximate
F with a band-limited function within any degree of accuracy by considering a slightly larger bandwidth compared to a value
W called the effective bandwith, say
, where
is a band limitation enlargement factor. The approximation error between
F and its
band limited version rapidly tends to zero as
increases. A value of
around
is generally sufficient to ensure that the band limitation error is well below the noise level affecting the measured data. Analytic upper bounds on the representation error are available [
2], which reinforce the rigorous theoretical foundation of the optimal sampling representation. Experimental results reported in [
11] fully confirm the results of the theory.
In summary, the above theory enables a simple and effective representation of the electromagnetic field using a number of field samples N that is only slightly greater than the minimum number of samples required by any linear representation of the field, namely .
4. An Intuitive Physical Explanation of the Optimal Sampling Method
As a preliminary step, let us consider the classic Shannon–Whittaker–Kotelnikov series representation of a bandlimited signal in the time domain:
where
,
, and
B is the one-sided angular bandwidth, i.e., the highest angular frequency.
The condition , which corresponds to sampling exactly at the Nyquist rate, implies that the time between samples is such that the phase variation in the signal’s fastest harmonic is exactly . This observation has profound implications and highlights the fundamental importance of phase variation. To clarify, a fundamental parameter in bandlimited function theory is the number of degrees of freedom within an observation interval T, given by the time–bandwidth product . This quantity, that equals the number of samples required in that interval at Nyquist rate, matches the number of zeros, i.e., the number of -phase intervals, of the signal’s fastest harmonic over the same interval.
Starting from the above observations, it is possible to discuss the optimal sampling theory from a different point of view based on phase variation in the field.
Consider the field radiated by a convex source and observed at a point
on the observation curve, placed at some distance from the source (see
Figure 3). In particular, take two elementary radiators having equal amplitude located at
and
. In a highly simplified model, the field at
can be approximated as
where
and the dependence of
and
on the abscissa
s and on the two source points is understood and not explicitly reported.
By factoring the mean and difference distances, and approximate
at the denominator (slowly varying amplitude) we have
The approximation in the denominator is generally valid in the Fraunhofer region. In the present work, however, this expression is used in a somewhat relaxed manner as a pedagogical device to elucidate the physical mechanism underlying optimal sampling.
From the formula, we can note that among all pairs of source points, the pair that produces the fastest local phase variation near an observation point
is the one for which the distances from
attain, respectively, the local maximum and minimum values; in the following we will denote the two extreme distances by
and
(see
Figure 4). Accordingly, for each
s we identify the two paths that maximize the local spatial frequency of the interference. Note that this requires some a priori information on the position and geometry of the source, that are used to identify the “extremal points” with reference to
.
Equation (
15) applied to the extremal points identifies two distinct phase terms:
To reduce the phase variation associated with the mean path, we constrain the observation curve to satisfy , so that the residual (locally dominant) phase variation along s is governed solely by . In other words, constraining suppresses the rapidly varying common-mode phase contribution associated with the average propagation path, leaving only the differential phase , which varies more slowly with s and therefore governs the local sampling rate.
As noted earlier, any other pair of source points produces a local phase variation along s that is no faster than this. Loosely speaking, these further interferences ’populate’ the bandwidth within W, giving a continuous (spatial) spectrum.
Based on the observation that, in the Shannon–Whittaker–Kotelnikov series, the phase advance between adjacent sampling points equals
radians at the highest (spatial) harmonic, we partition the observation curve into segments such that
changes by
from the beginning to the end of each segment. This “
radians between samples” rule stems from the fact that, after removing the common-mode term
, the field along the curve can be written in the form
, where
This is a complex baseband representation, so the Nyquist condition corresponds to a phase increment of
between consecutive samples. Moreover, the expression of
in Equation (
9) makes this connection explicit, since
Hence, the local sampling step along
s is determined by the Nyquist relation
which directly links the
-spacing criterion to the normalized differential coordinate defined in Equation (
9). This method yields a sampling density consistent with the Nyquist bound and ensures that, in the worst case, the phase excursion between neighboring samples does not exceed
radians.
In essence, the two functions
and
identify the natural geometric quantities underpinning field sampling. At a fundamental level, this geometry is fixed by the time propagation of the electromagnetic field from the sources to the observation point, which in the harmonic domain corresponds to phase. Indeed, in the geometry shown in
Figure 4, the
curves correspond to
, and the
curves correspond to
when the effective bandwidth
W is normalized to one. Note that, in the optimal sampling scheme, if the observation curve does not match the natural spatial sampling geometry (i.e., is not a
curve), the value of
is subtracted from the field to obtain the reduced field (see Equation (
5)). In [
3], it is rigorously demonstrated that this is the best choice to compensate for the extra mean phase shift (i.e., the extra time delay in a time-domain interpretation) that arises on curves where
is not constant, while
Appendix A of this paper provides an intuitive explanation of this result.
To better clarify the connection between the rigorous theory and the heuristic approach, let us consider a linear source of length
(see
Figure 5) and the curve
in the
half-space (i.e., a semi-elliptical observation curve) intersecting the
x-axis at
. The points
where
, which define the heuristic sampling positions, are plotted in the same figure.
For comparison, the curves
, adopted in the optimal sampling theory, are also plotted as dashed lines. The figure confirms that the sampling points obtained from this partition of the observation curve coincide with the sampling positions of the optimal strategy, given by the intersection points between the
curves as defined in [
3] and the observation curve.
In
Figure 6, the field on the observation curve and the sampling points are plotted, considering an antenna excited by random-amplitude, random-phase currents. The corresponding phase along the observation curve is plotted in
Figure 7. In this figure, the phase of the field is normalized to 0 radians at each sampling position to visualize the phase variation in the interval between two adjacent points. The normalized phase is denoted by
in the figure. The same representation using
instead of
s would change the spacing between the samples, making all intervals equal in length.
The plot shows that the phase variation remains within in almost all intervals, except one. This indicates that the sampling points defined by are not sufficient to represent the field without aliasing.
This observation is consistent with the oversampling prescribed by the optimal sampling theory. In practice, choosing the criterion
is equivalent to applying the optimal sampling strategy with a nominal bandwidth
w equal to the effective bandwidth
W. In the rigorous theory,
W plays the role of a critical value: the approximation error remains large for
, whereas it becomes small for
. To obtain a sufficiently small band limitation error, the theory requires a bandwidth
with
, i.e., an oversampling with respect to a signal of bandwidth
W. The interpolation error decreases rapidly as
increases [
3].
In the heuristic approach, this is equivalent to placing the samples at
defined by
. As an example, in
Figure 8 a value
is used to select the sampling points along the curve. The resulting sampling positions are shown in
Figure 9. Due to the shorter spacing between samples, the phase variation in the field now remains within
in all intervals, allowing a field reconstruction with negligible aliasing error (see
Figure 10).
As a last observation, this physics-based approach is, at least in principle, suitable for extension to 2D observation surfaces. This would be of practical relevance, since sampling on surface observation domains is cumbersome and, to the best of the author’s knowledge, not yet completely solved. A simple strategy is to leverage the results obtained on curves by considering the coordinate curves on the surface [
1,
3]. This approach matches the probe movement in near-field antenna measurement systems. However, the resulting sampling grid tends to be redundant. In particular, sampling grids derived from optimal sampling along curves exhibit a redundancy factor of
with respect to the NDF [
12], i.e., they require about
times more samples than the theoretical minimum, to which one must add the oversampling required to control the band limitation error. This can be easily illustrated in the case of planar sources, for which the visible spectral region corresponds to a disk of area
. In the optimal sampling framework, however, the spectral support is a square of side
, i.e., an area
. The resulting spectral area is therefore larger by a factor
, and this excess is associated with non-propagating contributions.
A possible strategy to extend the 1D approach, i.e., sampling based on local phase variation, to a field observed on a 2D surface is to consider the phase gradient. The 2D analogue of the 1D adaptive interval is an unstructured sampling mesh in which the placement of samples (i.e., the mesh vertices) is governed directly by the local phase variation, keeping the phase variation within along each mesh edge. It must be pointed out that reconstructing a signal from an unstructured grid is numerically more complex than the standard 2D Shannon–Whittaker–Kotelnikov expansion.
A more direct extension of rigorous one-dimensional (1D) sampling theory to two-dimensional (2D) surfaces involves identifying a closed-form warping transformation. A warping transformation is a nonlinear change in variables that converts a non-uniformly bandlimited field into a uniformly bandlimited one, thereby enabling classical Shannon sampling on a uniform grid [
13]. Consequently, the transformation adopted for optimal sampling along curves can be interpreted as a warping transformation derived from the physical characteristics of the function being sampled, effectively unifying physics and sampling theory. Identifying a closed-form warping transformation of the electromagnetic field over 2D surfaces remains an active area of research [
14,
15].