1. Introduction
Until recently, if a scientist or an engineer were asked what tensors are good for, probably in many cases they would have organized their answer around examples such as the Cauchy stress tensor or the tensors that arise in general relativity. However, these are examples of a more complex concept: that of a
tensor field. To give a first explanation of the difference, let us point out that a scalar field is a function defined on a portion of space (on which one can perform the various operations of Calculus, e.g., take the partial derivatives), whereas a scalar is just a real (or complex) number. To refine the explanation, one might discuss the difference between a vector field and a single vector. In the context of the present survey article, tensors are considered as mathematical objects per se. Thus, we disregard how tensors can vary in a portion of space, and look at them as single elements of a space of tensors, i.e. just the set of all tensors of the same type. In the last decades, it has become clearer and clearer that even this considerably simpler notion is of utmost applicative importance: a convincing account is provided by Landsberg’s book [
1], especially in Section 1.3 and the whole Part 3.
In most of the applications, vectors are either physical quantities which can be represented by arrows, or arrays of numbers. Since spaces of these two kinds of objects reveal the same fundamental structure, the twentieth century Mathematics has established the unifying concept of a
vector space. This abstract object turns out to be suitable for a much wider class of situations, including those in which vectors can vary in real space. Vector spaces are a kind of algebraic structure, that is, they are formally defined by means of the operations that can be performed on their elements. The multiplication of numbers by vectors can be extended by allowing more general kinds of “numbers”, which can be defined by means of another algebraic structure, named a
field (not to be confused with a field of varying objects in a portion of space). These are the building blocks of modern Linear Algebra, with which even the multilinear algebra, hence the tensor theory, can be set up. The survey is written using this language. Although not unknown, such an abstract mathematical approach might be a bit demanding for some tensor practitioners (cf. [
1], Subsection 0.3); therefore, it is likely they will prefer to accompany the reading with some of the many good books on abstract algebra.
Based on the work in [
1], Part 3, the importance of tensor decompositions can be appreciated. The perhaps most basic kind of decomposition leads to the notion of
tensor rank. It can be said that, to date, tensor rank is not understood well, as no general algorithm for determining it is known. The same is true for the symmetric rank of a symmetric tensor. Moreover, many of the techniques that have led to some success in tensor theory apply both to rank and to symmetric rank, and the assertion that they coincide (for a symmetric tensor) is the content of the Comon’s conjecture (see [
1], Exercise 2.6.6.5 and references therein). This encourages us to believe that a better understanding of the symmetric rank may shed some light on tensor rank.
When the base field is of characteristic zero, e.g., when dealing with real or complex numbers (as in most of the applicative uses of tensors), symmetric tensors are naturally identified with homogeneous polynomials. Even in positive characteristic, symmetric tensors and polynomials are “quite close” since, as a matter of facts, symmetric tensors can always be naturally identified with the socalled divided powers. From the polynomial viewpoint, the symmetric rank becomes the Waring rank, that is, the minimum number of summands that are required to express a given homogeneous polynomial as a sum of powers of linear forms. It follows that, in the characteristic zero case, to determine the maximum symmetric rank for symmetric tensors of given dimension n and order d, is the same as to determine the maximum Waring rank for degree d homogeneous polynomials in n variables. This is one of the most natural variants of the classical Waring problem on natural numbers. Now, if the symmetric rank was well understood, one probably would easily determine what is its maximum for symmetric tensors of given order and dimension. This leads us to hope that the techniques invented to find the maximum Waring rank for degree d forms in n variables might indicate some ways to understand tensor rank.
Following Geramita [
2], let us recall that, while the classical Waring problem was solved by Hilbert, its main variant remains open, i.e. determining the minimum number
$G\left(d\right)$ of summands that are needed to decompose every
sufficiently large natural number into a sum of
dth powers. From the polynomial viewpoint, a naturally analogous problem is to determine the maximum Waring rank of a
generic degree
d homogeneous polynomials in
n variables, where “generic” is meant in the sense that it commonly has in Algebraic Geometry. In this case, under the hypothesis that the base field is algebraic closed of characteristic zero, the answer is provided by the celebrated Alexander–Hirschowitz theorem, which deals with interpolation of sets of fat points, but which through the Terracini’s lemma has a direct translation in frame of the Waring rank (see [
3] for the original proof of the theorem and [
4], Section 7 for a good historical account of it). Henceforth, we shall refer to the Waring rank of a form simply as its
rank.
Interestingly, at the end of his review of the outstanding paper [
3] in the MathSciNet database, Fedor Zak wrote:
It would also be nice to know how many linear forms are required to express an arbitrary form of degree d as a sum of dth powers of linear forms.
In the mentioned Exposé [
2], Tony Geramita called this problem the
little (and the “generic version” solved by Alexander and Hirschowitz the
big) Waring problem for polynomials. However, as little as it can be, similar to a mouse, the problem is also escaping, and indeed it has remained open to date.
After Geramita’s Exposé, Johannes Kleppe was able to prove that the maximum rank of ternary quartics is seven, in his master thesis [
5] under the supervision of Kristian Ranestad. More precisely, that work contains a study of normal forms and ranks of ternary cubics and quartics, from which the maximum ranks can be obtained.
Some years later, at the opposite side of Kleppe’s result, which gives a sharp upper bound in two very specific cases, Corollary 1 [
6] provided us with a very mild upper bound in the general case. After about the same amount of time, that result was dramatically improved by Blekherman and Teitler (see [
7], Corollary 9). From the references in [
6], some information on earlier works on the subject can be obtained. For instance, the work in [
8], published in 1969, deals with the problem for arbitrary fields, with a particular care to detect in which cases there is actually a maximum for the Waring rank (as it is always the case when the base field is algebraically closed of characteristic zero). We believe that any list of references one might try to work out for the problem of our interest will likely be far from complete, since it is of a natural and elementary nature and dates back to at least the first half of the past century. For instance, B. Segre at the beginning of [
9], Section 96, mentioned a question raised in [
10].
The introduction of Buczyński and Teitler’s paper [
11] is a good source of information on what is known on the maximum Waring rank, up to recent times. As tensor theory, as well as its related algebrogeometric aspects, is currently a very active research field, a good deal of new results have been discovered since then, and they can be of use to determine that maximum. However, if we strictly focus on the problem, as far as we know, there are only two new facts that improve ([
11], Table 1). The first one is [
12], Proposition 3.3, which sensibly improves the Blekherman and Teitler’s upper bound when
$n=3$. The other is [
13], Theorem 2.5, which extends to even degrees the Buczyński and Teitler’s lower bound established in [
11], Theorem 1 (in the sense that it raises by one the lower bound given by the maximum rank of monomials, which for an even degree
d is
$({d}^{2}+2d)/4$).
According to the aforementioned results, for ternary forms over algebraically closed fields of characteristic zero both the lower and the upper bounds are of order $O({d}^{2}/4)$, and the gap is $d1$. The main purpose of the present survey article is to convey the basic ideas that led to those bounds.
2. Symmetric Algebra and Apolarity
In this section, we give an overview of some fundamental facts that are worthy of being recalled in the present context. The technical treatments of basic tensor theory and apolarity that can be encountered in the literature may vary considerably. Instead of fixing one of them, we only review the main statements: whatever reference the reader is willing to adopt, the basic definitions will likely be compatible with (or at least adaptable to) our assumptions. The exposition is organized so that, by adding sufficient details, one can get a coherent theoretical development.
2.1. The Tensor Algebra of a Vector Space
We take the view that the tensor algebra $T\left(V\right)$ can be any fixed $\mathbb{K}$algebra that contains V as a $\mathbb{K}$vector subspace and such that for each $\mathbb{K}$vector space homomorphism of V into a (commutative or not) $\mathbb{K}$algebra A there exists a unique $\mathbb{K}$algebra homomorphism $T\left(V\right)\to A$ that extends it. The elements of $T\left(V\right)$ are called tensors, the multiplication in $T\left(V\right)$ is denoted by ⊗ and ${t}_{0}\otimes {t}_{1}$ is the tensor product of ${t}_{0}$ and ${t}_{1}$. However, $V\otimes V$ does not denote the set of all ${v}_{0}\otimes {v}_{1}$ with ${v}_{0},{v}_{1}\in V$, but the vector subspace of $T\left(V\right)$ generated by that set. The tensor power$V\otimes \cdots \otimes V=:{V}^{\otimes d}$ can be similarly defined; its elements are said to have orderd. The multiplication ${\mu}_{d}:V\times \cdots \times V\to {V}^{\otimes n}$, $\left({v}_{1},\dots ,{v}_{n}\right)\mapsto {v}_{1}\otimes \cdots \otimes {v}_{n}$, is a universal dlinear map, that is, every dlinear map $m:V\times \cdots \times V\to W$ factors as $\phi \circ {\mu}_{d}$ for a uniquely determined vector space homomorphism $\phi :{V}^{\otimes d}\to W$.
This approach may be considered rather abstract, but it might be said that $T\left(V\right)$ is no more abstract than $\mathbb{C}$: similar to how one extends $\mathbb{R}$ by adding a square root of $1$ and all objects that are consequently needed to keep the operations and their properties, here one extends V and $\mathbb{K}$ by adding tensor products of vectors and all objects that are consequently needed to keep the operations of V, and to build a multiplication with the usual properties (apart from commutativity). The above characterization of $T\left(V\right)$ (up to algebra isomorphisms) makes precise the intuitive idea, and indeed a similar characterization holds for $\mathbb{C}$: for every $\mathbb{R}$algebra A that contains a square root x of $1$ (for instance, A could be the algebra of endomorphisms of vectors in the plane and x a rotation by ${90}^{\circ}$), there exists a unique $\mathbb{R}$algebra homomorphism $\mathbb{C}\to A$ that sends the imaginary unit into x.
2.2. The Symmetric Algebra of a Vector Space
The definition of the symmetric algebra $S\left(V\right)$ can be given in the same way, except for the fact that now the target $\mathbb{K}$algebra A in the characteristic property is required to be commutative. A very natural $\mathbb{K}$algebra homomorphism $T\left(V\right)\to S\left(V\right)$ arises: the one which extends the identity map of V. Elements of $S\left(V\right)$ are not exactly symmetric tensors, though in most cases the two things can be identified, but rather polynomials. Indeed, if ${\left({x}_{i}\right)}_{i\in I}$ denotes a (possibly infinite) basis of the vector space V, any element of $S\left(V\right)$ can be written as a polynomial in the ${x}_{i}$s, and two elements are equal if and only if for each monomial they carry the same coefficient.
2.3. Symmetric Tensors
To define symmetric tensors, first notice that for each permutation $\sigma $ of $\{1,\dots ,n\}$, by the mentioned universal property of tensor powers we have an automorphism ${\sigma}_{V}$ of ${V}^{\otimes n}$ such that ${v}_{1}\otimes \cdots \otimes {v}_{n}\mapsto {v}_{\sigma \left(1\right)}\otimes \cdots \otimes {v}_{\sigma \left(n\right)}$. A tensor in ${V}^{\otimes n}$ is said to be symmetric if it is invariant under ${\sigma}_{V}$ for all permutation $\sigma $ of $\{1,\dots ,n\}$. For instance, $v\otimes w+w\otimes v$ and $v\otimes v\otimes w+v\otimes w\otimes v+w\otimes v\otimes v$ are symmetric tensors for whatever choice of $v,w\in V$. A sum of symmetric tensors of different orders can be called symmetric as well (although very seldom one encounters such sums).
2.4. The Symmetric Product
The set
$\overline{S}\left(V\right)$ of symmetric tensors is a vector subspace, but not a subalgebra of
$T\left(V\right)$ (apart from the trivial cases when
$dimV\le 1$). However, a meaningful multiplication in
$\overline{S}\left(V\right)$ can be introduced. The
symmetric product${s}_{0}{s}_{1}$ of two symmetric tensors of given orders
${d}_{0},{d}_{1}$ can be defined as the sum of
${\sigma}_{V}({s}_{0}\otimes {s}_{1})$ with
$\sigma $ varying on the
$({d}_{0},{d}_{1})$shuffles, that are the permutations of
$\{1,\dots ,{d}_{0}+{d}_{1}\}$ such that
$\sigma \left(1\right)<\cdots <\sigma \left({d}_{0}\right)$ and
$\sigma ({d}_{0}+1)<\cdots <\sigma ({d}_{0}+{d}_{1})$. For instance,
for whatever choice of
$u,v,w\in V$.
There is a unique way to extend this operation on the whole of $\overline{S}\left(V\right)$, that preserves the distributive law. Note that with this definition we have ${v}^{d}=d!{v}^{\otimes d}=d!v\otimes \cdots \otimes v$ (d factors).
2.5. A Variant of the Symmetric Product in Characteristic Zero
We mention that there is a much more popular variant of the symmetric product, which is in use when
$\mathbb{K}$ is
$\mathbb{C}$, or more generally is of characteristic zero (see, e.g., [
1], 2.6.3). It can be presented as follows. Every
$\mathbb{K}$vector space automorphism
$\alpha $ of any
$\mathbb{K}$algebra
A induces another multiplication on
A: the one that makes
$\alpha $ a
$\mathbb{K}$algebra automorphism, when it replaces the original one on the target (but not on the domain). In particular, if a sequence
${\left({a}_{d}\right)}_{d\ge 0}$ of nonzero scalars is given, the multiplication by
${a}_{d}$ on order
d symmetric tensors gives a
$\mathbb{K}$vector space automorphism of
$\overline{S}\left(V\right)$, hence a multiplication in it. The mentioned symmetric product in
$\overline{S}\left(V\right)$ is given by the sequence
${a}_{d}:=1/d!$. For instance, this definition gives
(compare with Equation (
1)); note also that now
${v}^{d}={v}^{\otimes d}$.
The reason for the popularity of the “modified” symmetric product is explained by the relationship between symmetric tensors and polynomials: it makes the restriction $\overline{S}\left(V\right)\to S\left(V\right)$ of the natural map $T\left(V\right)\to S\left(V\right)$ a $\mathbb{K}$algebra isomorphism. In positive characteristic, a multiplication with such a desirable property can not be hoped for, simply because in this case the restriction $\overline{S}\left(V\right)\to S\left(V\right)$ is not bijective (unless $dimV\le 1$). It also worthy of being remarked that in characteristic zero, $\overline{S}\left(V\right)$ with the previous symmetric product is naturally isomorphic to $S\left(V\right)$ as well, but in this case through the $\mathbb{K}$algebra homomorphism $S\left(V\right)\to \overline{S}\left(V\right)$ that arises simply because $\overline{S}\left(V\right)$ is a commutative $\mathbb{K}$algebra containing V.
Although the subject of the present work is the maximum symmetric rank in the characteristic zero case, we prefer to assume that $\overline{S}\left(V\right)$ is equipped with the general symmetric product (which is defined regardless of the characteristic of $\mathbb{K}$), not with the modified one.
2.6. Tensors on the Dual and Multilinear Forms
Let us consider the dual space
${V}^{*}$ and denote by
${V}^{*d}$ the space of
dlinear forms. A
dlinear form
$\mu :{V}^{d}\to \mathbb{K}$ and a
${d}^{\prime}$linear form
${\mu}^{\prime}:{V}^{{d}^{\prime}}\to \mathbb{K}$ give rise to the
$(d+{d}^{\prime})$linear form
This allows us to define a
$\mathbb{K}$algebra structure on
Since ${V}^{*}$ embeds (as a summand) in $M\left(V\right)$, one comes with a natural $\mathbb{K}$algebra homomorphism $T\left({V}^{*}\right)\to M\left(V\right)$, which is always injective and turns out to be an isomorphism if and only if $dimV<\infty $. On the other hand, to give a dlinear form on V is the same as to give a linear form on ${V}^{\otimes d}$, and $T\left(V\right)$, as a matter of facts, is a graded algebra whose degree d component is ${V}^{\otimes d}$ (regardless of the finiteness assumption). Therefore, $M\left(V\right)$ is isomorphic as a vector space to the graded dual of $T\left(V\right)$. In conclusion, when $dimV<\infty $, the graded dual of $T\left(V\right)$ is isomorphic, as a graded vector space, to $T\left({V}^{*}\right)$.
Let us also point out that, given
${l}_{1},\dots ,{l}_{d}\in {V}^{*}$, the image of the tensor product
${l}_{1}\otimes \cdots \otimes {l}_{d}$ in
$M\left(V\right)$ takes value
on
$\left({v}_{1},\dots ,{v}_{d}\right)$, and this is also the value taken on
${v}_{1}\otimes \cdots \otimes {v}_{d}$ by the image in the graded dual of
$T\left(V\right)$.
For the symmetric algebra, we have the following facts. To begin with, $S\left(V\right)$ is a graded algebra as well, and the natural homomorphism $T\left(V\right)\to S\left(V\right)$ is a graded one, and as a matter of facts is surjective on each graded component. Hence, the graded dual of $S\left(V\right)$ embeds into the graded dual of $T\left(V\right)$, which is isomorphic to $M\left(V\right)$. However, similarly to what happens for ${V}^{\otimes d}$, to give a dlinear form on the symmetric power ${S}^{d}\left(V\right)$ (the subspace of degree d elements of $S\left(V\right)$) is the same as to give a symmetricdlinear form on V. This easily implies that the image of the dual of ${S}^{d}\left(V\right)$ through the embedding into $M\left(V\right)$ is precisely the subspace of dlinear symmetric forms. Next, it is not difficult to show that an order d tensor in $T\left({V}^{*}\right)$ is symmetric if and only if its image through the embedding $T\left({V}^{*}\right)\hookrightarrow M\left(V\right)$ is a symmetric dlinear map. In the finite dimensional case, the converse is also true: a dlinear form is symmetric if and only if it corresponds to a symmetric tensor. In conclusion, when $dimV<\infty $, the graded dual of $S\left(V\right)$ is isomorphic, as a graded $\mathbb{K}$vector space, to $\overline{S}\left({V}^{*}\right)$.
According to the definition of the symmetric product, given
${l}_{1},\dots ,{l}_{n}\in {V}^{*}$, the image of
${l}_{1}\cdots {l}_{d}$ in
$M\left(V\right)$ takes value
on
$\left({v}_{1},\dots ,{v}_{d}\right)$, that is, the
permanent of the matrix
${\left({l}_{i}\left({v}_{j}\right)\right)}_{i,j}$. This is also the value taken on
${v}_{1}\cdots {v}_{d}$ by the image of
${l}_{1}\cdots {l}_{d}$ in the graded dual of
$S\left(V\right)$.
2.7. Evaluation
The algebra ${\mathbb{K}}^{V}$ of all functions $V\to \mathbb{K}$ is commutative and contains ${V}^{*}$. Hence, there is a canonical $\mathbb{K}$algebra homomorphism $S\left({V}^{*}\right)\to {\mathbb{K}}^{V}$. Note also that the value of the image of $f\in S\left({V}^{*}\right)$ on v is the image of through the evaluation homomorphism $S\left({V}^{*}\right)\to \mathbb{K}$ that extends the evaluation at v map ${V}^{*}\to \mathbb{K}$ ($l\mapsto l\left(v\right)$). If $f\in S\left({V}^{*}\right)$ is considered as a polynomial, the image in ${\mathbb{K}}^{V}$ is the corresponding polynomial function.
For a form $f\in {S}^{d}\left({V}^{*}\right)$, the corresponding symmetric dlinear form can be considered as a polarization of the corresponding polynomial function (though it is customary to affect that dlinear form by a factor $1/d!$).
2.8. Symmetric Tensors on the Dual and Divided Powers
The symmetric product may be preferred over its variant in characteristic zero, not only because it is defined regardless to the characteristic of
$\mathbb{K}$, but also because it makes
$\overline{S}\left(V\right)$ an algebra of divided powers, in a quite simple way: it suffices to set
${v}^{\left[d\right]}:={v}^{\otimes d}$. A perhaps more familiar way to introduce divided powers is in the context of duality, as done, e.g., in [
14], Appendix A. To link the discussion earlier to the Iarrobino and Kanev’s approach:
Since an isomorphic image of a tensor algebra is a tensor algebra as well, we can assume that
$T\left({V}^{*}\right)$ is chosen so that
$\overline{S}\left({V}^{*}\right)$ actually
is the graded dual of
$S\left(V\right)$. Similarly, if
$R=\mathbb{K}\left[{x}_{1},\dots ,{x}_{r}\right]={\oplus}_{d\ge 0}{R}_{d}$ is a (graded) ring of polynomials, we can assume
$S\left({R}_{1}\right)=R$. Let us also assume
$V:={R}_{1}$ (under a suitable definition of polynomial rings, this assumption imposes no restriction on
V). Setting
$\mathcal{D}:=\overline{S}\left({V}^{*}\right)$ and denoting by
${\mathcal{D}}_{d}$ the subspace of order
d symmetric tensors, we get exactly in the situation at the beginning of [
14], Appendix A. Let us employ a few lines below to check that setting
${l}^{\left[d\right]}:={l}^{\otimes d}$, subsequent definitions of [
14], Appendix A, are automatically fulfilled.
Let
$\left({X}_{1},\dots ,{X}_{r}\right)$ be the base of
${\mathcal{D}}_{1}$, dual to
$\left({x}_{1},\dots ,{x}_{r}\right)$ (that is,
${X}_{i}\left({x}_{j}\right)$ is 0 when
$i\ne j$ and 1 when
$i=j$). Note that
${X}_{i}{X}_{j}$ takes value 1 on
${x}_{i}{x}_{j}$ when
$i\ne j$ (it is indeed given by Equation (
3), and
${l}_{i}\left({x}_{j}\right)={l}_{j}\left({v}_{i}\right)=0$), but the value is 2 when
$i=j$, that is,
${X}_{i}^{2}\left({x}_{i}^{2}\right)=2$. Instead, we have
${X}_{i}^{\left[2\right]}\left({x}_{i}\right)={X}_{i}^{\otimes 2}\left({x}_{i}\right)=1$, as it follows from the evaluation rule Equation (
2) (or also from
${X}_{i}^{2}=2!{X}_{i}^{\otimes 2}$). More generally, we have that
${X}_{1}^{\left[{d}_{1}\right]}\cdots {X}_{r}^{\left[{d}_{r}\right]}$ takes value 0 on
${x}_{1}^{{d}_{1}^{\prime}}\cdots {x}_{r}^{{d}_{r}^{\prime}}$ when
$\left({d}_{1},\dots ,{d}_{r}\right)\ne \left({d}_{1}^{\prime},\dots ,{d}_{r}^{\prime}\right)$ and value 1 when
$\left({d}_{1},\dots ,{d}_{r}\right)=\left({d}_{1}^{\prime},\dots ,{d}_{r}^{\prime}\right)$. This agrees with [
14], Definition A.1. Since the number of
$(d,{d}^{\prime})$shuffles is
$(d+{d}^{\prime})!/(d!{d}^{\prime}!)$, we have
hence
in agreement with [
14], (A.0.5). Similar calculations lead to
in agreement with [
14], Definition A.8.
2.9. The Contraction Map
Given
$p\in {R}_{d}$,
$f\in {\mathcal{D}}_{d+{d}^{\prime}}=Hom\left({R}_{d+{d}^{\prime}},\mathbb{K}\right)$ and denoting by
${\mu}_{p}:{R}_{{d}^{\prime}}\to {R}_{d+{d}^{\prime}}$ the vector space homomorphism given by the multiplication by
p, the composition
$f\circ {\mu}_{p}$ belongs to
${\mathcal{D}}_{{d}^{\prime}}$ and is called the
contraction of f by p. The bilinear operations
${R}_{d}\times {\mathcal{D}}_{d+{d}^{\prime}}\to {\mathcal{D}}_{{d}^{\prime}}$ (assuming
${\mathcal{D}}_{{d}^{\prime}}=\left\{0\right\}$ when
${d}^{\prime}<0$) extend to a unique bilinear operation
$R\times \mathcal{D}\to \mathcal{D}$, that can be called
contraction map (in agreement with [
14], Definition A.2). If
${l}_{1},\dots ,{l}_{d}\in {R}_{1}$ and
$\phi \in M\left(V\right)$ is the
$(d+{d}^{\prime})$linear form corresponding to
f, the contraction of
f by
${l}_{1}\cdots {l}_{d}$ corresponds to the
${d}^{\prime}$linear form obtained by fixing
d arguments of
$\phi $ equal to
${l}_{1},\dots ,{l}_{d}$:
$\left({v}_{1},\dots ,{v}_{{d}^{\prime}}\right)\mapsto \phi \left({l}_{1},\dots ,{l}_{d},{v}_{1},\dots ,{v}_{{d}^{\prime}}\right)$. This property may also be used to give an alternative definition of the contraction, because of the characteristic property of the symmetric powers (the
dlinear assignment on all
$({l}_{1},\dots ,{l}_{d})\in {R}_{1}^{d}$ determines a homomorphism
${R}_{d}={S}^{d}\left({R}_{1}\right)\to {V}^{*{d}^{\prime}}$, whose image is canonically isomorphic to
${\mathcal{D}}_{{d}^{\prime}}$ and which depends on
$f\in {\mathcal{D}}_{d+{d}^{\prime}}$ in a linear way). For this reason, the contraction can also be called
insertion and denoted by ┙ (this attitude is perhaps more common in the context of alternating forms).
2.10. Contraction and Derivatives
The ordinary directional derivative of differentiable (real valued) functions fulfills the Leibnitz rule
${\partial}_{v}\left(fg\right)=\left({\partial}_{v}f\right)g+f{\partial}_{v}g$ and on linear forms is nothing but the evaluation on
v. These two properties can be used to characterize the derivative of polynomials in
$S\left({V}^{*}\right)$ along
$v\in V$. Indeed, for each
d, we can define
${\partial}_{v}$ on
${S}^{d}{V}^{*}$ as the unique operator into
${S}^{d1}{V}^{*}$ such that
for all
${l}_{1},\dots ,{l}_{d}\in {V}^{*}$ (with the hat denoting omission). Then,
${\partial}_{v}$ extends to the whole of
$S\left({V}^{*}\right)$ by additivity. Using again the fact that linear operators on
${S}^{d}{V}^{*}$ are characterized by their values on the products
${l}_{1}\cdots {l}_{d}$, one can easily check that the Leibnitz rule holds in
$S\left({V}^{*}\right)$ and that
${\partial}_{v}$ is the unique extension of the evaluation on
v in
${V}^{*}$ with this property. A partial derivative is obviously a directional derivative along a basis vector
${x}_{i}$.
Let
${l}_{1},\dots ,{l}_{d}\in {V}^{*}$,
${v}_{1},\dots ,{v}_{d}\in V$, and
f be the image in
${\mathcal{D}}_{d}$ of
${l}_{1}\cdots {l}_{d}\in {S}^{d}{V}^{*}$ (that is, the product
${l}_{1}\cdots {l}_{d}$ in the ring
$\mathcal{D}$). From the evaluation rule Equation (
3) follows that the contraction of
f by
${v}_{1}$ take the same value on
${v}_{2}\cdots {v}_{d}$ as
(is basically a Laplacelike expansion of the permanent along the first row), which is the image of
${\partial}_{{v}_{1}}\left({l}_{1}\cdots {l}_{d}\right)\in {S}^{d1}{V}^{*}$. Using additivity and again the fact that operators on symmetric powers are determined by their values on products of vectors, we conclude that, for every
$v\in V$, the partial derivative
${\partial}_{v}$ and the contraction by
v are compatible via the canonical homomorphism
$S\left({V}^{*}\right)\to \overline{S}\left({V}^{*}\right)=\mathcal{D}$.
A constant coefficient linear partial differential operator on $S\left({V}^{*}\right)$ is a linear combination of compositions of directional (or partial) derivatives. The set D of such operators is a commutative $\mathbb{K}$algebra with multiplication given by composition. Hence, $v\mapsto {\partial}_{v}$ extends in a unique way to a $\mathbb{K}$algebra homomorphism $S\left(V\right)=R\to D$. The image of each $p\in R$ in D can be denoted by ${\partial}_{p}$.
2.11. Apolarity
When $\mathbb{K}$ is of characteristic zero, both canonical homomorphisms $S\left({V}^{*}\right)\to \mathcal{D}$ and $R\to D$ are isomorphisms. In this case, we can assume that $S\left({V}^{*}\right)=\mathcal{D}$ (and under a suitable definition of polynomials also $R=D$ could be assumed). This way the contraction map becomes a bilinear map $S\left(V\right)\times S\left({V}^{*}\right)\to S\left({V}^{*}\right)$ such that each p acts as the constant coefficient linear differential operator ${\partial}_{p}$ (e.g., the contraction of ${X}_{1}{X}_{2}^{2}$ by $3{x}_{1}{x}_{2}+{x}_{2}^{2}$ is $3{\partial}_{{x}_{1}}{\partial}_{{x}_{2}}\left({X}_{1}{X}_{2}^{2}\right)+{\partial}_{{x}_{2}}{\partial}_{{x}_{2}}\left({X}_{1}{X}_{2}^{2}\right)=2{X}_{1}+6{X}_{2}$).
From the coordinatefree definitions is quite easy to recognize that the contraction map is invariant with respect to the canonical actions of
$GL\left(V\right)$ on
V and on
${V}^{*}$ (cf. also [
14], Proposition A3(i)). As Ehrenborg and Rota reported in [
15], Introduction, for each fixed degree there is a unique invariant bilinear form
${S}^{d}\left(V\right)\times {S}^{d}\left({V}^{*}\right)\to \mathbb{K}$, which has been much used since the nineteenth century by classical invariant theorists. They also say that this form can be called
apolar bilinear form, that the subject of apolarity has been related with the symbolic method in classical invariant theory, and that an efficient treatment can be given in frames of Hopf algebras.
Apolarity for univariate polynomials of the same degree can also be disguised as an explicit formula with alternating signs. To understand why, it may be useful to have a quick look on what happens for the homogeneous version of such polynomials, that is, for binary forms. From a geometric, projective viewpoint, (nonzero) vectors can be viewed as points and linear forms as hyperplanes. For binary forms, that is, when
$dimV=2$ and the projective picture is a line, hyperplanes are (singletons of) points. From the algebraic viewpoint, this amounts to the existence of an isomorphism
$V\to {V}^{*}$, unique up to scalar factors (or, equivalently, to the existence of a unique, up to scalar factors, nondegenerate alternating form on
V). In coordinates,
${a}_{0}{x}_{0}+{a}_{1}{x}_{1}$ corresponds to
${a}_{1}{X}_{0}{a}_{0}{X}_{1}$ (or to a scalar multiple of it). The extension
$S\left(V\right)\stackrel{\sim}{\to}S\left({V}^{*}\right)$ of this isomorphism allows us to equivalently describe apolarity as a bilinear form on
$S\left(V\right)$ alone. Restricting the attention on (homogeneous) polynomials of the same degree, we get a bilinear form on
${S}^{d}\left(V\right)$, which is alternating for
d odd and symmetric for
d even. In coordinates:
In terms of univariate polynomials of an assigned degree
d:
Some classical results deal with this kind of apolarity, with $\mathbb{K}=\mathbb{C}$. For instance, Grace’s theorem is sometimes named Grace’s apolarity theorem for this reason. Nowadays, there is a basic result, which is largely referred to as the apolarity lemma. It plays a fundamental role in proving the bounds on Waring ranks we aim to present in this article. We end this section by setting up the technical environment of our presentation. In the next section, we state a version of the apolarity lemma.
2.12. Standing Assumptions
From the usual geometric viewpoint, forms are regarded as hypersurfaces. To begin with, for every element of
$S\left({V}^{*}\right)$, the vanishing locus of the corresponding polynomial function is an affine hypersuperface in
V. However, a projective viewpoint is perhaps more fruitful. We define the projective space
$\mathbb{P}\left(V\right)$ as the set of onedimensional subspace of
V,
$\u2329\phantom{\rule{0.166667em}{0ex}}v\phantom{\rule{0.166667em}{0ex}}\u232a$ with
$v\ne 0$, and the hypersurface corresponding to a form
$f\in {S}^{d}\left({V}^{*}\right)$ as the zero locus in
$\mathbb{P}\left(V\right)$ of the corresponding degree
d homogeneous function on
V. Sometimes geometric features of the hypersurface defined by a form
f, such as the singular locus, can give information on the rank of
f (cf. [
1], Theorem 9.2.1.4).
However, rank determination often uses the dual viewpoint, from which forms are considered as points in a space where powers of linear forms constitute a Veronese variety (see [
1], 4.3.7). Sometimes, having a simultaneous look on both viewpoint has been fruitful, and perhaps a systematic investigation with this double viewpoint could be worthy of being pursued. However, that it is not our goal here, we prefer to facilitate the interchange between
V and
${V}^{*}$ by working on an arbitrary pair of (finite dimensional) vector spaces, with a given perfect pairing between one another. To denote such spaces, we follow a kind of
abstract index notation, using upper indices for one of the two spaces and lower indices for the other. In a purely algebraic context, especially one in which powers play an important role, this might be considered bad practice. However, if one takes care of not assigning an independent meaning to
x, the use of
${x}^{i}$ causes no ambiguities. An advantage of this choice is to have a notation that can be more promptly translated (and provide insight) in physics contexts where tensors are widely used.
As we have anticipated, we are interested in the case when $\mathbb{K}$ is algebraically closed of characteristic zero (e.g., $\mathbb{K}=\mathbb{C}$). Hence, apolarity will be assumed on a (dual) pair of symmetric algebras, which we denote by ${S}^{\u2022}$ and ${S}_{\u2022}$. From our preferred geometric viewpoint, elements of ${S}^{\u2022}$ are considered as polynomial functions on the degree 1 component ${S}_{1}$ of ${S}_{\u2022}$ (in accordance with the abstract index notation, where forms take upper indices and vectors lower indices), but elements of ${S}^{\u2022}$ act also as constant coefficient linear differential operators on ${S}_{\u2022}$. Note that, to fit the first description into the previously outlined technical treatment, one needs to identify ${S}_{\u2022}$ with $S\left(V\right)$ and ${S}^{\u2022}$ with $S\left({V}^{*}\right)$, whereas, to fit the other, one needs the converse. Polynomials whose rank is to be studied will live in ${S}_{d}$.
Following from the above said, now we set up more formally our ground technical framework for the subsequent sections.
We assume that $\mathbb{K}$ is algebraically closed and of characteristic zero. The span of a subset $X\subseteq V$ is denoted by $\u2329\phantom{\rule{0.166667em}{0ex}}X\phantom{\rule{0.166667em}{0ex}}\u232a$ (or also $\u2329\phantom{\rule{0.166667em}{0ex}}{v}_{1},\dots ,{v}_{n}\phantom{\rule{0.166667em}{0ex}}\u232a$ when $X=\{{v}_{1},\dots ,{v}_{n}\}$). We define the projective space $\mathbb{P}\left(V\right)$ as the set $\left\{\u2329\phantom{\rule{0.166667em}{0ex}}v\phantom{\rule{0.166667em}{0ex}}\u232a:v\in V\setminus \left\{0\right\}\right\}$ of onedimensional subspaces of V. We fix two symmetric algebras ${S}_{\u2022}$ and ${S}^{\u2022}$, whose degree d components are denoted by ${S}_{d}$ and ${S}^{d}$. We assume $dim{S}_{1}<\infty $, $dim{S}^{1}<\infty $, ${S}_{\u2022}=S\left({S}_{1}\right)$, ${S}^{\u2022}=S\left({S}^{1}\right)$. We assume that a perfect pairing ${S}^{1}\times {S}_{1}\to \mathbb{K}$ is given. By the value $f\left(v\right)$ of $f\in {S}^{\u2022}$ at $v\in {S}_{1}$, we mean the value at v of the image of f through ${S}^{\u2022}\stackrel{\sim}{\to}S\left({S}_{1}^{*}\right)\to {\mathbb{K}}^{{S}_{1}}$, where the first isomorphism is induced by the given perfect pairing. The ideal $I\left(X\right)\subseteq {S}^{\u2022}$ of a subset $X\subseteq \mathbb{P}\left({S}_{1}\right)$ is the homogenous ideal with degree d components given, for each d, by all $f\in {S}^{d}$ such that $f\left(v\right)=0$ for all $\u2329\phantom{\rule{0.166667em}{0ex}}v\phantom{\rule{0.166667em}{0ex}}\u232a\in X$. The apolar bilinear map ${S}^{\u2022}\times {S}_{\u2022}\to {S}_{\u2022}$ is the map induced by the earlier defined bilinear map $S\left(V\right)\times S\left({V}^{*}\right)\to S\left({V}^{*}\right)$, when $V={S}^{1}$, through the isomorphism $S\left({{S}^{1}}^{*}\right)\stackrel{\sim}{\to}{S}_{\u2022}$ induced by the given perfect pairing. ${\partial}_{f}x$ denotes the value of the apolar map on $(f,x)\in {S}^{\u2022}\times {S}_{\u2022}$. f and x are said to be apolar to each other when ${\partial}_{f}x=0$, and the ideal $Ann\phantom{\rule{0.166667em}{0ex}}x\subseteq {S}^{\u2022}$ of all f apolar to x is called the apolar ideal of x. If $W\le {S}_{d}$ is a subspace, ${W}^{\perp}$ denotes the set of all $f\in {S}^{d}$ that are apolar to all elements of W. Similarly, for a subspace $W\le {S}^{d}$, ${W}^{\perp}$ denotes the set of all $x\in {S}_{d}$ that are apolar to all elements of W.

From the discussion above, it follows that apolarity induces an isomorphism ${S}^{d}\to {S}_{d}^{*}$ for each d, therefore gives a perfect pairing in each degree. Hence, ${W}^{\perp}$ denotes nothing but the orthogonal complement with respect such a perfect pairing. The notation $Ann\phantom{\rule{0.166667em}{0ex}}x$ for the apolar ideal complies with the notion of the annihilator of an element of a module, because ${S}_{\u2022}$ is structured as an ${S}^{\u2022}$module by apolarity. We prefer not to use the quite common notation ${x}^{\perp}$ for the apolar ideal (we speak about orthogonality only in a fixed degree).
When one needs schemes (for which we assume the definitions in [
16]), it turns out that a point of
$\mathbf{Proj}\phantom{\rule{0.166667em}{0ex}}S\left({V}^{*}\right)$ rational over
$\mathbb{K}$ is a maximal nonirrelevant homogeneous ideal in
$S\left({V}^{*}\right)$ such that its intersection with
${S}^{1}\left({V}^{*}\right)={V}^{*}$ is a hyperplane. However, every hyperplane in
${V}^{*}$ is the hyperplane of forms that vanish on some point in
$\mathbb{P}\left(V\right)$. This gives the canonical identification of
$\mathbb{P}\left(V\right)$ with the set of
$\mathbb{K}$points of
$\mathbf{Proj}\phantom{\rule{0.166667em}{0ex}}S\left({V}^{*}\right)$.
3. Basic Results
3.1. Apolarity Lemma
Let us preliminary point out that
Remark 1. From Equation (
4)
, it follows that, given $f\in {S}^{d}$ and $v\in {S}_{1}$ such that $f\left(v\right)=0$, for every $g\in {S}^{{d}^{\prime}}$, we have However, ${\partial}_{g}{\partial}_{f}{v}^{d+{d}^{\prime}}=0$ for all $g\in {S}^{{d}^{\prime}}$ implies that ${\partial}_{f}{v}^{d+{d}^{\prime}}=0$ (because apolarity gives a perfect pairing in degree ${d}^{\prime}$). Of course, ${\partial}_{f}$ also vanishes on ${v}^{{d}^{\prime}}$ when ${d}^{\prime}<d$. Therefore, if $f\left(v\right)=0$ then ${\partial}_{f}$ vanishes on all powers of v.
By additivity, we conclude that every $f\in I\left(\{\u2329\phantom{\rule{0.166667em}{0ex}}{v}_{1}\phantom{\rule{0.166667em}{0ex}}\u232a,\cdots ,\u2329\phantom{\rule{0.166667em}{0ex}}{v}_{r}\phantom{\rule{0.166667em}{0ex}}\u232a\}\right)$ is apolar to every power sum ${{v}_{1}}^{d}+\cdots +{{v}_{r}}^{d}$, $d>0$ and, more generally, to every linear combination of ${{v}_{1}}^{d},\dots ,{{v}_{r}}^{d}$, $d>0$.
Now, we prove the following version of the apolarity lemma.
Lemma 1. Let $x\in {S}_{d}$ with $d>0$, and $X:=\{\u2329\phantom{\rule{0.166667em}{0ex}}{v}_{1}\phantom{\rule{0.166667em}{0ex}}\u232a,\cdots ,\u2329\phantom{\rule{0.166667em}{0ex}}{v}_{r}\phantom{\rule{0.166667em}{0ex}}\u232a\}\subset \mathbb{P}{S}_{1}$. Then, Proof. Suppose that $x\in \u2329\phantom{\rule{0.166667em}{0ex}}{{v}_{1}}^{d},\dots ,{{v}_{r}}^{d}\phantom{\rule{0.166667em}{0ex}}\u232a$, that is, x is a linear combination of ${{v}_{1}}^{d},\dots ,{{v}_{r}}^{d}$. By Remark 1, every $f\in I\left(X\right)$ is apolar to such linear combination, hence is apolar to x. Therefore, $I\left(X\right)\subseteq Ann\phantom{\rule{0.166667em}{0ex}}x$.
Conversely, let us suppose that
$I\left(X\right)\subseteq Ann\phantom{\rule{0.166667em}{0ex}}x$. By the evaluation of Equation (
4), it follows that
$f\in {S}^{d}$ vanishes on
$v\in {S}_{1}$ if and only if it is apolar to
${v}^{d}$. In other terms, the set of all
$f\in {S}^{d}$ that vanish on
$v\in {S}_{1}$ is the orthogonal complement of
${v}^{d}$ with respect to the perfect pairing given by apolarity in degree
d. Hence
Since $I\left(X\right)\subseteq Ann\phantom{\rule{0.166667em}{0ex}}x$, we have in particular that x is orthogonal to $I\left(X\right)\cap {S}^{d}$. Hence, $x\in \u2329\phantom{\rule{0.166667em}{0ex}}{{v}_{1}}^{d},\dots ,{{v}_{r}}^{d}\phantom{\rule{0.166667em}{0ex}}\u232a$. □
A general and detailed version of the apolarity lemma can be found in [
14], Lemma 1.15. It holds in every characteristic and uses divided powers. Lemma 1 is basically equivalent to [
14], Lemma 1.15(i) and (ii), restricted to the characteristic zero case.
To illustrate the lemma with a simple (nearly trivial) example, let ${S}_{\u2022}:=\mathbb{C}\left[{x}_{0},{x}_{1}\right]$, ${S}^{\u2022}:=\mathbb{C}\left[{x}^{0},{x}^{1}\right]$, with $\left({x}^{0},{x}^{1}\right)$ being the dual basis of $\left({x}_{0},{x}_{1}\right)$. The evaluation of a polynomial $p=p\left({x}^{0},{x}^{1}\right)\in {S}^{\u2022}$ on ${a}^{0}{x}_{0}+{a}^{1}{x}_{1}\in {S}_{1}$ is just $p\left({a}^{0},{a}^{1}\right)$. When p is homogeneous of degree d and $p\left(v\right)=0$ for a $v\in {S}_{1}$, then p vanishes on all scalar multiples of v, that is, on all elements of $\u2329\phantom{\rule{0.166667em}{0ex}}v\phantom{\rule{0.166667em}{0ex}}\u232a$. If $v\ne 0$, we can say that $\u2329\phantom{\rule{0.166667em}{0ex}}v\phantom{\rule{0.166667em}{0ex}}\u232a\in \mathbb{P}{S}_{1}$ is a root of p.
Let us find the sum of squares decompositions of
$f:={x}_{0}{x}_{1}\in {S}_{2}$ using the apolarity lemma. We have
${\partial}_{{x}^{0}}f={x}_{1}$ and
${\partial}_{{x}^{1}}f={x}_{0}$, hence
$Ann\phantom{\rule{0.166667em}{0ex}}f$ has no degree 1 homogeneous nonzero elements. In degree 2 we have
$Ann\phantom{\rule{0.166667em}{0ex}}f\cap {S}^{2}=\u2329\phantom{\rule{0.166667em}{0ex}}{{x}^{0}\phantom{\rule{0.166667em}{0ex}}}^{2},{{x}^{1}\phantom{\rule{0.166667em}{0ex}}}^{2}\phantom{\rule{0.166667em}{0ex}}\u232a$. Obviously,
${S}^{d}\subset Ann\phantom{\rule{0.166667em}{0ex}}f$ for all
$d\ge 3$. Now, for a finite set
$X=\{\u2329\phantom{\rule{0.166667em}{0ex}}{v}_{1}\phantom{\rule{0.166667em}{0ex}}\u232a,\cdots ,\u2329\phantom{\rule{0.166667em}{0ex}}{v}_{r}\phantom{\rule{0.166667em}{0ex}}\u232a\}\subset \mathbb{P}{S}_{1}$ of
r (distinct) points,
$I\left(X\right)$ is the set of all (polynomial) multiples in
${S}^{\u2022}$ of the polynomial
$p\in {S}^{r}$ with roots precisely
$\u2329\phantom{\rule{0.166667em}{0ex}}{v}_{1}\phantom{\rule{0.166667em}{0ex}}\u232a,\cdots ,\u2329\phantom{\rule{0.166667em}{0ex}}{v}_{r}\phantom{\rule{0.166667em}{0ex}}\u232a$. Hence, by the apolarity lemma, for every homogeneous
$p\in Ann\phantom{\rule{0.166667em}{0ex}}f\cap {S}^{r}$ that has
r (distinct) roots
$\u2329\phantom{\rule{0.166667em}{0ex}}{v}_{1}\phantom{\rule{0.166667em}{0ex}}\u232a,\cdots ,\u2329\phantom{\rule{0.166667em}{0ex}}{v}_{r}\phantom{\rule{0.166667em}{0ex}}\u232a$, we have
$f\in \u2329\phantom{\rule{0.166667em}{0ex}}{{v}_{1}}^{2},\dots ,{{v}_{r}}^{2}\phantom{\rule{0.166667em}{0ex}}\u232a$. It easily follows that
f can be decomposed as a sum of squares of appropriate scalar multiples of
${v}_{1},\dots {v}_{r}$. For instance,
$p:={{x}^{0}\phantom{\rule{0.166667em}{0ex}}}^{2}{{x}^{1}\phantom{\rule{0.166667em}{0ex}}}^{2}\in Ann\phantom{\rule{0.166667em}{0ex}}f\cap {S}^{2}$ gives rise to the decomposition
3.2. The Classically Known Results on Maximum Rank
From the elementary theory of quadratic forms, known since long time, follows that the rank of a quadratic form $f\in {S}_{2}$, equals the rank of its representing matrix with respect to whatever given (ordered) basis. Hence, the maximum rank equals $dim{S}_{1}$, that is, the number of indeterminates (if ${S}_{\u2022}$ is considered as a ring of polynomials).
To find the maximum rank of binary forms of given degree, apolarity is very effective. Indeed, let us consider the following simple description of the apolar ideal of a binary form.
Proposition 1. Let $f\in {S}_{d}\setminus \left\{0\right\}$, with $dim{S}_{1}=2$. Then, $Ann\phantom{\rule{0.166667em}{0ex}}f$ is generated by a form $a\in {S}^{s}$ and a form $b\in {S}^{d+2s}$ for some integer $s\le (d+2)/2$.
Proof. See [
14], Theorem 1.44(iv). □
As reported in [
14], this classical result is due to Macaulay. When
$dim{S}_{1}=2$, the ideal of a set of
r distinct points in
$\mathbb{P}{S}_{1}$ is generated by a homogeneous form. If the form
a in the statement of Proposition 1 is squarefree, from Lemma 1 follows that the rank of
f is
s, and some appropriate pairwise nonproportional roots of
a in
${S}_{1}$ give the linear forms of a
dth power sum decomposition. Taking also into account that
a and
b must be coprime, since they generate an ideal that contains
${S}^{d+1}$, we also deduce that if
a is not squarefree, then the rank is
$d+2s$ and for every finite subset
$X\subset \mathbb{P}{S}_{1}$ there exists a
dth power sum decomposition such that
$\u2329\phantom{\rule{0.166667em}{0ex}}v\phantom{\rule{0.166667em}{0ex}}\u232a\notin X$ for every linear form
v in it.
From the above, it is clear that the rank of f is at most d, and can be d only if $Ann\phantom{\rule{0.166667em}{0ex}}f$ contains the square of a linear form. Given a basis $\left({x}_{0},{x}_{1}\right)$ of ${S}_{1}$, the apolar ideal of ${{x}_{0}}^{d1}{x}_{1}$ contains ${{x}^{1}}^{2}$, with $\left({x}^{0},{x}^{1}\right)$ being the dual basis. We conclude that the maximum rank of binary forms of degree $d>0$ is d.
In [
9], Sections 96 and 97, one finds that the maximum ranks of ternary and quaternary cubics are 5 and 7, respectively. As we mentioned in the Introduction, beyond these classical results, only two new cases have been recently worked out: the maximum rank is 7 for ternary quartics (which has been determined in [
5,
17]) and 10 for ternary quintics (see [
11,
18]).
To our knowledge, no other values for the maximum rank have been determined to date; the known values can therefore be summarized in the following
Table 1.
3.3. Elementary Bounds on Maximum Waring Rank
Let us recall a common geometric viewpoint on Waring rank. Given
$f\in {S}_{d}\setminus \left\{0\right\}$, we have a point
$\u2329\phantom{\rule{0.166667em}{0ex}}f\phantom{\rule{0.166667em}{0ex}}\u232a\in \mathbb{P}{S}_{d}$ and we have to express
f as sum of
dth powers. The set of (spans of)
dth powers of linear forms is an algebraic variety in
$\mathbb{P}{S}_{d}$: the image of the embedding
$\mathbb{P}{S}_{1}\to \mathbb{P}{S}_{d}$,
$\u2329\phantom{\rule{0.166667em}{0ex}}v\phantom{\rule{0.166667em}{0ex}}\u232a\mapsto \u2329\phantom{\rule{0.166667em}{0ex}}{v}^{d}\phantom{\rule{0.166667em}{0ex}}\u232a$. This embedding turns out to be equivalent to a much studied embedding: the Veronese embedding (also called
duple embedding: see, e.g., [
16], Chapter I, Exercise 2,12). Its image is sometimes called the Veronese variety (see [
1] 4.3.7). The problem of finding a power sum decomposition of
f is equivalent to the problem of finding a set of points
X in the Veronese variety such that
$f\in \u2329\phantom{\rule{0.166667em}{0ex}}X\phantom{\rule{0.166667em}{0ex}}\u232a$, that is, such that the point
$\u2329\phantom{\rule{0.166667em}{0ex}}f\phantom{\rule{0.166667em}{0ex}}\u232a$ lies in the projective span of
X (one has to take into account that a scalar multiple of a
dth power is a
dth power as well, since
$\mathbb{K}$ is algebraically closed). Since the Veronese variety spans
$\mathbb{P}{S}_{d}$, the Waring rank is well defined for all forms, and it is at most
where
$n:=dim{S}_{1}$ and the parenthesized notation stands for the binomial coefficient. This gives an elementary upper bound on rank (which could easily be slightly lowered, but we are not interested in doing this here).
Let us now consider the union
U of all projective spans of
r distinct points in the Veronese variety. Clearly, for every
$\u2329\phantom{\rule{0.166667em}{0ex}}f\phantom{\rule{0.166667em}{0ex}}\u232a\notin U$, the rank of
f is greater than
r. Elementary tools of algebraic geometry allow one to estimate the dimension of the Zariski closure
$\overline{U}$ of
U (see [
1], 4.9.5 or [
16], Chapter I, Section 2, p. 10), which is called the
$(r1)$th secant variety (of the Veronese variety). Roughly speaking, the set of all groups of
r points in the Veronese variety, which has dimension
$n1=dim\mathbb{P}{S}_{1}$, is of dimension
$r(n1)$. Since most of these groups spans a subspace of dimension
$r1$, the expected dimension of
$\overline{U}$ is
$rn1$. When this number does not reach the dimension
$\left(\genfrac{}{}{0pt}{}{d+n1}{n1}\right)1$ of the entire space
$\mathbb{P}{S}_{d}$, there exist forms with rank greater than
r. Hence, the maximum rank in
${S}_{d}$ is at least
(where the external parentheses denote the upper integer part; a similar notation
$\lfloor \dots \rfloor $ will be used for the lower integer part).
Note that when
$d=2$, the lower bound is
$\lceil (n+1)/2\rceil $, meanwhile the maximum rank is
n. It is also worthy of being mentioned that the estimate of
$dim\overline{U}$ fails for
$d=2$ (and
$n,r\ge 2$). More generally, that estimate fails when most points in
$\overline{U}$ lie on infinitely many spans. In this case, the dimension of the secant variety drops, and the now classical theorem by Alexander and Hirschowitz gives the complete list of
$n,d,r$ for which this happens (we refer the reader to the exposition in [
4]). It turns out that for
$d\ge 3$ the above lower bound can be raised by one in exactly four cases:
$(n,d)\in \left\{\right(3,4),(4,4),(5,3),(5,4\left)\right\}$.
Given $n,d$, if r is the least value for which $\overline{U}=\mathbb{P}{S}_{d}$ then, by some basic algebrogeometric considerations which we skip here, for all $\u2329\phantom{\rule{0.166667em}{0ex}}f\phantom{\rule{0.166667em}{0ex}}\u232a$ in a nonempty open Zariski subset of $\mathbb{P}{S}_{d}$, f is actually of rank r. In this situation, it is customary to say that r is the rank of a generic form in ${S}_{d}$. In the context of tensor rank, the striking outcome of the Alexander–Hirschowitz theorem is indeed the exact value of the rank of a generic form (the lower bounds in the exceptional cases are only some of the many consequences). In the following sections, we review the enhanced lower and upper bounds that have been found recently.
4. Lower Bounds
To find a good lower bound on the set of the symmetric ranks of all symmetric tensors over $\mathbb{K}$ of given order d and dimension n, it suffices to find a form in ${S}_{d}$, when $dim{S}_{1}=n$, with high Waring rank.
The (few) lower bounds which we are aware of have been obtained by finding some special forms of high rank. Since the rank of a generic form gives a lower bound, the challenge is to exceed it. In this section, we present the special forms of high rank that have given the best known lower bounds.
4.1. What Monomials Tell Us
To begin with, let us consider binary forms, for which ranks are quite well understood. The maximum rank of degree d binary forms is d, while the rank of a generic degree d binary form is $\lfloor (d+2)/2\rfloor $ (it can be deduced from Proposition 1). Moreover, a degree d binary form of maximum rank can be turned into a monomial by a change of coordinates. Hence, for binary forms, the maximum rank is reached by monomials. For quadrics, whose rank is obviously well understood, the maximum rank of monomials is two (unless $dim{S}_{1}\le 1$), meanwhile the maximum rank is reached by generic forms (and equals $dim{S}_{1}$, that from a polynomial viewpoint is the number of indeterminates).
The rank of all monomials has been determined by Carlini, Catalisano e Geramita in [
19]. In dimension three, it turns out that the monomial
$x{y}^{s}{z}^{s}$ is of rank
${(s+1)}^{2}$ and
$x{y}^{s1}{z}^{s}$ of rank
$s(s+1)$. This gives a lower bound that asymptotically approaches
${d}^{2}/4$ for the rank of ternary forms of degree
d, while a generic form has rank asymptotically approaching
${d}^{2}/6$. According to [
12], Proposition 3.4, the asymptotic estimate of maximum rank for ternary forms is actually
${d}^{2}/4$. When the number of variables is four or greater, the maximum rank of monomials does not exceed the rank of a generic form of the same degree.
In view of the above, a first guess on maximum rank could be that the maximum rank is reached either by monomials or by generic forms. However, for ternary quartics, for which the maximum rank is known from Kleppe’s master thesis [
5,
17], it exceeds by one the maximum rank of both the monomials and the generic forms. The maximum rank exceeds both the maximum rank, of the monomials and of the generic forms, for ternary quintics too (see [
11,
18]). Buczyński and Teitler in [
11] also found forms in more than three variables with rank exceeding by one the rank of generic forms. An improvement by one might seem not too exciting but, at least for ternary forms, one cannot hope to go much farther. Indeed, the upper bound given in [
12], Proposition 3.3 shows that the maximum rank of a degree
d ternary form can exceed the maximum rank of monomials by at most
d. Thus, the initial guess may be modified by expecting that the maximum rank could only slightly exceed the maximum rank of either monomials or generic forms.
For a detailed discussion on maximum rank of monomials, we refer the reader to [
20]. Let us now outline how the rank of monomials has been bounded from below, and how that technique has been enhanced by Buczyński and Teitler, to exceed the previously known lower bounds on maximum ranks.
4.2. What Hilbert Functions Tell Us
The
Hibert function of a graded module
${\oplus}_{d\in \mathbb{Z}}{M}_{d}$ over the graded
$\mathbb{K}$algebra
${S}^{\u2022}$ can be simply defined as the function that on each
$d\in \mathbb{Z}$ takes value
${dim}_{\mathbb{K}}{M}_{d}$. This is a fundamental tool in algebraic geometric, and is still much studied. To let readers who are not acquainted with algebraic geometry get a taste of the fundamental nature of Hilbert function, let us mention that the degree and the dimension of an algebraic set can easily be get from a naturally associated Hilbert function. More precisely, let
$X\subseteq \mathbb{P}\left({S}_{1}\right)$ be the set of all points on which some system of homogenous forms in
${S}^{\u2022}$ vanishes. Then, the Hilbert function
${H}_{X}$ of
${S}^{\u2022}/I\left(X\right)$ coincides with a polynomial
${p}_{X}$ (the
Hilbert polynomial of
X) for all sufficiently large degrees. The degree of
${p}_{X}$ gives the dimension
n of
X and
$n!$ times the leading coefficient of
${p}_{X}$ gives the degree of
X. In the case when
X is a finite set of
r points, which by Lemma 1 is of our interest here, we get that
${H}_{X}\left(d\right)=r$ for all sufficiently large
d. Below, we take a few lines to directly show this fact in an elementary way; readers who are interested in the general properties of Hilbert functions can find them in many basic textbooks of algebraic geometry (e.g., in [
16], Chapter I, Section 7).
Let
$X=\left\{\u2329\phantom{\rule{0.166667em}{0ex}}{v}_{1}\phantom{\rule{0.166667em}{0ex}}\u232a,\dots ,\u2329\phantom{\rule{0.166667em}{0ex}}{v}_{r}\phantom{\rule{0.166667em}{0ex}}\u232a\right\}\subseteq \mathbb{P}\left({S}_{1}\right)$ be a set of
r points and let
${H}_{X}$ be the Hilbert function of
${S}^{\u2022}/I\left(X\right)$. The degree
d component of that quotient is
${S}^{d}/I{\left(X\right)}_{d}$, where
$I{\left(X\right)}_{d}={S}^{d}\cap I\left(X\right)$ is the space of degree
d forms that vanish on
X. From the evaluation in Equation (
4), it follows that
$I{\left(X\right)}_{d}={\u2329\phantom{\rule{0.166667em}{0ex}}{{v}_{1}}^{d},\dots ,{{v}_{r}}^{d}\phantom{\rule{0.166667em}{0ex}}\u232a}^{\perp}$ (this fact has been already noticed in the proof of Lemma 1). It follows that
Note that it is easy to find
${l}^{1},\dots ,{l}^{r1}\in {S}^{1}$ such that
${l}^{i}\left({v}_{i}\right)=0$ and
${l}^{i}\left({v}_{r}\right)\ne 0$ for each
$i\in \{1,\dots ,r1\}$. It follows that when
$d\ge r1$ the hyperplane
${({{l}^{1}}^{dr+2}{l}^{2}\cdots {l}^{r1})}^{\perp}<{S}_{d}$ contains
${{v}_{1}}^{d},\dots ,{{v}_{r1}}^{d}$, but not
${{v}_{r}}^{d}$. In a similar way, it can be found hyperplanes that do not contain a given
${v}_{i}$ but contain all
${v}_{j}$ with
$j\ne i$. This shows that
${{v}_{1}}^{d},\dots ,{{v}_{r}}^{d}$ are linearly independent, hence
Therefore, ${H}_{X}\left(d\right)=r$ for all sufficiently large d ($\ge r1$, in this case).
4.3. What Hyperplane Sections Tell Us
Let
${\overline{S}}^{\u2022}={S}^{\u2022}/I$ be the quotient of
${S}^{\u2022}$ by a homogeneous ideal
I and suppose that
$\overline{l}\in {\overline{S}}^{1}$ is not a zero divisor in
${\overline{S}}^{\u2022}$. Then, the multiplication by
$\overline{l}$ in
${\overline{S}}^{\u2022}$ injects each homogeneous component
${\overline{S}}^{d}$ into
${\overline{S}}^{d+1}$. Let
H and
${H}^{\prime}$ be, respectively, the Hilbert functions of
${\overline{S}}^{\u2022}$ and of its quotient
${\overline{S}}^{\u2022}/\left(\overline{l}\right)$ over the ideal generated by
$\overline{l}$. Thus, we have
Consequently, we also have
In this way, relevant properties of H can be deduced from properties of ${H}^{\prime}$.
The quotient ${\overline{S}}^{\u2022}/\left(\overline{l}\right)$ is naturally isomorphic to the quotient of ${S}^{\u2022}$ over the ideal $I+\left(l\right)$, with $l\in {S}^{1}$ being a representative of $\overline{l}$. When I is the ideal of an algebraic set $X\subseteq \mathbb{P}\left({S}^{1}\right)$ (that is, the set of all points where some set of homogeneous elements of ${S}^{\u2022}$ vanish), the algebraic set defined by $I+\left(l\right)$ (that is, the set of all points where all the homogeneous elements of $I+\left(l\right)$ vanish) is the intersection of X with the hyperplane given by l. From a geometric viewpoint, the idea is that features of hyperplane sections of X give relevant information on X. This idea is ubiquitous in algebraic geometry.
Let us see what we get in the case of our interest. When
X is a finite set of
r points, if a product of homogeneous elements
$xy$ vanishes on
X but
y does not, then
x must vanish on some point of
X. Conversely, if
x vanishes on some point of
X, it is not difficult to find a nonzero
y in some
${S}^{d}$ such that
$xy$ vanishes on
X. Therefore, to find
$\overline{l}\in {S}^{\u2022}/I\left(X\right)$ that is not a zero divisor is to find an hyperplane that does not meet
X. In this case, since we know that
H takes values
r for
$d\ge r1$, we conclude that
with
${H}^{\prime}$ being the Hilbert function of
${S}^{\u2022}/(I\left(X\right)+\left(l\right))$.
4.4. Rank of Monomials
The result we have just discussed, in conjunction with the apolarity lemma gives a way to bound the rank of
$f\in {S}_{d}$ from below. If we fix
$l\in {S}^{1}\setminus \left\{0\right\}$, for every finite set of
r points
X such that
$I\left(X\right)\subseteq Ann\phantom{\rule{0.166667em}{0ex}}f$ we obviously have
$I\left(X\right)+\left(l\right)\subseteq Ann\phantom{\rule{0.166667em}{0ex}}f+\left(l\right)$. Denoting by
${H}_{f}$ the Hilbert function of
${S}^{\u2022}/(Ann\phantom{\rule{0.166667em}{0ex}}f+\left(l\right))$, we have
${H}^{\prime}\left(i\right)\ge {H}_{f}\left(i\right)$ for all
i, hence
Taking into account Lemma 1, we have that every power sum decomposition $f={{v}_{1}}^{d}+\cdots +{{v}_{r}}^{d}$ for which no $\u2329\phantom{\rule{0.166667em}{0ex}}{v}_{i}\phantom{\rule{0.166667em}{0ex}}\u232a$ lies on the hyperplane ${\u2329\phantom{\rule{0.166667em}{0ex}}l\phantom{\rule{0.166667em}{0ex}}\u232a}^{\perp}$, has at least $b\left(l\right)$ summands. On this basis, a first rough idea to find a lower bound on the rank of f is to find the minimum $b\left(l\right)$, with $\u2329\phantom{\rule{0.166667em}{0ex}}l\phantom{\rule{0.166667em}{0ex}}\u232a$ varying in an infinite subset of $\mathbb{P}\left({S}^{1}\right)$ such that each point of $\mathbb{P}\left({S}_{1}\right)$ lies on at most a finite number of the hyperplanes ${\u2329\phantom{\rule{0.166667em}{0ex}}l\phantom{\rule{0.166667em}{0ex}}\u232a}^{\perp}$ (e.g., one may take an irreducible curve in $\mathbb{P}\left({S}^{1}\right)$ contained in no hyperplane). However, Carlini, Catalisano and Geramita followed another interesting path.
For whatever
X, the ideal
$\left(I\left(X\right):l\right)=\left\{g\in {S}^{\u2022}:gl\in I\left(X\right)\right\}$ is clearly the ideal of the set
${X}^{\prime}:=X\setminus \mathbb{P}{\u2329\phantom{\rule{0.166667em}{0ex}}l\phantom{\rule{0.166667em}{0ex}}\u232a}^{\perp}$. Hence, for the Hilbert function
${H}_{f}^{\prime}$ of
${S}^{\u2022}/\left((Ann\phantom{\rule{0.166667em}{0ex}}f:l)+\left(l\right)\right)$,
cannot exceed the number of points in
${X}^{\prime}$, and consequently the number of points in
X, for whatever
X. This holds for whatever choice of
$l\in {S}^{1}\setminus \left\{0\right\}$, which therefore can be chosen to maximize the above sum.
Now, let us consider a positive degree monomial $f={{x}_{1}}^{{a}_{1}}\cdots {{x}_{n}}^{{a}_{n}}$ for a given basis $\left({x}_{1},\dots ,{x}_{n}\right)$ of ${S}_{1}$. Let $\left({x}^{1},\dots ,{x}^{n}\right)$ be the dual basis in ${S}^{1}$, and note that a monomial ${{x}^{1}}^{{b}_{1}}\cdots {{x}^{n}}^{{b}_{n}}$ is apolar to f if and only if ${b}_{i}>{a}_{i}$ for some i. Moreover, for two different (monic) monomials m and ${m}^{\prime}$ that are not apolar to f, we have that ${\partial}_{m}f$ and ${\partial}_{{m}^{\prime}}f$ cannot be proportional. This easily implies that $Ann\phantom{\rule{0.166667em}{0ex}}f$ is the ideal generated by ${{x}^{1}\phantom{\rule{0.166667em}{0ex}}}^{{a}_{1}+1},\dots ,{{x}^{n}\phantom{\rule{0.166667em}{0ex}}}^{{a}_{n}+1}$.
With no loss of generality, we can assume
${a}_{1}\le \cdots \le {a}_{n}$, and let
${a}_{i}$ be the first nonzero exponent. It is quite easy to recognize that
$\left(Ann\phantom{\rule{0.166667em}{0ex}}f:{x}^{i}\right)$ is generated by
and
$\left(Ann\phantom{\rule{0.166667em}{0ex}}f:{x}^{i}\right)+\left({x}^{i}\right)$ by
To calculate that for the Hilbert function
${H}_{f}^{\prime}$ of
${S}^{\u2022}/\left((Ann\phantom{\rule{0.166667em}{0ex}}f:{x}^{i})+\left({x}^{i}\right)\right)$, one has
is not difficult (and quite easy if one is familiar with Hilbert functions).
Since we are concerned with lower bounds on rank, we could end here the subsection. However, to agree that
$r:=\left({a}_{i+1}+1\right)\cdots \left({a}_{i+1}+1\right)$ is actually the rank of
f is quite easy because, similar to what is smartly remarked in [
19], we have that
is the ideal of a set of
r distinct points and is contained in
$Ann\phantom{\rule{0.166667em}{0ex}}f$.
4.5. Beyond Monomials and Generic Forms
As anticipated before, once the rank of monomials has been determined, one can find ternary monomials with rank much higher than the rank of generic forms of the same degree (which is known by the Alexander–Hirschowitz theorem). When the number of indeterminates is four or greater, the rank of generic forms can not be exceeded by monomials. We give now a brief account of how Buczyński and Teitler were able to beat both ternary monomials and generic quaternary forms, with one and the same argument. A more informative description can be directly found in [
11].
Let $f\in {S}_{d}$ and $l\in {S}^{1}\setminus \left\{0\right\}$. In the calculations before, the sum of all values of the Hilbert function of quotient algebras of the type ${S}^{\u2022}/(I+\left(l\right))$ turned out to be useful. When $I=Ann\phantom{\rule{0.166667em}{0ex}}f$, that sum bounds from below the number of summands of a decomposition whose linear forms are outside the hyperplane ${\u2329\phantom{\rule{0.166667em}{0ex}}l\phantom{\rule{0.166667em}{0ex}}\u232a}^{\perp}$. When $I=(Ann\phantom{\rule{0.166667em}{0ex}}f:l)$, that sum directly bounds from below the rank of f. Note that the sum under consideration is nothing but the dimension of ${S}^{\u2022}/(I+\left(l\right))$ as a $\mathbb{K}$vector space. We can give the following useful description of this dimension, using a relation between $I+\left(l\right)$ and $(I:l)$ that one often encounters when dealing with hyperplane sections.
When
${S}^{\u2022}/I$ is finitedimensional, we have
Since the kernel of the homomorphism
is
$(I:l)$, we deduce that
When
$I=Ann\phantom{\rule{0.166667em}{0ex}}f$,
${S}^{\u2022}/I$ is called the
apolar algebra, and its dimension the
apolar length of
f. They can be denoted by
${A}_{f}$ and
$alf$. Note that even when
$I=(Ann\phantom{\rule{0.166667em}{0ex}}f:l)$,
${S}^{\u2022}/I$ is an apolar algebra. Indeed, for whatever
$x,y\in {S}^{\u2022}$, we have
that is,
$(Ann\phantom{\rule{0.166667em}{0ex}}f:y)=Ann\phantom{\rule{0.166667em}{0ex}}{\partial}_{y}f$. Hence, when
$I=(Ann\phantom{\rule{0.166667em}{0ex}}f:l)$, the quotient
A is the apolar algebra of
${\partial}_{l}f$.
From the formula $(Ann\phantom{\rule{0.166667em}{0ex}}f:y)=Ann\phantom{\rule{0.166667em}{0ex}}{\partial}_{y}f$ and the fact that apolarity is a perfect pairing in every degree, we get another interesting fact: the apolar length of f equals the dimension of the vector space of all ${\partial}_{y}f$ with $y\in {S}^{\u2022}$.
We end up with
the former being a lower bound on the number of summands of a decomposition of
f whose linear forms are outside the hyperplane
${\u2329\phantom{\rule{0.166667em}{0ex}}l\phantom{\rule{0.166667em}{0ex}}\u232a}^{\perp}$, and the latter a lower bound on the rank of
f. Based on these remarks, a good knowledge of apolar algebras, which can be obtained for instance from [
14], clearly gives precious information on lower bounds on rank.
A first obvious step is to try to maximize
$al{\partial}_{l}fal{\partial}_{{l}^{2}}f$. To use a coordinate description, let us fix dual bases
$\left({x}_{1}\dots ,{x}_{n}\right)$ and
$\left({x}^{1}\dots ,{x}^{n}\right)$ of
${S}_{1}$ and
${S}^{1}$. To get
$al{\partial}_{{l}^{2}}f=0$ we choose
$l={x}_{1}$ and consider a form
Then,
${\partial}_{l}f=g$, and we have to choose
g with maximum apolar length. From [
14], we can find the value of that maximum and learn that it is reached by a generic
g (that is, for all
g in a suitable nonempty open set in the Zariski topology). In conclusion, there exist degree
d forms in
n indeterminates with rank not less than the maximum apolar length of degree
$d1$ forms in
$n1$ indeterminates. Surprisingly, that maximum equals the maximum rank of degree
d monomials when
$n=3$, and the rank of generic forms of degree
d when
$n=4$ and
d is odd.
When
$(Ann\phantom{\rule{0.166667em}{0ex}}f:l)=Ann\phantom{\rule{0.166667em}{0ex}}{\partial}_{l}f$ is considered instead of
$Ann\phantom{\rule{0.166667em}{0ex}}f$ to get the lower bound, one might hope that for some special
f some of the linear forms might be forced to lie on the hyperplane
${\u2329\phantom{\rule{0.166667em}{0ex}}l\phantom{\rule{0.166667em}{0ex}}\u232a}^{\perp}$. However, for geometric reasons, we expect that forms of high rank have many decompositions, which can therefore easily escape out of the hyperplane. Note also that, when
f is a monomial
${{x}_{1}}^{{a}_{1}}\cdots {{x}_{n}}^{{a}_{n}}$,
$alfal{\partial}_{{x}^{i}}f=al{\partial}_{{x}^{i}}fal{\partial}_{{{x}^{i}}^{2}}f$ whenever
${a}_{i}\ne 0$, so that we have no loss in cutting out the part on the hyperplane. This might give some indication on why in the high rank examples found in [
11] the first thing considered is to raise the value of
$alfal{\partial}_{{x}^{i}}f$For a form f of the type ${x}_{1}g+k$, one can raise $alfal{\partial}_{{x}^{1}}f$ by one by lowering $alg$ by one, but at the cost of lowering ${al}_{{x}^{1}}fal{\partial}_{{{x}_{1}}^{2}}f$ too. This causes a problem for decompositions that have some linear forms on the hyperplane ${\u2329\phantom{\rule{0.166667em}{0ex}}{x}^{1}\phantom{\rule{0.166667em}{0ex}}\u232a}^{\perp}$. The idea was to show that for suitable choices of the form k, such decompositions must have at least two forms on the hyperplane. To this end, note that if a decomposition of f involve exactly one linear form v that lies in ${\u2329\phantom{\rule{0.166667em}{0ex}}{x}_{1}\phantom{\rule{0.166667em}{0ex}}\u232a}^{\perp}$, then $f{v}^{d}$ has a decomposition with all linear factors outside that hyperplane. Using the fact (pointed out before) that the apolar length equals the dimension of the vector space of all derivatives, it turns out that $al(f{v}^{d})al{\partial}_{{x}^{1}}(f{v}^{d})$ can be kept hight for whatever choice of v. This is incompatible with the fact that for some v, $f{v}^{d}$ has a decomposition with all linear factors outside ${\u2329\phantom{\rule{0.166667em}{0ex}}{x}^{1}\phantom{\rule{0.166667em}{0ex}}\u232a}^{\perp}$.
In view of determination of maximum rank, the Buczyński–Teitler lower bound is particularly interesting because, in conjunction with the upper bound in [
18], shows that the maximum rank of ternary quintics is ten. Now, at the end of [
12], Introduction, a possible guess for the maximum rank for ternary forms of an arbitrarily given degree
d is outlined, and if it is correct then the Buczyński–Teitler lower bound is the best possible for
d odd (and
$n=3$). In [
13], the lower bound for ternary forms given by monomials is raised by one for even degrees too, and if the guess in [
12] is correct, it cannot be improved further. Basically, the lower bound in [
13] follows the second advice in [
11], Remark 19, but uses a more specific example, similar to that in [
11], Theorem 18 (which gives the lower bound of ten for ternary quintics), and the arguments are of a purely algebraic nature (do not involve geometric dimension counts).
6. Summary
Let us summarize the state of the knowledge on the maximum rank
${r}_{\mathrm{m}ax}(n,d)$ of forms of degree
$d>0$ in
n variables, which has been presented in this article. For ternary forms:
For
$n\ge 4$, we have the lower and upper bounds given by the rank
${r}_{\mathrm{g}en}$ of generic forms and its double, which hold in general for the rank with respect to a variety. The enhancements obtained in [
11,
23] allow raising by one the lower bound when
d is odd, and lower the upper bound by one. Let us mention that, according to the list in [
1], 5.4.1, the exceptional cases from the Alexander and Hirschowitz’s theorem,
$(3,4),(4,4),(5,3),(5,4)$, are included in the special cases for which the upper bound
$2{r}_{\mathrm{g}en}$ can be lowered
by two according to Theorem 3.9 in [
23].
The other special cases are a bit more cumbersome to be detected, so we do not take them into account in the following summary:
with
To give a more concrete idea of the above values, we explicitly report some ranges in
Table 2, which enrich
Table 1.
Let us conclude by recalling that, if the symmetric rank was well understood, it would be easy to determine ${r}_{\mathrm{m}ax}(n,d)$. That is why the techniques invented to find ${r}_{\mathrm{m}ax}(n,d)$ may hopefully indicate some ways to understand tensor rank, which would be a considerable achievement, because of the recently recognized high applicative interest of this topic.