1. Introduction
In the context of the time independent Schrödinger equation:
the elements of a symmetry group G commute with
H [
1,
2,
3]:
Commutativity requires that H and G do not mix each other’s invariant spaces so that the eigenfunctions of H must (barring accidental degeneracy) transform according to the irreducible representations of the symmetry group. The methods of group representations can help determine the eigenvalues and eigenvectors of H.
Suppose that
H does not commute with a particular finite unitary group G but that the behavior of
H under the group elements is known and given by:
Then a symmetrized operator,
Hsym, is projected from
H by symmetry averaging, summing over all transformations and dividing by the order of the group:
The symmetrized operator commutes with all elements of G:
and is Hermitian.
The original operator is the sum:
where
HR is the remainder. If
Hsym is nonzero and the remainder
HR is sufficiently small in some sense, perturbation theory may give acceptable approximations to the eigenvalues and eigenvectors of
H. A number of groups may be used for any particular system, but the optimum choice for analysis would generally be the largest group yielding a nonzero
Hsym. In general, the larger the group, the more the reduction in analytical requirements.
This approach may be compared to established perturbation methods in which the total Hamiltonian is expressed as the sum of a high-symmetry zero-order Hamiltonian plus a symmetry lowering perturbation: . Symmetry-averaging is systematic, ensures that HR has no identity component and provides a relationship between reduced matrix elements of Hsym and matrix elements of H.
It is usual to have a matrix representation, [
H], of the Hamiltonian over a suitable basis of a Hilbert space, either as an initial definition or as a first step in analysis. If the basis is symmetry-adapted to the group G the matrix of the symmetrized operator [
Hsym] will be reduced although [
H] will not be. By a corollary to the symmetry-generation theorem [
4,
5,
6], the reduced matrix elements of [
Hsym] are simple averages of certain elements of [
H] making the construction of [
Hsym] much more efficient than the computationally demanding projection. Therefore, even though
Hsym may not be the Hamiltonian of a real molecule [
Hsym] is readily determined.
Symmetry-averaging as a projection is presented in the next section and the use of the symmetry-adapted basis for symmetry-averaging is described in the third section. In the fourth section, the simplest example of a heteronuclear diatomic molecule corresponding to a two-vertex graph with one edge is presented. In the fifth section, symmetry-averaging of a three vertex system to two different groups is described. The sixth section is the conclusion.
2. Symmetry-averaging
Using the methods of finite group algebra, the identity for a finite group is given in terms of simple projectors as:
where the sum is over the
M irreducible representations of the group and the projectors
eα are commuting orthogonal idempotents:
This is the notation of Littlewood [
7] and Weyl [
8] in which irreducible representations are indexed by lower case Greek letters. It is convenient to take the lead term,
α = 1 to correspond to the totally symmetric irreducible representation occurring in every group. Then the corresponding projector is the symmetrizer:
where the sum is over all g elements of the group. The effect of the symmetrizer on an arbitrary operator is to project out the totally symmetric component:
where the “∘” operation is defined as shown to be the appropriate linear combination of equivalence transformations. Then
Hsym commutes with all elements of the group. The original operator is expressed as the sum of
Hsym and a remainder:
Since the right hand side of Equation 10 is the average of all the system orientations produced by the group transformations, the term symmetry-averaging is appropriate.
3. Symmetry-adapted basis
The elements of a basis symmetry-adapted to a group G transform irreducibly under the elements of the group according to:
where
α denotes an irreducible representation of G,
and
r index the rows and columns of the
f(
α) ×
f(
α) matrix [
Ga]
α,
ρ distinguishes repeated irreducible representations, and
ω identifies the complete space. The representation of the group on a symmetry-adapted basis is completely reduced. A symmetry-adapted basis is suitably-conditioned if all of the |
ω;
ραr〉,
ρ = 1,…,
f(
ω;
α), transform according to identical matrices [
Ga]
α. Here,
f(
ω;
α) is the number of times the
α irreducible representation occurs, given by the usual character formula. Matrix elements of an operator
H are expressed on this basis as:
If the operator is defined as a matrix [
H]
ω on some defining basis: {|
ωi〉,
i = 1,…,
f(
ω)}, it is necessary to perform the transformation to the symmetry-adapted basis:
On a symmetry-adapted basis, the elements of an operator,
Hsym, that commutes with G satisfy the relation [
9]:
where the far right factor is a reduced matrix element. Since these elements are zero unless
α equals
and
r equals
, and the reduced matrix elements are independent of
r, the matrix is partitioned into scalar submatrices on the symmetry-adapted basis. Reordering the basis to group together elements with the same
α and
r gives a partition with
f(
α) identical
f(
ω;
α)×
f(
ω;
α) blocks down the diagonal resulting in a factored secular polynomial. These blocks are the reduced matrices. Suppose, for example, that a basis symmetry-adapted to the point group C
3v contains one
A1 and two
E irreducible representations:
then the corresponding representation of the group element
Ga is:
The matrix of an operator that commutes with the group has the form:
A reordering of the basis to group together elements with the same
r:
results in further reduction to block diagonal form:
illustrating the power of group theory.
The partitioning of matrix (17) into scalar blocks illustrates the invariances that give rise to the symmetry-generation theorem [
6]. If
Hsym is obtained by symmetry-averaging of
H, as in equation (10), and if the basis is suitably conditioned, then the trace of a block in matrix (17) is the sum of traces of corresponding transformed bocks of the matrix of
H:
. Then as a corollary to the symmetry-generation theorem the reduced matrix elements of
Hsym on a suitably conditioned symmetry-adapted basis are given by:
That is, the reduced matrix elements of the symmetry-averaged operator are simple averages of particular matrix elements of the original operator over the symmetry-adapted basis.
4. Two dimensional matrix
The simplest case is a two vertex graph corresponding to a two dimensional Hermitian matrix:
The symmetry group to be used is of order two and consists of the identity and the transposition (12):
The effect of the transposition on the matrix is:
so that symmetry averaging yields:
Let
then:
as in Equation 6. Clearly, the smaller the value of
d the closer [
H] is to [
Hsym] and the more appropriate this choice of symmetry. A symmetry-adapted basis is:
where {|1〉, |2〉} is the defining basis. The matrix of
Hsym on this basis is:
and for the difference,
HR:
Evidently, as a perturbation [HR] makes only a second order correction.
Consider the numerical case
a = −40.0,
b = −30.0 and
c = −5.00. The exact eigenvalues to three significant figures are:
From [Hsym] the zero-order eigenvalues are −40.0 and −30.0. With the second order correction, these become −42.5 and −27.5.
5. Three-vertex examples
The next two illustrations refer to a weighted three-edge, three-vertex graph:
corresponding to the matrix:
This graph could be embedded in Euclidean three-space so that the groups considered would be point groups such as Cs, D3 or C3v. If the effect of these operations is to permute the vertices of the graph, the Hamiltonian matrix will be permuted. These are the examples considered here. If the weights associated with two of the vertices, say numbers one and two, are closer to each other than to the third, then it would be appropriate to consider a reflection plane through vertex three and bisecting the edge between vertex one and two. The effect on the Hamiltonian matrix would be exchange of index one with two. On the other hand, if the weights associated with all three vertices are similar, then the point group C3v would be suitable, transforming the Hamiltonian matrix by the permutation group S3.
S2 Averaging
First consider the choice of G as the two-fold symmetry group
. Using symmetry averaging on the original the invariant matrix is:
where
Using a symmetry adapted basis will give the reduced matrix directly. This basis is:
and the matrix on this basis is:
The reduced matrix elements of
are then:
so that the invariant matrix on this basis is:
The repetition of the A irreducible representation gives rise to a 2×2 block that must be diagonalized while the non-repeated eigenvalue can simply be read off the diagonal.
S3 Averaging
The S
3 symmetry averaged matrix is:
where
and
It is unnecessary, however, to obtain this matrix, since [
H] can be directly transformed to a symmetry-adapted basis. Assuming that the elements of the defining basis are centered at the vertices and transform in the same way, a symmetry-adapted basis is:
where the C
3v names of the irreducible representations of have been used. On this basis, the matrix is:
From equation 20 the reduced matrix elements of [
Hsym]
ω are then:
so that the invariant matrix is:
and the difference matrix is:
It is instructive to consider some numerical values. The first case appears to have little symmetry.
The symmetry averaged component on the symmetry adapted basis is:
with the difference matrix:
The first order solutions are −30.0, −19.0, and −11.0. The exact solutions to three significant figures are −34.4, −17.4, −8.18.
A second example may be considered more symmetrical since the elements of the difference matrix are smaller.
which gives the same symmetrized component as before on the symmetry adapted basis:
with a difference matrix of:
The first order solutions are −30.0, −15.4, and −14.6, close to the exact eigenvalues to three significant figures of −30.1, −15.4, −14.5. The elements of the difference matrix are smaller by a factor of ten then in the previous case.
6. Conclusion
Since symmetry-averaging by projection requires a sum of triple matrix products it becomes computationally impractical as the sizes of basis and symmetry group increase. As shown here this calculation is unnecessary on a suitably conditioned symmetry-adapted basis so that symmetry-averaging is practical even for large systems.
The examples given here are purposely extremely small and simple; nevertheless, important properties of symmetry averaging can be determined from them. First, even if the system seems to have no symmetry, symmetry-averaging can be useful. For the two-vertex example the symmetry determined eigenvalues are good to second order. In the S2 treatment of the three vertex example the exact eigenvalues separate into two groups similar to the approximate result. In the second S3 treatment, the smaller difference term gives an indication of how symmetry averaging can be useful. In this case, the approximate eigenvalues are within 1% or less of the exact values.
It is anticipated that ab initio and DFT calculations can profit from this theory. Fock and Kohn-Sham matrices would be reduced on a suitably-conditioned symmetry-adapted basis and the matrix elements determined by the method described here.
A more general extended approach is to project out all the different irreducible symmetry components of the parent H. The noninvariant remainder HR would then be the sum of several irreducible tensorial operators. The Wigner-Eckart theorem can then be used to evaluate the matrix elements of these operators. This approach is being developed.