Indicatively, assume on a periodic basis, and in particular at the beginning of each update period, that an updater discusses his/her payoff with another randomly-chosen inspectee, who is not necessarily an updater himself/herself. If the two interacting inspectees have equal payoffs, then the updater retains his/her strategy. If, however, they have a payoff gap, then the updater is likely to revise his/her strategy, subject to how significant their payoffs’ variance is.
Regarding the inspector’s response, we no longer consider his/her strategy to be the choice of the inspection frequency (recall the inspector’s dilemma in the standard game setting whether to inspect or not). Instead, we take into account the overall effort the inspector devotes to his/her inspection activity. In particular, we identify this generic term as the fraction of the available budget that he/she invests on his/her objective, assuming that the inspection budget controls every factor of his/her effectiveness (e.g., the inspection frequency, the no-detection probability, the false alarms, etc.).
At each update event, we assume that the inspector is limited to the same finite, renewable available budget B. Without experiencing any policy-adjusting costs, he/she aims at maximising his/her payoff against each different distribution of the inspectees’ strategies at the least possible cost. Additionally, we assume that at each time point, he/she is perfectly informed about the inspectees’ collective behaviour. Therefore, we treat the inspector as a rational, straightforward, payoff maximising player. This suggestion is described as the best response principle.
Under this framework, the crime distribution in the population of the inspectees is subject to evolutionary pressure over time. Thus, the term evolutionary is introduced to describe the inspection game. It turns out that more efficient strategies gradually become dominant.
3.1. Analysis
We begin our analysis by assuming that the inspectees choose their strategies within a finite bounded set of strategies , generating increasingly illegal profits. Their group’s state space is then the set of sequences of non-negative integers , denoting the occupation frequency of strategy . Equivalently, it is the set of sequences of the corresponding relative occupation frequencies , where .
We consider a constant number of inspectees, namely we have for each group’s state n. Provided that the population size N is sufficiently large (formally valid for through the law of large numbers), we approximate the relative frequencies with , denoting the probabilities with which the strategies are adopted. To each strategy i we assign an illegal profit , characterizing compliers, and a strictly increasing punishment fine , with . We assume that is constant, namely that ’s form a one-dimensional regular lattice.
As explained, the inspector has to deal with an evolving crime distribution in the population of the inspectees,
. We define the set of probability vectors
, such that:
We introduce the inner product notation to define the group’s expected (average) illegal profit by . Respectively, we define the group’s expected (average) punishment fine by . We also define the inspector’s invested budget against crime distribution by and the inspector’s efficiency by . The last function measures the probability with which a violator is detected given that the inspector invests budget b. To depict a plausible scenario, we assume that perfect efficiency cannot be achieved within the inspector’s finite available inspection budget B (namely, the detection probability is strictly smaller than one, ).
Assumption 1. The inspector’s efficiency, , is a twice continuously differentiable, strictly increasing, strictly concave function, satisfying: An inspectee who plays strategy either escapes undetected with probability and obtains an illegal profit or gets detected with probability and is charged with a fine . Additionally, every inspectee receives a legal income r, regardless of his/her strategy being legal or illegal.
Therefore, to an inspectee playing strategy
i, against the inspector investing budget
b, we assign the following inspectee’s payoff function:
Accordingly, we need to introduce a payoff function for the inspector investing budget b against crime distribution . Recall that the inspector is playing against a large group of inspectees and intends to fight their collective illegal behaviour. For his/her macroscopic standards, the larger the group is, the less considerable absolute values corresponding to a single agent (i.e., ) are.
To depict this inspector’s subjective evaluation, we introduce the inspector’s payoff function as follows:
where
κ is a positive scaling constant and
, respectively
, denotes the expected (average) illegal profit, respectively the expected (average) punishment fine, at time
t. Without loss of generality, we set
. Note that the inspector’s payoff always obtains a finite value, including the limit
.
As already mentioned, an updater revises his/her strategy with a switching probability depending on his/her payoff’s difference with the randomly-chosen individual’s payoff, with whom he/she exchanges information. Then, for an updater playing strategy
i and exchanging information with an inspectee playing strategy
j, we define this switching probability by
, for a timespan
, where:
and
is an appropriately-scaled normalization parameter.
This transition process dictates that in every period following an update event, the number of inspectees playing strategy i is equal to the corresponding sub-population in the previous period, plus the number of inspectees having played strategies and switching into strategy i, minus the number of inspectees having played strategy i and switching into strategies .
Hence, we derive the following iteration formula:
which can be suitably reformulated, taking the limit as
, into an equation resembling the well-known replicator equation (see, e.g., [
33]):
Remark 1. We have used here a heuristic technique to derive Equation (
9)
, bearing in mind that we consider a significantly large group of interacting individuals (formally valid for the limiting case of an infinitely large population). A rigorous derivation can be found in [34]. In agreement with our game setting (i.e., the myopic hypothesis), (
9) is not a best-response dynamic. However, it turns out that successful strategies, yielding payoffs higher than the group’s average payoff, are subject to evolutionary pressure. This interesting finding of our setting, which is put forward in the above replicator equation, simply states that although the inspectees are not considered to be strictly rational maximizers (but instead myopic optimizers), successful strategies propagate into their population through the imitation procedure. This characteristic classifies (
9) into the class of the payoff monotonic game dynamics [
35]. Before proceeding further, it is important to state that our setting is quite different from the general setting of standard evolutionary game theory. Unlike standard evolutionary games, in our approach there are no small games of a fixed number of players through which successful strategies evolve. On the contrary, at each step and throughout the whole procedure, there is only one
players game taking place (see also the game in
Section 5).
Regarding the inspector’s interference, the best response principle states that at each time step, against the crime distribution he/she confronts, the inspector aims to maximise his/her payoff with respect to his/her available budget:
On the one hand, the inspector chooses his/her fine policy strategically in order to manipulate the evolution of the future crime distribution. On the other hand, at each period, he/she has at his/her disposal the same renewable budget
B, while he/she is not charged with any policy adjusting costs. In other words, the inspector has a period-long planning horizon regarding his/her financial policy. Therefore, he/she instantaneously chooses at each step his/her response
b that maximises his/her payoff (
6) against the prevailing crime distribution.
Let us define the inspector’s best response (optimum budget), maximising his/her payoff (
6) against the prevailing crime distribution, by:
Having analytically discussed the inspectees’ and the inspector’s individual dynamic characteristics, we can now combine them and obtain a clear view of the system’s dynamic behaviour as a whole.
In particular, we substitute the inspector’s best response (optimum budget)
into the system of ordinary differential equations (ODEs) (
9), and we obtain the corresponding system governing the evolution of the non-cooperative game described above:
Recall that through (
12) we aim to investigate the evolution of illegal behaviour within a large group of interacting, myopically-maximising inspectees (bureaucrats) under the pressure of a rationally-maximising inspector (incorruptible superior).
Without loss of generality, we set
. Let us also introduce the following auxiliary notation:
Proposition 1. A probability vector is a singular point of (
12)
, namely it satisfies the system of equations:if and only if there exists a subset , such that for , and for . Proof. For any
such that
,
, System (
12) reduces to the same one, but only with coordinates
(notice that
I must be a proper subset of
S). Then, for the fixed point condition to be satisfied, we must have
for every
.
☐
The determination of the fixed points defined in Proposition 1 and their stability analysis, namely the deterministic evolution of the game, clearly depend on the form of . One can identify two control elements that appear in and thus govern the game dynamics; the functional control and the control parameter B. We have set the fine to be a strictly increasing function, and we consider three eventualities regarding its convexity; (i) linear; (ii) convex; (iii) concave. To each version we assign a different inspector’s punishment profile. Indicatively, we claim that a convex fine function reveals an inspector who is lenient against relatively low collective violation, but rapidly jumps to stricter policies for increasing collective violation. Contrarily, we claim that a concave fine function reveals an inspector who is already punishing aggressively even for a relatively low collective violation. Finally, we assume that a linear fine function represents the ‘average’ inspector. We also vary in each case the constant (linear)/increasing (convex)/decreasing (concave) gradient of function . Accordingly, we vary the size of the finite available budget B. The different settings we establish with these control parameter variations and, therefore, the corresponding dynamics we obtain in each case have clear practical interpretation providing useful insight into applications. For example, the fine function can be, and usually is, defined by the inspector himself (think of different fine policies when dealing with tax evasion), while the level of budget B is decided from the benevolent principal by whom the inspector is employed. The detection efficiency is not regarded as an additional control since it characterizes the inspector’s behaviour endogenously. However, say the inspector has an excess of budget, then he/she could invest it in improving his/her expertise (e.g., his/her technical know-how) that is related with his/her efficiency indirectly. Then, he/she could improve . We do not engage with this case.
3.2. Linear Fine
Equivalently to (
11), for a linear fine
, the inspector’s best response (optimum budget) can be written as:
We check from (
15) that we cannot have
for every
, since at least for
, it is
. However, depending on the size of
B, we may have
for every
.
Then, it is reasonable to introduce the following notation:
where
is not necessarily deliverable, i.e.,
may not belong to
. One should think of this critical value
as a measure of the adequacy of the inspector’s available budget
B, namely as the ‘strength’ of his/her available budget. Obviously, if
, the inspector benefits from exhausting all his/her available budget when dealing with
, while, if
, the inspector never needs to exhaust
B in order to achieve an optimum response.
Theorem 2. Let Assumption 1 hold. The unit vectors , , lying on the vertices of the d-simplex, are fixed points of (
12)
. Moreover:- (i)
If , there is additionally a unique hyperplane of fixed points, - (ii)
If and , there are additionally infinitely many hyperplanes of fixed points,
Proof. In any case, the unit probability vectors
satisfy (
14), since by definition of
, it is
,
, whilst it is
.
The setting we introduce with Assumption 1 ensures that
is a continuous, non-decreasing, surjective function. In particular, we have that
is strictly increasing in
and
is strictly increasing in
. Hence:
- (i)
When
, there is a unique
satisfying
. This unique
is generated by infinitely many probability vectors,
, forming a hyperplane of points satisfying (
14).
- (ii)
When
and
, every
satisfies
. Each one of these infinitely many
is generated by infinitely many probability vectors,
, forming infinitely many hyperplanes of points satisfying (
14).
☐
We refer to the points as pure strategy fixed points, to emphasize that they correspond to the group’s strategy profiles such that every inspectee plays the same strategy i. Accordingly, we refer to the points , , as mixed strategy fixed points, to emphasize that they correspond to the group’s strategy profiles such that the inspectees are distributed among two or more available strategies.
Before proceeding with the general stability results, we present the detailed picture in the simplest case of three available strategies generating increasingly illegal profits including compliance.
Figure 2 implies a budget
B such that the inspector never exhausts it. For a relatively low
B or for an overly lenient
(see
Figure 2a and
Figure 3a), the pure strategy fixed point
is asymptotically stable. Increasing though
B or, accordingly, toughening up
(see
Figure 2b), a hyperplane of asymptotically-stable mixed strategy fixed points appears. Depending on the critical parameter
, we may have infinitely many hyperplanes of asymptotically stable mixed strategy fixed points (see
Figure 3b,d). Finally, keeping
B constant, the more we increase the slope of
, the closer this(ese) hyperplane(s) moves towards compliance (see
Figure 2c and
Figure 3c,e). We generalize these results into the following theorem.
Theorem 3. Let Assumption 1 hold. Consider the fixed points given by Theorem 2. Then:
(i) the pure strategy fixed point is a source, thus unstable; (ii) the pure strategy fixed points , , are saddles, thus unstable; (iii) the pure strategy fixed point is asymptotically stable when ; otherwise, it is a source, thus unstable; (iv) the mixed strategy fixed points , are asymptotically stable.
☐
3.3. Convex/Concave Fine
Let us introduce the auxiliary variable
. As usual, using the inner product notation, we define the corresponding group’s expected/average value by
, where
. Then, equivalently to (
11) and (
15), the inspector’s best response/optimum budget can be written as:
For every
, let us introduce as well the parameter:
Lemma 1. For a convex fine, is strictly decreasing in i for constant j (or vice versa), while for a concave fine, is strictly increasing in i for constant j (or vice versa). Furthermore, for a convex fine, is strictly decreasing in for constant , while for a concave fine, is strictly increasing in for constant .
Theorem 4. Let Assumption 1 hold. The unit vectors , , lying on the vertices of the d-simplex are fixed points of (
12)
. Moreover, there may be additionally up to internal fixed points , living on the support of two strategies , uniquely defined for each pair of strategies; they exist given that the following condition applies respectively: Proof. In any case, the unit probability vectors
satisfy (
14), since by the definition of
, it is
,
, whilst it is
,
.
Consider a probability vector
satisfying (
14), such that
,
,
and
. Then, from Proposition 1,
should satisfy
,
, namely the fraction
should be constant
, and equal to
.
To satisfy this, according to Lemma 1, the complement set may not contain more than two elements, namely the distributions may live on the support of only two strategies.
For such a distribution
, such that
and
,
, where
, we get:
The setting we introduce with Assumption 1 ensures that
is a continuous, non-decreasing, surjective function. In particular, we have that
is strictly increasing in
and
is strictly increasing in
. Then, for any
to exist, namely for (
20) to hold in each instance, the following condition must hold respectively:
☐
We refer to the points as double strategy fixed points, to emphasize that they correspond to the group’s strategy profiles, such that the inspectees are distributed between two available strategies.
Again, we present the detailed picture in the simplest case of three available strategies generating increasingly illegal profits including compliance. Like above, in
Figure 4 and
Figure 5, we observe how the interplay of the key parameters
B and
affect the game dynamics.
The general pattern is similar to
Figure 2 and
Figure 3. Initially, the pure strategy fixed point
appears to be asymptotically stable (see
Figure 4a–c and
Figure 5a), while gradually, either increasing
B or toughening up
, this unique asymptotically stable fixed point shifts towards compliance. However, the shift here takes place only through double strategy fixed points and not through hyperplanes of fixed points. In particular, for a concave
, it occurs only through the fixed point
living on the support of the two border strategies (see
Figure 4d–j), while for a convex
, it occurs through the fixed points
,
, namely only through the fixed points living on the support of two consecutive strategies (see
Figure 5b–j).
Proposition 2. For a convex fine: (i) the set of double strategy fixed points contains at most one fixed point living on the support of two consecutive strategies; (ii) there is at most one pure strategy fixed point satisfying: Proof. (i) Assume there are two double strategy fixed points
, both living on the support of two consecutive strategies. According to Theorem 4, both of them should satisfy (
19). However, since it is
, then from Assumption 1, it is also
, and since the fine is convex, then from Lemma 1, it is also
. Overall, we have that:
which contradicts the initial assumption.
(ii) Assume there are two pure strategy fixed points
,
, both satisfying (
22). However, since it is
, then from Assumption 1, it is also
, and since the fine is convex, then from Lemma 1, it is also
. Overall, we have that:
which contradicts the initial assumption.
☐
We generalize the results that we discussed above on the occasion of
Figure 4 and
Figure 5, into the following theorem.
Theorem 5. Let Assumption 1 hold. Consider the fixed points given by Theorem 4. Then:For a concave fine: (i) the pure strategy fixed point is a source, thus unstable; (ii) the pure strategy fixed points , are saddles, thus unstable; (iii) the double strategy fixed points are saddles, thus unstable; (iv) the double strategy fixed point is asymptotically stable; (v) the pure strategy fixed point is asymptotically stable when does not exist; otherwise, it is a source, thus unstable.
For a convex fine: (i) the pure strategy fixed point is a source, thus unstable; (ii) the double strategy fixed points , are saddles, thus unstable; the double strategy fixed points are asymptotically stable; the pure strategy fixed points , are saddles, thus unstable; the pure strategy fixed point is a source, thus unstable. * When no exists, the pure strategy fixed point satisfying (22) is asymptotically stable.
☐