Next Article in Journal
Back to the Future: Revisiting the Application of an Enzyme Kinetic Equation to Maize Development Nearly Four Decades Later
Next Article in Special Issue
A Generic Model Accounting for the Interactions among Pathogens, Host Plants, Biocontrol Agents, and the Environment, with Parametrization for Botrytis cinerea on Grapevines
Previous Article in Journal
Interaction of Preventive, Cultural, and Direct Methods for Integrated Weed Management in Winter Wheat
Previous Article in Special Issue
Evaluation of the ‘Irish Rules’: The Potato Late Blight Forecasting Model and Its Operational Use in the Republic of Ireland
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Effects of Quantitative Ordinal Scale Design on the Accuracy of Estimates of Mean Disease Severity

1
Division of Biometrics, Department of Agronomy, National Chung Hsing University, Taichung 40227, Taiwan
2
Department of Statistics and Information Science, Fu Jen Catholic University, New Taipei City 24205, Taiwan
3
Department of Plant Pathology, National Chung Hsing University, Taichung 40227, Taiwan
4
USDA-ARS-SEFTNRL, 21 Dunbar Road, Byron, GA 31008, USA
5
Graduate Institute of Food Safety, National Chung Hsing University, Taichung 40227, Taiwan
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Agronomy 2019, 9(9), 565; https://doi.org/10.3390/agronomy9090565
Submission received: 14 August 2019 / Revised: 8 September 2019 / Accepted: 17 September 2019 / Published: 19 September 2019

Abstract

:
Estimates of plant disease severity are crucial to various practical and research-related needs in agriculture. Ordinal scales are used for categorizing severity into ordered classes. Certain characteristics of quantitative ordinal scale design may affect the accuracy of the specimen estimates and, consequently, affect the accuracy of the resulting mean disease severity for the sample. The aim of this study was to compare mean estimates based on various quantitative ordinal scale designs to the nearest percent estimates, and to investigate the effect of the number of classes in an ordinal scale on the accuracy of that mean. A simulation method was employed. The criterion for comparison was the mean squared error of the mean disease severity for each of the different scale designs used. The results indicate that scales with seven or more classes are preferable when actual mean disease severities of 50% or less are involved. Moreover, use of an amended 10% quantitative ordinal scale with additional classes at low severities resulted in a more accurate mean severity compared to most other scale designs at most mean disease severities. To further verify the simulation results, estimates of mean severity of pear scab on samples of leaves from orchards in Taiwan demonstrated similar results. These observations contribute to the development of plant disease assessment scales to improve the accuracy of estimates of mean disease severities.

1. Introduction

In agricultural research, estimates of disease severity are needed to understand and predict yield loss due to disease, for monitoring and forecasting epidemics, for comparing treatments or genotypes for disease resistance, and so on. Accuracy of the mean disease severity is critical in these studies to ensure there is no failure of analysis [1]. Here, accuracy refers to the deviation between the true value of a quantity (in this case, the actual mean disease severity) and estimates of that value based on the sampling process [2]. The reliability (or precision) of this mean estimate is the extent of the deviations occurring when the mean severity is repeatedly estimated using the same sampling process. The definition of the mean accuracy should not be confused with the definition of accuracy used in measurement science and as applied to estimates of disease severity on individual specimens—where accuracy can be defined as the degree of closeness of estimated or measured severity to a recognized actual value [2,3].
When estimated visually, plant disease severity can be defined as the proportion of symptoms and/or signs of the disease at the scale of a plant organ, a plant, a field plot, or more rarely a field [2]. During visual severity estimation, an individual will assign a disease severity score to a specimen using one of several rating scales [4]. A special type of ordinal scale, a quantitative ordinal scale, which comprises a number of intervals of known numeric ranges, is commonly used when estimating plant disease severity. The intervals are most often based on the percent area with symptoms. The most widely used quantitative ordinal scale is the Horsfall-Barratt (HB) scale which divides the percent scale into 12 consecutive, logarithmically increasing (from 0 to 50% severity), and then decreasing intervals (from 50% to 100% severity) numbered 1 to 12, where 1 = 0, 2 = 0+–3, 3 = 3+–6, 4 = 6+–12, 5 = 12+–25, 6 = 25+–50, 7 = 50+–75, 8 = 75+–88, 9 = 88+–94, 10 = 94+–97, 11 = 97+–100, and 12 = 100% disease [5]. Several other similar scales have been developed that subdivide the percent scale into different numbers of intervals and interval sizes [6,7,8,9,10,11,12,13]. For subsequent data analysis, it is recommended that the mid-point of the scale interval be used if applying parametric methods [2,14], although the ordinal ratings may be analyzed directly if non-parametric methods are used [14].
Nutter and Esker [15] demonstrated the barely noticeable difference in the ability to differentiate two severities between 25 and 50%. Due to the resolution of many ordinal scales in this range, accurate estimates of severity using the 0–100% ratio scale (i.e., nearest percent estimate, NPE) are almost always going to be more informative compared with those from an ordinal scale [16], but the latter has the perceived advantage of being easier for raters to learn, and quicker and cheaper to implement. However, due to the structure of some ordinal scales, additional variability may be added into the samples [17]. This is because certain characteristics of quantitative ordinal scale design might cause the information to be distorted in the process of transforming NPE data into ordinal disease severity scores. To what extent is the information distorted, and what impact does it have on the accuracy of the mean value and variance of the sample mean? If the distortion is severe, it will result in inaccurate mean estimates. The consequence of inaccurate estimates of mean disease severity could be misleading when the estimates are used as predictors of crop loss, or for other decision-making purposes [2]. This issue must be understood before steps can be taken for practitioners in plant pathology to optimize the designs of ordinal scales in situations where the scales are preferred. Furthermore, an incorrect scale rating for a specimen estimate can exacerbate the inaccuracy [18].
The number of classes in an ordinal scale has been discussed by Kranz [19,20] and Hau et al. [21] in relation to scales with unequal-sized categories. Kranz [19] suggested that about seven classes were optimal. Furthermore, the underlying distribution of severities may have ramifications for the number of classes needed on a scale to accurately estimate the mean disease severity [21]. Kranz [22] believed that the maximum likely severity to be encountered should be considered, and the scale set accordingly. However, these studies neither compared these interval-scale estimates to nearest percent estimates (NPEs), nor did they investigate the effects of sample size on the accuracy of the resulting mean estimate of severity.
Several studies have compared different scales for their accuracy, precision, agreement, and reliability of disease severity estimates [23,24,25,26]. The studies demonstrate that estimates of disease severity using the HB scale are less precise compared to NPEs using the 0–100% ratio scale. Nita et al. [26] reported that estimates using a 5% increment scale were closer to NPE estimates compared to those estimates using the HB scale. They recommended using an equal-interval scale for estimating disease severity rather than the unequal-interval HB scale, at least where the use of a disease scale is preferred. Madden et al. [2] and Nutter and Esker [15] contend that equal-increment scales may be preferable to the HB scale. An advantage of equal-sized ordinal scales is that they can be treated as continuous data. However, to ensure accuracy of estimates of specimens with severity <10%, Bock et al. [4] suggested that even equal-interval ordinal scales should be designed to accommodate severities at the low end of the spectrum to minimize overestimation of low mean severities.
Several studies have used simulation modeling to compare ordinal and other scale types to study aspects of disease assessment. Hartung and Piepho [16] evaluated three different rating scales regarding the accuracy, precision, and time needed for scoring. Chiang et al. [27] explored issue related to the use of the disease severity index (DSI) on accuracy. The objective of these studies was to determine which assessment method was most accurate. Besides investigating the accuracy and precision of disease severity estimates, some studies have explored the impact of the relative values resulting from different rating scales using hypothesis testing [14,17,28,29,30]. The criteria used to compare rating scales was the power of the hypothesis test for comparing treatment means using simulated rater estimates relative to the means based on actual data. Chiang et al. [17] showed that linear quantitative ordinal scales with 5%–10% intervals and additional classes at lower severity appear to be equivalent to NPEs for hypothesis testing. Also, intervals in the mid-range not exceeding 10% was recommended. Furthermore, if there are too many classes, the assumed advantages of speed and simplicity offered by the scale will be lost. Thus, it was suggested that a 10% (not 5%) linear quantitative ordinal scale with sensitivity to low disease severity is preferable to a non-linear ordinal scales (e.g., HB scale) for assessing disease severity.
Few studies have determined scale characteristics that maximize the accuracy of estimates of mean disease severity. The purpose of this study is to compare various ordinal scales to the 0 to 100% ratio scale and explore the effect of the number of classes in an ordinal scale on the accuracy of the mean estimate. To determine the accuracy of the mean, we used the mean squared error (MSE) of the mean disease severity estimates for each scales type, applying a simulation approach as described by Hartung and Piepho [16] and Chiang et al. [27]. To validate these results, we assessed the severity of pear scab (caused by Venturia nashicola Tanaka & Yamamoto) in 6 pear orchards in Taiwan. Pear scab is a particularly destructive disease of pear in Taiwan and elsewhere [31,32,33,34]. Specifically, we asked the following questions, (i) What is the effect of equal or unequal intervals on the mean estimate? (ii) How does interval width affect the mean estimate? (iii) What is the optimal number of classes in a quantitative ordinal scale to maximize the accuracy of the mean? Finally, we used these results to develop a more accurate assessment scale for assessing the severity of pear scab.

2. Materials and Methods

2.1. Assessment Methods

We adopted a two-stage approach to explore the effects of ordinal scale structure (linear or non-linear) and class number on accuracy and imprecision. The scales used in the first stage to explore scale structure are as follows:
1. NPE: disease severity estimated by the raters to the nearest 1%.
2. HB: non-linear, unequal intervals with 12 categories: 0, 0+–3, 3+–6, 6+–12, 12+–25, 25+–50, 50+–75, 75+–88, 88+–94, 94+–97, 97+–99% and 100% [5].
3. EI10: a linear, 10-class 10% interval scale with equal intervals from 0% to 100%.
4. AM10: an amended linear 10% interval scale with 10 classes emphasizing low severities: 0, 0+–1, 1+–4, 4+–10, 10+–20, 20+–30, 30+–40, 40+–50, 50+–70, 70+–100% disease) [17]. Here, only severities <50% were emphasized because leaves often abscise if the disease becomes too severe, making it difficult to obtain samples with severity >50% [22]. Thus, only two classes (50+–70, 70+–100% disease) account for severities ≥50% for assessment in methods 4 through 7.
The scales used in the second stage to explore the effect number of classes are as follows:
5. AM20: AM10 but with 20% linear intervals for disease severities 10% to 50% (with 8 classes: 0, 0+–1, 1+–4, 4+–10, 10+–30, 30+–50, 50+–70, 70+–100% disease).
6. AM5: AM10 but with 5% intervals for disease severities 10% to 50% with 14 classes: 0, 0+–1, 1+–4, 4+–10, 10+–15, 15+–20, 20+–25, 25+–30, 30+–35, 35+–40, 40+–45, 45+–50, 50+–70, 70+–100% disease).
7. AM10f: AM10 but finer intervals at ≤10% disease with 11 classes: 0, 0+–1, 1+–3, 3+–6, 6+–10, 10+–20, 20+–30, 30+–40, 40+–50, 50+–70, 70+–100% disease).

2.2. Simulation Method

To obtain specimen estimates of disease severity, the simulation was performed according to a positively skewed beta-distribution. A beta-distribution, defining the distribution of a random variable on the closed unit interval 0–1 is commonly used for a proportion and can be made very flexible by choosing different shape parameters (α and β) [35,36,37]. In this study, the means (α/(α + β)) of the random variables of the beta-distributions (given the two chosen parameters: α and β) were set at 1%, 5%, 10%, 20%, 30%, and 40% severities to demonstrate a range of possible “actual” mean disease severities (%). Furthermore, we selected three different variance parameter settings (representing situations of large, medium, and small sample variation) for the fixed means to illustrate the effects of the variance between estimates with the different assessment methods (Figure 1).

2.3. Criterion for Comparison: Mean Squared Error (MSE)

The MSE takes into account both the variance (precision) and the bias (accuracy) of a mean severity, so it provides a useful method for comparing several assessment scales [38]. The MSE is calculated as follows:
MSE ( X ¯ ) = E ( X ¯ μ ) 2 = Variance ( X ¯ ) + [ Bias ( X ¯ ) ] 2
where μ is a population mean; E is, an average, a mathematical EXPECTATION (each amount is multiplied by the corresponding probability). X ¯ represents the average disease severity of the specimens.
Variance ( X ¯ ) = Variance ( X ) n
where Variance ( X ¯ ) is the variance of the sampling distribution of X ¯ ; and its square root is the standard error of the mean severity. Here, Bias ( X ¯ ) was used to investigate the effects of bias (over and underestimates) in estimates of the mean disease severities.
Bias ( X ¯ ) = E ( X ¯ ) μ

2.4. Simulation Framework

The simulation steps are as follows:
  • We simulated n sample size values (from 10 to 100 in increments of 5) from a beta-distribution with the preselected specific mean severity and variance for that sample (which might represent a plot in a field). These n simulated values on the continuous percentage scale, defined by the beta distribution of a random variable on the closed unit interval 0–1, represent the NPEs.
  • The resulting NPEs were converted to the appropriate classes for assessment methods 2–7. These scale data were subsequently converted to the appropriate midpoint value of each class for subsequent analysis [2].
  • The MSEs of mean disease severity estimates for each of the different scales were calculated (Equation (1)).
  • The corresponding variances and biases were calculated (equations 2 and 3, respectively).
  • The Monte Carlo simulation process was repeated 10,000 times. To present the results comparing assessment methods, we plot MSEs (or variance or bias) on the y-axis against sample size values (from n = 10 to 100 in n = 5 increments) on the x-axis at each of a range of “actual” mean disease severities (1%, 5%, 10%, 20%, 30%, and 40% [6 mean disease severities]) and disease severity variances (representing large, medium, and small variation [3 variances]). Thus, the results are presented in a montage (a 6 × 3 array of figures). For example, R1C2 (chart in row 1, column 2) indicates the chart in the second column in the first row, and presents the relationships between the MSE, variance, or bias of the mean disease severity estimates and sample size (n) for each of the different scales used, at an “actual” mean disease severity of 1%, and the medium variation.

2.5. Pear scab Assessment

Leaves of pear with scab were collected from six orchards in Dongshi District, Taichung County in Taiwan from September–November 2018. The data sets consist of 658, 389, 269, 453, 209, and 244 diseased leaves. Images of the leaves were captured with a digital camera (Sony CyberShot DSC-F717, 5.0 megapixels; SONY Corporation, Tokyo, Japan), and the area with symptoms of scab on each leaf was measured using Assess© V2.0 (APS Press, Image Analysis Software for Plant Disease Quantification) [39] to obtain “actual” values of the disease severities. Subsequently, we used the data sets to gauge the performance of the different quantitative ordinal scales described in order to assess the accuracy of estimates of mean disease severity. The procedures used were the same as described for the previous simulation steps. Briefly, simulations were undertaken using the estimated severity data to replace the values sampled from beta-distributions in step 1 of the simulation framework.
All statistical analyses, simulations, and graphics were performed in R [40]. We used the cloud for data storage. The data can be found at the following link: https://drive.google.com/drive/folders/1tqM7ZPyiCxETfsDrTRKgGgVzQ_ASs5jZ?usp=sharing

3. Results

3.1. Comparison to Determine the Effect of Scale Structure

Most MSEs decreased significantly as sample size (n) increased (Figure 2). Moreover, the decrease in magnitude was rapid as variation increases with the fixed mean severity (e.g., R2C1, R3C1, and R4C1 of Figure 2). NPEs most often had the smallest MSE, although, at higher mean severities (≥20%), there was little difference between the methods. Thus, overall, the performance of the 0% to 100% ratio scale was superior to all other scales tested when the MSE was used as the criterion for comparison. This is unsurprising because the process of simulating NPEs did not include any bias factor, so the mean of the NPEs would always be close to the “actual” severities. That is, using the NPE scale did not distort any information.
Comparing the four quantitative ordinal scales (Figure 2), MSEs were increased at lower severities (≤10%) when assessments were made using the EI10 scale (equal intervals of 10%). A sharp contrast exists between the effect of using the EI10 and the other assessment methods; e.g., R1C1, R1C2, R1C3, and R2C1 of Figure 2. Differences among MSEs using the other assessment methods (HB and AM10 scales) were minor at low severities (≤10%), but MSEs using the HB scale were larger at severities of 20% and 30%, especially when variation in the sample was small (e.g., R5C3 of Figure 2). The likely reason for this is that the HB scale intervals of 12+ to 25% and 25+ to 50% are so wide (13% and 25%, respectively) that when mid-points are taken for analysis they more often result in less accurate specimen estimates compared with what is achieved based on NPEs, the EI10 or the AM10 scales.
The results show that variance decreased as the sample size increased, with the initial decline being steep (Figure 3). There were no noticeable differences in variances among the different assessment methods. The decrease is inversely proportional to the square root of the sample size, and the pattern of change in relation to sample size is most often similar to those for the MSEs (Figure 2). Thus most of the differences in magnitudes of the MSEs can be ascribed to the variance (or precision) of the sample estimates, with the exception of the ordinal scale with large intervals (EI10) at severities of 1% or 5% which resulted in grossly inflated MSEs. These inflated MSEs when using the EI10 scale were due to biases that were greatest at severities of 1% and 5% as compared to the other scales (e.g., R1C1, R1C2, and R2C1 of Figure 4). Use of the HB scale resulted in large biases at severities of 20% and 30% (e.g., R4C3 and R5C3 of Figure 4). With the AM10 scale, biases occurred at a severity of 40%, but as with the other assessment scales at 40% severity, the bias was relatively small (<1%) (e.g., R6C2 and R6C3 of Figure 4).

3.2. Comparison to Determine the Effect of Number of Classes

The first stage determined that the AM10 scale (10% quantitative ordinal scale with additional classes to accommodate low disease severities) is the optimal ordinal scale type (Figure 2, Figure 3 and Figure 4). In the second stage, we investigated the effect of the number of classes (and therefore resulting interval widths). MSEs were greater when the interval width was 20% (AM20 scale) at mean severities of 10% or less (see R2C1, R2C2, R3C2, and R3C3 of Figure 5). However, at severities of 20% or more, the differences in MSEs between the scales with different class numbers are small (Figure 5). Both the scale with 5% intervals at ≥10% severity (AM5 scale) and the scale with 10% intervals at ≥10% severity (AM10 scale) had equivalent MSEs. Furthermore, MSEs were similar when using the scale that incorporated finer interval classes at <10% severity (AM10f scale) compared to when using the AM10 scale.
The relationship between variance and sample size at all levels of variation (large, medium, and small) was similar for all scale types and also similar to that for the MSE, regardless of class number (Figure 6). Only in some cases, for very small sample sizes, was the variance of the AM20 scale slightly greater compared to the other scales. Moreover, as noted above, most differences in the magnitude of the MSE can be attributed to the contribution made by its variance, or lack of precision, rather than bias.
Sample size had little or no effect on the bias, regardless of variation (large, medium, and small) (Figure 7). Using the AM20 scale at severities ≤30% resulted in the largest biases. Biases when using the AM10 scale were similar or only slightly higher when compared to those obtained when using the AM5 or AM10f scales.
Based on the MSE, variance, and bias, the performance of the AM10 scale is not inferior to that of the AM5 or AM10f scales. A linear structure is preferred. Too many classes in an ordinal scale could negate the assumed advantages of speed and simplicity offered by these scales. Based on these results and criteria, we recommend the AM10 scale structure as an optimal ordinal scale structure for estimating plant disease severity. Furthermore, the results indicate that scales with ≥7 classes are preferable, where severity ≤50% disease is emphasized.

3.3. Analysis of Severity of Pear Scab Data

Based on the image analysis measurements, the six plots had actual mean disease severities of 8%, 13%, 20%, 32%, 41%, and 49% with corresponding variances of 0.0091%, 0.0146%, 0.0146%, 0.0192%, 0.0287%, and 0.0231%, respectively (Figure 8).
The pattern of MSE, variance, and bias for a mean scab severity of 8% were similar to that for a mean severity of 10% with large variation in the simulated studies (i.e., R3C1 of Figure 2, Figure 3, Figure 4, Figure 5, Figure 6 and Figure 7); the EI10 and AM20 scales did not perform well. Moreover, the overall patterns for a mean scab severity of 13% are similar to that for a mean disease severity of 10% with medium variation in the simulated studies (R3C2 of Figure 2, Figure 3, Figure 4, Figure 5, Figure 6 and Figure 7); in this case, the HB and AM20 scales did not perform well. The pattern of MSE, variance, and bias for a mean scab severity of 20% is similar to that of a mean severity of 20% with medium variation in the simulated studies (R4C2 of Figure 2, Figure 3, Figure 4, Figure 5, Figure 6 and Figure 7); in this case the performance of the HB scale was particularly poor, with the greatest MSE and bias compared to other scales. The mean scab severity of 32% presented a similar pattern to those of a mean severity of 30% with medium variation in the simulated studies (R5C2 of Figure 2, Figure 3, Figure 4, Figure 5, Figure 6 and Figure 7); again, the HB scale had the greatest bias. With a mean scab severity of 41% or 49%, the performance of the AM5, AM10 and AM10f scales were all slightly inferior to the other scales based on all three criteria, but this is unlikely to be an issue as the mean scab severity was close to 50% and is usually in a range of less interest. Based on MSE, variance, and bias, the AM5, AM10, and AM10f scales all performed well at mean scab severities <41%. NPEs performed well at all mean disease severities. Thus, the trends for both simulated and empirical data are consistent.
We developed a ranking system using the simulated data and the pear scab assessment data to support the claim that NPEs, the AM5, AM10f, and AM10 scales had the lowest MSEs compared to the other methods (Table 1). For the seven assessment methods, we calculated the sum of MSEs from n = 10 to 100 in n = 5 increments at the specific mean severities and variances for the simulated and real data, resulting in aggregate MSEs at each mean severity for each variance, and for the pear scab assessment data. Subsequently, we summed these aggregate MSEs for each of the mean disease severities as an “overall” measure of accuracy. Based on the magnitude of these overall MSE data, we assigned a rank from 1 to 7 for each of the assessment methods, where 1 indicated the smallest MSE, and therefore the most accurate assessment method.
As expected, the smallest MSE (ranked #1) is observed when assessments are made using NPEs. Although the AM10 scale was consistently ranked as 3, 4 or 5, there is only a slight difference in the overall MSE values compared with the AM5 and AM10f methods (ranked #2, 3 or 4). However, too many divisions in an ordinal scale might negate the assumed advantages of speed and simplicity offered by an ordinal scale. The AM5 or AM10f scales both have more classes to differentiate (14 and 11 classes, respectively). Thus, the AM10 scale (10 classes) is recommended as the most appropriate ordinal scale design to use for disease severity assessment. In future studies where severity estimates of pear scab (or other diseases) are needed using a quantitative ordinal scale, the scale is recommended to conform to that described for the AM10 scale.

4. Discussion

Plant pathologists often obtain data on disease severity using visual assessments made by individuals who assign a disease severity score to each specimen in a sample using one of several rating scales [4]. Quantitative ordinal scales are usually preferred for convenience and speed of rating [2]. However, due to their structure, certain quantitative ordinal scales might introduce additional variability into the samples [17]. Thus, the objective of this study was to build on previous research [14,17] and to seek the “optimal” design for this scale type (i.e., equal or unequal classes, the width of intervals, and numbers of classes) so as to maximize the accuracy of estimates of mean disease severity.

4.1. “Optimal” Design for an Ordinal Scale

Our results indicate that an amended 10% interval ordinal scale (AM10) was the most appropriate ordinal scale design for researchers who desired to base disease severity estimation on an ordinal scale; the AM10 scale yielded a high level of accuracy in estimates of mean disease severity, while minimizing the number of classes (10 classes). In particular, the additional low classes included in the AM10 scale were more sensitive to the ranges of severity at which many diseases frequently occur and at which they needed to be estimated. The HB scale (and other lower resolution, non-linear ordinal scales), the EI10 scale (10 equal 10% intervals) scale, and the AM20 scale (20% linear intervals at severity ≥10%) generally had larger MSEs (disease severity estimates were not reliable and/or accurate as NPEs and the AM10 scale). Furthermore, when the AM5 scale (5% linear intervals at severity ≥10%) or the AM10f scale (an additional interval at severity ≤10%) were included for comparison, neither the MSE, variance, or bias indicated they had an advantage over the AM10 scale. Both the AM5 and AM10f scales would further negate the assumed advantages of simplicity (fewer classes) offered by ordinal scales. Additional findings of this study indicate that, when preparing quantitative ordinal scales for rating disease severity, scales with ≥7 classes were preferable where severities ≤50% disease was emphasized. Additional classes will be needed at severities >50%. These should be at equal 10% intervals based on the ability of raters to detect a just noticeable difference [15]. In contrast, Kranz [19] and Hau et al. [21] had previously conjectured that a minimum of 7 and 8 classes should be recommended for the 0% to 100% scale. The pear scab severity estimates we collected from the pear plots further support the simulation results.

4.2. Effects of Statistical Distributions of Diseased Leaves

There is generally a positively skewed distribution for the disease at low severities. Bock et al. [4] states that as estimates cannot be <0%; all estimates at low disease severities (<10%) are subject to a “barrier” at the low end (i.e., 0%). As disease approaches mid-range severity, it resembles a normal distribution. For example, the data set used to compare the severity of Septoria leaf blotch (SLB) in Chiang et al. [28] exhibit these characteristics. In our study, the “actual” mean disease severities had specified statistical distributions of diseased leaves (Figure 1). The different statistical distributions affected the MSE of mean disease severity estimates for each of the different scales used for estimation (Figure 2 and Figure 5). Thus, our results support the statement of Hau et al. [21] that the characteristics of the distribution of diseased leaves in the population and the number of classes in a disease scale affect the accuracy of the estimate of mean disease severity; very skewed populations rated using a scale with few categories (<7) result in inaccurate mean estimates. After 30 years, we tested and provided strong support for these claims using both simulation studies and empirical data sets.

4.3. Properties of the AM10 Quantitative Ordinal Scale

The AM10 scale is based on a 10% linear scale emphasizing severities ≤50% disease and had additional classes at low severities ≤10%. AM10 could be regarded as a linear ordinal scale with sensitivity emphasized at low disease severity (<10%). Although it is technically a non-linear scale, at severities ≥10%, it resembles a linear scale. So in this regard, the AM10 scale can be considered a hybrid scale, part linear, and part non-linear. The conclusions here are consistent with those of a previous study based on estimates of the severity of citrus canker by different raters [14]. Whereas the present study explored the issue of the accuracy of the mean disease severity, the earlier study compared the relative differences due to scale type when comparing treatments using a hypothesis test based on the probability of a type II error.
One thing to note is that when using the AM10 scale, larger biases may occur at severities of 40% because only two classes (50+–70, 70+–100% disease) account for severities ≥50%. Thus, as some diseases can exceed 50% severity, maintaining a linear 10% interval from 10% to 100% will help avoid undesirable bias in the upper range (but increasing the number of classes).

4.4. Mean Squared Error (MSE), Variance, and Bias

The criterion for comparison of the scales was the MSE of the mean disease severity estimates ( X ¯ ). The MSE ( X ¯ ) is separated into variance ( X ¯ ) (precision) and square of the bias ( X ¯ ). We found that most differences in the magnitudes of the MSE could be ascribed to its variance. Although biases due to scale type accounted for only a small portion of the MSE, it played a crucial role in comparison among assessment methods, especially at low disease severities (≤10%). Biases resulting from the use of ordinal disease assessment scales almost invariably resulted in overestimation (Equation (3) Bias ( X ¯ ) = E ( X ¯ ) μ > 0) compared to the actual disease severity. On the other hand, bias was not observed for NPEs. Disease estimates at low actual severities usually have a positively skewed distribution, and estimates of mean disease severity are inflated by one or a few very large values. When using EI10, the MSEs of the mean disease severity estimates were particularly high at severities of 1% and 5%. The critical factor here was the bias. When NPEs were transformed into values on a 10% equal-interval scale, the mean of individual units sampled was greater than that of the original NPEs. The result was the data were distorted after the transformation and could not represent the “actual” mean disease severity (%) in a population at low severities (1% or 5%) due to the positively skewed distribution noted above. Although the distribution of severities resembled a normal distribution as mean severity approaches 50%, there were only two classes (50+–70, 70+–100%) which include severities ≥50% for assessment methods 4 through 7; this resulted from the use of a coarse measurement scale at the right as opposed to a finer scale at the left side of a normal distribution.
Thus, for the fixed numbers of classes (10 classes) in the disease scales tested, the AM10 scale is recommended, despite slight bias at actual severities close to 40%, which seldom need to be considered for practical use [22]. Thus, the results presented here are of great value with regard to that range of disease most often observed in the field for many pathosystems. Thus, as noted in the previous section, where greater resolution is needed at severities >50%, a continued linear 10% scale is recommended.

4.5. Rationale for Simulation Studies

In this study, a two-stage approach was performed to firstly compare the scale structure (linear or non-linear), and secondly the number of classes (interval size) in a scale. We wanted to determine which scales minimized inaccuracy and imprecision. It is rational to explore the two factors in turn. We found that there was little variation among biases based on sample size, allowing us to conclude that biases are stable regardless of sample size (n). The reason for this phenomenon is that the simulation process was repeated 10,000 times, making it possible to average the 10,000 means of each individual sample size for each situation. If the simulation process was repeated fewer times, there would be extreme fluctuations in the biases (K-S Chiang, unpublished data). Therefore, more repetitions (in these cases 10,000 times) are required in order to compare various NPE and ordinal scale estimates.

4.6. Direction for Future Studies

In this study, the process of simulating the estimates did not take rater biases into account; only the effects of biases due to assessment method were studied. Although the effects of rater bias (over and underestimates) on estimates of disease severity of individual specimens on hypothesis testing using different assessment methods was previously explored [28,29], in this study, we were investigating the effect of the scale per se, on MSE, variance, and bias of the sample mean. Thus we focused on the issue of absolute values to provide information regarding which assessment method is preferable to use to provide the most accurate estimates of the mean disease severity. Taking rater biases into account for mean disease severity estimates will be an area for future studies.

5. Conclusions

We believe that the results of this study will contribute to the development of improved disease assessment quantitative ordinal scales. Use of better-designed ordinal scales will be helpful in improving the accuracy of disease severity estimates in plant disease epidemiology and related areas of research. The recommendations are general and applicable across most or many plant disease assessment situations.

Author Contributions

K.S.C., H.I.L., and J.R.T. conceived and designed the research. K.S.C., H.I.L., J.R.T., and C.H.B. analyzed the data and wrote the manuscript. K.S.C. and W.H.C. collected the data from the field. All authors approved the submission.

Funding

This research was supported by the Ministry of Science and Technology of Taiwan, R.O.C. (MOST 107-2313-B-005-031).

Acknowledgments

We are grateful to the reviewers for helpful comments which improved the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bock, C.H.; Chiang, K.S.; Del Ponte, E.M. Accuracy of plant specimen disease severity estimates: Concepts, history, methods, ramifications and challenges for the future. CAB Reviews: Pers. Ag. Vet. Sci. Nutr. Nat. Res. 2016, 11, 1–13. [Google Scholar] [CrossRef]
  2. Madden, L.V.; Hughes, G.; van den Bosch, F. The Study of Plant. Disease Epidemics; APS Press: St. Paul, MN, USA, 2007. [Google Scholar]
  3. Nutter, F.W., Jr. Disease assessment terms and concepts. In Encyclopedia of Plant. Pathology; Maloy, O.C., Murray, T.D., Eds.; John Wiley and Sons: New York, NY, USA, 2001; pp. 312–323. [Google Scholar]
  4. Bock, C.H.; Poole, G.; Parker, P.E.; Gottwald, T.R. Plant disease severity estimated visually, by digital photography and image analysis, and by hyperspectral imaging. Crit. Rev. Plant. Sci. 2010, 29, 59–107. [Google Scholar] [CrossRef]
  5. Horsfall, J.G.; Barrat, R.W. An improved grading system for measuring plant disease. Phytopathology 1945, 35, 655. [Google Scholar]
  6. Hunter, R.E.; Roberts, D.D. A disease grading system for pecan scab. Pecan Q. 1978, 12, 3–6. [Google Scholar]
  7. Koitabashi, M. New biocontrol method for parsley powdery mildew by the antifungal volatiles-producing fungus Kyu-W63. J. Gen. Plant. Pathol. 2005, 71, 280–284. [Google Scholar] [CrossRef]
  8. Kobriger, K.M.; Hagedorn, D.J. Determination of bean root rot potential in vegetable production fields of Wisconsin’s central sands. Plant. Dis. 1983, 67, 177–178. [Google Scholar] [CrossRef]
  9. Kora, C.; McDonald, M.R.; Boland, G.J. Epidemiology of Sclerotinia rot of carrot caused by Sclerotinia sclerotiorum. Can. J. Plant. Pathol. 2005, 27, 245–258. [Google Scholar] [CrossRef]
  10. Lazarovits, G.; Hill, J.; Patterson, G.; Conn, K.L.; Crump, N.S. Edaphic soil levels of mineral nutrients, pH, organic matter, and cationic exchange capacity in the geocaulosphere associated with potato common scab. Phytopathology 2007, 97, 1071–1082. [Google Scholar] [CrossRef] [PubMed]
  11. Nitzan, N.; Cummings, T.F.; Johnson, D.A.; Miller, J.S.; Batchelor, D.L.; Olsen, C.; Quick, R.A.; Brown, C.R. Resistance to root galling caused by the powdery scab pathogen Spongospora subterranea in potato. Plant. Dis. 2008, 92, 1643–1649. [Google Scholar] [CrossRef] [PubMed]
  12. Slopek, S.W. An improved method of estimating percent leaf area diseased using a 1 to 5 disease assessment scale. Can. J. Plant. Pathol. 1989, 11, 381–387. [Google Scholar] [CrossRef]
  13. Vieira, R.F.; Paula, T.J., Jr.; Carneiro, J.E.S.; Teixeira, H.; Queiroz, T.F.N. Management of white mold in type III common bean with plant spacing and fungicide. Trop. Plant. Pathol. 2012, 37, 95–101. [Google Scholar]
  14. Bock, C.H.; Gottwald, T.R.; Parker, P.E.; Ferrandino, F.; Welham, S.; van den Bosch, F.; Parnell, S. Some consequences of using the Horsfall-Barratt scale for hypothesis testing. Phytopathology 2010, 100, 1030–1041. [Google Scholar] [CrossRef] [PubMed]
  15. Nutter, F.W., Jr.; Esker, P.D. The role of psychophysics in phytopathology:The Weber-Fechner law revisited. Eur. J. Plant. Pathol. 2006, 114, 199–213. [Google Scholar] [CrossRef]
  16. Hartung, K.; Piepho, H.P. Are ordinal rating scales better than percent ratings? A statistical and “psychological” view. Euphytica 2007, 155, 15–26. [Google Scholar] [CrossRef]
  17. Chiang, K.S.; Liu, S.C.; Bock, C.H.; Gottwald, T.R. What interval characteristics make a good categorical disease assessment scale? Phytopathology 2014, 104, 575–585. [Google Scholar] [CrossRef] [PubMed]
  18. Bock, C.H.; Wood, B.W.; van den Bosch, F.; Parnell, S.; Gottwald, T.R. The effect of Horsfall-Barratt category size on the accuracy and reliability of estimates of pecan scab severity. Plant. Dis. 2013, 97, 797–806. [Google Scholar] [CrossRef] [PubMed]
  19. Kranz, J. Schätzklassen für Krankheitsbefall. Phytopath. Z. 1970, 69, 131–139. [Google Scholar] [CrossRef]
  20. Kranz, J. Measuring plant disease. In Experimental Techniques in Plant. Disease Epidemiology; Kranz, J., Rotem, J., Eds.; Springer: New York, NY, USA, 1988; pp. 35–50. [Google Scholar]
  21. Hau, B.; Kranz, J.; Konig, R. Fehler beim Schätzen von Befallsstärken bei Pflanzenanzenkrankheiten. Z. Pflanzenkrankh. Pflanzenschutz 1989, 96, 649–674. [Google Scholar]
  22. Kranz, J. A study on maximum severity in plant disease. Travaux dédiés à G. Viennot-Bourgin 1977, 16, 9–73. [Google Scholar]
  23. Bock, C.H.; Parker, P.E.; Cook, A.Z.; Gottwald, T.R. Visual rating and the use of image analysis for assessing different symptoms of citrus canker on grapefruit leaves. Plant. Dis. 2008, 92, 530–541. [Google Scholar] [CrossRef] [PubMed]
  24. Bock, C.H.; Parker, P.E.; Cook, A.Z.; Riley, T.; Gottwald, T.R. Comparison of assessment of citrus canker foliar symptoms by experienced and inexperienced raters. Plant. Dis. 2009, 93, 412–424. [Google Scholar] [CrossRef] [PubMed]
  25. Forbes, G.A.; Korva, J.T. The effect of using a Horsfall-Barratt scale on precision and accuracy of visual estimation of potato late blight severity in the field. Plant. Pathol. 1994, 43, 675–682. [Google Scholar] [CrossRef]
  26. Nita, M.; Ellis, M.A.; Madden, L.V. Reliability and accuracy of visual estimation of Phomopsis leaf blight of strawberry. Phytopathology 2003, 93, 995–1005. [Google Scholar] [CrossRef] [PubMed]
  27. Chiang, K.S.; Liu, H.I.; Bock, C.H. A discussion on disease severity index values: I. warning on inherent errors and suggestions to maximize accuracy. Ann. Appl. Biol. 2017, 171, 139–154. [Google Scholar] [CrossRef]
  28. Chiang, K.S.; Bock, C.H.; El Jarroudi, M.; Delfosse, P.; Lee, I.H.; Liu, H.I. Effects of rater bias and assessment method on disease severity estimation with regard to hypothesis testing. Plant. Pathol. 2016, 65, 523–535. [Google Scholar] [CrossRef]
  29. Chiang, K.S.; Bock, C.H.; Lee, I.H.; El Jarroudi, M.; Delfosse, P. Plant disease severity assessment—How rater bias, assessment method and experimental design affect hypothesis testing and resource use efficiency. Phytopathology 2016, 106, 1451–1464. [Google Scholar] [CrossRef] [PubMed]
  30. Chiang, K.S.; Liu, H.I.; Tsai, J.W.; Tsai, J.R.; Bock, C.H. A discussion on disease severity index values. Part II: Using the disease severity index for null hypothesis testing. Ann. Appl. Biol. 2017, 171, 490–505. [Google Scholar] [CrossRef]
  31. Cho, E.K.; Cho, W.T.; Lee, E.J. The causal organism of pear scab in Korea. Korean J. Mycol. 1985, 13, 263–265. [Google Scholar]
  32. Li, B.; Zhao, H.; Li, B.; Xu, X.M. Effects of temperature, relative humidity and duration of wetness period on germination and infection by conidia of the pear scab pathogen (Venturia nashicola). Plant. Pathol. 2003, 52, 546–552. [Google Scholar] [CrossRef]
  33. Park, P.; Ishii, Y.; Adachi, Y.; Kanematsu, S.; Ieki, H.; Umemoto, S. Infection behavior of Venturia nashicola, the cause of scab on Asian pears. Phytopathology 2000, 90, 1209–1216. [Google Scholar] [CrossRef]
  34. Sun, S.K. Fruit Diseases in Taiwan; Shih Way Publishers: Taichung, Taiwan, 2001. [Google Scholar]
  35. Carlin, B.P.; Louis, T.A. Bayes and Empirical Bayes Methods for Data Analysis, 2nd ed.; Chapman and Hall/CRC Press: Boca Raton, FL, USA, 2000. [Google Scholar]
  36. Gelman, A.; Carlin, J.B.; Stern, H.S.; Rubin, D.B. Bayesian Data Analysis; Chapman and Hall/CRC Press: Boca Raton, FL, USA, 1995. [Google Scholar]
  37. Liu, S.C.; Chiang, K.S.; Lin, C.H.; Chung, W.C.; Lin, S.H.; Yang, T.C. Cost analysis in choosing group size when group testing for Potato virus Y in the presence of classification errors. Ann. Appl. Biol. 2011, 159, 491–502. [Google Scholar] [CrossRef]
  38. Bain, L.J.; Engelhardt, M. Introduction to Probability and Mathematical Statistics, 2nd ed.; Duxbury Press: Pacific Grove, CA, USA, 1992; pp. 283–284. [Google Scholar]
  39. Lamari, L. ASSESS: Image Analysis Software for Plant Disease Quantification; APS Press: St. Paul, MN, USA, 2002. [Google Scholar]
  40. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2018. [Google Scholar]
Figure 1. A beta-distribution was used to simulate the random variables representing disease severity. The means of the random variables of the beta-distributions were set at 1%, 5%, 10%, 20%, 30%, and 40% severities to demonstrate a range of possible “actual” mean disease severities (%). Furthermore, we selected three different variance parameter settings (representing large, medium, and small variation) for the fixed means to illustrate the effect of the variance of estimates with the different assessment methods.
Figure 1. A beta-distribution was used to simulate the random variables representing disease severity. The means of the random variables of the beta-distributions were set at 1%, 5%, 10%, 20%, 30%, and 40% severities to demonstrate a range of possible “actual” mean disease severities (%). Furthermore, we selected three different variance parameter settings (representing large, medium, and small variation) for the fixed means to illustrate the effect of the variance of estimates with the different assessment methods.
Agronomy 09 00565 g001
Figure 2. The relationships between the mean squared error (MSE) of mean disease severity estimates and sample size (n) (from n = 10 to 100 in n = 5 increments) for each of the different scales used, over a range of “actual” mean disease severities (1%, 5%, 10%, 20%, 30%, and 40%) and variance (represented by large, medium, and small variation). The simulation process was repeated 10,000 times. Assessment methods: (i) NPE (nearest percentage estimate); (ii) HB (Horsfall-Barratt scale); (iii) AM10 (10% linear scale emphasizing severities ≤50% disease, and with additional intervals at severities <10%); (iv) EI10 (10% linear scale).
Figure 2. The relationships between the mean squared error (MSE) of mean disease severity estimates and sample size (n) (from n = 10 to 100 in n = 5 increments) for each of the different scales used, over a range of “actual” mean disease severities (1%, 5%, 10%, 20%, 30%, and 40%) and variance (represented by large, medium, and small variation). The simulation process was repeated 10,000 times. Assessment methods: (i) NPE (nearest percentage estimate); (ii) HB (Horsfall-Barratt scale); (iii) AM10 (10% linear scale emphasizing severities ≤50% disease, and with additional intervals at severities <10%); (iv) EI10 (10% linear scale).
Agronomy 09 00565 g002
Figure 3. The relationships between the variance of mean disease severity estimates and sample size (n) (from n = 10 to 100 in n = 5 increments) for each of the different scales used, over a range of “actual” mean disease severities (1%, 5%, 10%, 20%, 30%, and 40%) and variance (represented by large, medium, and small variation). The simulation process was repeated 10,000 times. Assessment methods: (i) NPE (nearest percentage estimate); (ii) HB (Horsfall-Barratt scale); (iii) AM10 (10% linear scale emphasizing severities ≤50% disease, and with additional intervals at severities <10%); (iv) EI10 (10% linear scale).
Figure 3. The relationships between the variance of mean disease severity estimates and sample size (n) (from n = 10 to 100 in n = 5 increments) for each of the different scales used, over a range of “actual” mean disease severities (1%, 5%, 10%, 20%, 30%, and 40%) and variance (represented by large, medium, and small variation). The simulation process was repeated 10,000 times. Assessment methods: (i) NPE (nearest percentage estimate); (ii) HB (Horsfall-Barratt scale); (iii) AM10 (10% linear scale emphasizing severities ≤50% disease, and with additional intervals at severities <10%); (iv) EI10 (10% linear scale).
Agronomy 09 00565 g003
Figure 4. The relationships between the bias of mean disease severity estimates and sample size (n) (from n = 10 to 100 in n = 5 increments) for each of the different scales used, over a range of “actual” mean disease severities (1%, 5%, 10%, 20%, 30%, and 40%) and variance (represented by large, medium, and small variation). The simulation process was repeated 10,000 times. Assessment methods: (i) NPE (nearest percentage estimate); (ii) HB (Horsfall-Barratt scale); (iii) AM10 (10% linear scale emphasizing severities ≤50% disease, and with additional intervals at severities <10%); (iv) EI10 (10% linear scale).
Figure 4. The relationships between the bias of mean disease severity estimates and sample size (n) (from n = 10 to 100 in n = 5 increments) for each of the different scales used, over a range of “actual” mean disease severities (1%, 5%, 10%, 20%, 30%, and 40%) and variance (represented by large, medium, and small variation). The simulation process was repeated 10,000 times. Assessment methods: (i) NPE (nearest percentage estimate); (ii) HB (Horsfall-Barratt scale); (iii) AM10 (10% linear scale emphasizing severities ≤50% disease, and with additional intervals at severities <10%); (iv) EI10 (10% linear scale).
Agronomy 09 00565 g004
Figure 5. The relationships between the mean squared error (MSE) of mean disease severity estimates and sample size (n) (from n = 10 to 100 in n = 5 increments) for each of the different scales used, over a range of “actual” mean disease severities (1%, 5%, 10%, 20%, 30%, and 40%) and variance (represented by large, medium, and small variation). The simulation process was repeated 10,000 times. Assessment methods: (i) NPE (nearest percentage estimate); (ii) AM10 (10% linear scale emphasizing severities ≤50% disease, and with additional intervals at severities <10%); (iii) AM20: similar to the structure of AM10 (both scales have the same structure at ≤10%), but with 20% intervals for disease severities from 10% to 50%; (iv) AM5: similar to the structure of AM10 (both scales have the same structure at ≤10%), but with 5% intervals for disease severities 10 to 50%; and (v) AM10f: similar to the structure of the AM10 scale, but with finer intervals at ≤10% disease.
Figure 5. The relationships between the mean squared error (MSE) of mean disease severity estimates and sample size (n) (from n = 10 to 100 in n = 5 increments) for each of the different scales used, over a range of “actual” mean disease severities (1%, 5%, 10%, 20%, 30%, and 40%) and variance (represented by large, medium, and small variation). The simulation process was repeated 10,000 times. Assessment methods: (i) NPE (nearest percentage estimate); (ii) AM10 (10% linear scale emphasizing severities ≤50% disease, and with additional intervals at severities <10%); (iii) AM20: similar to the structure of AM10 (both scales have the same structure at ≤10%), but with 20% intervals for disease severities from 10% to 50%; (iv) AM5: similar to the structure of AM10 (both scales have the same structure at ≤10%), but with 5% intervals for disease severities 10 to 50%; and (v) AM10f: similar to the structure of the AM10 scale, but with finer intervals at ≤10% disease.
Agronomy 09 00565 g005
Figure 6. The relationships between the variance of mean disease severity estimates and sample size (n) (from n = 10 to 100 in n = 5 increments) for each of the different scales used, over a range of “actual” mean disease severities (1%, 5%, 10%, 20%, 30%, and 40%) and variance (represented by large, medium, and small variation). The simulation process was repeated 10,000 times. Assessment methods were: (i) NPE (nearest percentage estimate); (ii) AM10 (10% linear scale emphasizing severities ≤50% disease, and with additional intervals at severities <10%); (iii) AM20: Similar to the structure of AM10 (both scales have the same structure at ≤10%), this scale has wider ordinal scale intervals for disease severities 10 through 50% based on a 20% linear scale; (iv) AM5: Similar to the structure of AM10 (both scales have the same structure at ≤10%), this scale has narrower ordinal scale intervals for disease severities 10 through 50% based on a 5% linear scale; (v) AM10f: Similar to the structure of AM10, this scale has 10% linear intervals at >10%, but has finer intervals at ≤10% disease.
Figure 6. The relationships between the variance of mean disease severity estimates and sample size (n) (from n = 10 to 100 in n = 5 increments) for each of the different scales used, over a range of “actual” mean disease severities (1%, 5%, 10%, 20%, 30%, and 40%) and variance (represented by large, medium, and small variation). The simulation process was repeated 10,000 times. Assessment methods were: (i) NPE (nearest percentage estimate); (ii) AM10 (10% linear scale emphasizing severities ≤50% disease, and with additional intervals at severities <10%); (iii) AM20: Similar to the structure of AM10 (both scales have the same structure at ≤10%), this scale has wider ordinal scale intervals for disease severities 10 through 50% based on a 20% linear scale; (iv) AM5: Similar to the structure of AM10 (both scales have the same structure at ≤10%), this scale has narrower ordinal scale intervals for disease severities 10 through 50% based on a 5% linear scale; (v) AM10f: Similar to the structure of AM10, this scale has 10% linear intervals at >10%, but has finer intervals at ≤10% disease.
Agronomy 09 00565 g006
Figure 7. The relationships between the bias of mean disease severity estimates and sample size (n) (from n = 10 to 100 in n = 5 increments) for each of the different scales used, with conditioning on a range of “actual” mean disease severities (1%, 5%, 10%, 20%, 30%, and 40%) and variance (represented by large, medium, and small variation). The simulation process was repeated 10,000 times. Assessment methods were: (i) NPE (nearest percentage estimate); (ii) AM10 (10% linear scale emphasizing severities ≤50% disease, and with additional intervals at severities <10%); (iii) AM20: Similar to the structure of AM10 (both scales have the same structure at ≤10%), this scale has wider ordinal scale intervals for disease severities 10 through 50% based on a 20% linear scale; (iv) AM5: Similar to the structure of AM10 (both scales have the same structure at ≤10%), this scale has narrower ordinal scale intervals for disease severities 10 through 50% based on a 5% linear scale; (v) AM10f: Similar to the structure of AM10, this scale has 10% linear intervals at >10%, but has finer intervals at ≤10% disease.
Figure 7. The relationships between the bias of mean disease severity estimates and sample size (n) (from n = 10 to 100 in n = 5 increments) for each of the different scales used, with conditioning on a range of “actual” mean disease severities (1%, 5%, 10%, 20%, 30%, and 40%) and variance (represented by large, medium, and small variation). The simulation process was repeated 10,000 times. Assessment methods were: (i) NPE (nearest percentage estimate); (ii) AM10 (10% linear scale emphasizing severities ≤50% disease, and with additional intervals at severities <10%); (iii) AM20: Similar to the structure of AM10 (both scales have the same structure at ≤10%), this scale has wider ordinal scale intervals for disease severities 10 through 50% based on a 20% linear scale; (iv) AM5: Similar to the structure of AM10 (both scales have the same structure at ≤10%), this scale has narrower ordinal scale intervals for disease severities 10 through 50% based on a 5% linear scale; (v) AM10f: Similar to the structure of AM10, this scale has 10% linear intervals at >10%, but has finer intervals at ≤10% disease.
Agronomy 09 00565 g007
Figure 8. Results in relation to the actual mean severity of peach scab (caused by Venturia nashicola) as measured using image analysis on leaves from six plots in Dongshi District, Taichung County, Taiwan. Plots were sampled from September to November 2018. The results show the relationships between the mean squared error (MSE), or variance, or bias of the mean disease severity estimates and sample size (n) (from 10 to 100 in 5 increments) for each of the different scales used, at the six mean disease severities of 8%, 13%, 20%, 32%, 41%, and 49% with corresponding variances of 0.0091%, 0.0146%, 0.0146%, 0.0192%, 0.0287%, and 0.0231%, respectively. Assessment methods: (i) NPE (nearest percentage estimate); (ii) HB (Horsfall-Barratt scale); (iii) EI10 (equal interval 10% linear scale); (iv) AM10 (10% linear scale emphasizing severities ≤50% disease, and with additional intervals at severities <10%); (v) AM20: Similar to the structure of AM10 (both scales have the same structure at ≤10%), this scale has wider ordinal scale intervals for disease severities 10 through 50% based on a 20% linear scale; (vi) AM5: Similar to the structure of AM10 (both scales have the same structure at ≤10%), this scale has narrower ordinal scale intervals for disease severities 10 through 50% based on a 5% linear scale; (vii) AM10f: Similar to the structure of AM10, this scale has 10% linear intervals at >10%, but has finer intervals at ≤10% disease.
Figure 8. Results in relation to the actual mean severity of peach scab (caused by Venturia nashicola) as measured using image analysis on leaves from six plots in Dongshi District, Taichung County, Taiwan. Plots were sampled from September to November 2018. The results show the relationships between the mean squared error (MSE), or variance, or bias of the mean disease severity estimates and sample size (n) (from 10 to 100 in 5 increments) for each of the different scales used, at the six mean disease severities of 8%, 13%, 20%, 32%, 41%, and 49% with corresponding variances of 0.0091%, 0.0146%, 0.0146%, 0.0192%, 0.0287%, and 0.0231%, respectively. Assessment methods: (i) NPE (nearest percentage estimate); (ii) HB (Horsfall-Barratt scale); (iii) EI10 (equal interval 10% linear scale); (iv) AM10 (10% linear scale emphasizing severities ≤50% disease, and with additional intervals at severities <10%); (v) AM20: Similar to the structure of AM10 (both scales have the same structure at ≤10%), this scale has wider ordinal scale intervals for disease severities 10 through 50% based on a 20% linear scale; (vi) AM5: Similar to the structure of AM10 (both scales have the same structure at ≤10%), this scale has narrower ordinal scale intervals for disease severities 10 through 50% based on a 5% linear scale; (vii) AM10f: Similar to the structure of AM10, this scale has 10% linear intervals at >10%, but has finer intervals at ≤10% disease.
Agronomy 09 00565 g008
Table 1. The mean square error (MSE) for the seven assessment methods summed from n = 10 to 100 in n = 5 increments for each specific mean severity and variance (large, medium, and small) for the simulated data sets and the estimated pear scab severity data. The sums of the MSEs for all the mean disease severities were calculated (“Overall”), and a ranking (1 to 7) assigned for each assessment method, where 1 = lowest MSE.
Table 1. The mean square error (MSE) for the seven assessment methods summed from n = 10 to 100 in n = 5 increments for each specific mean severity and variance (large, medium, and small) for the simulated data sets and the estimated pear scab severity data. The sums of the MSEs for all the mean disease severities were calculated (“Overall”), and a ranking (1 to 7) assigned for each assessment method, where 1 = lowest MSE.
Data SetAssessment Method 1Mean Disease SeveritiesOverallRank
0.010.050.100.200.300.40
Simulation study (large variance)NPE8.54 22.55 78.35 239.49 437.87 451.96 1238.75 1
HB42.39 35.98 100.32 250.65 429.65 438.37 1297.36 6
EI10388.52 105.69 111.06 241.90 427.15 442.51 1716.84 7
AM1012.78 26.60 83.61 255.08 434.88 445.73 1258.68 4
AM2015.15 43.92 101.50 261.36 432.26 439.79 1293.98 5
AM512.19 23.62 81.04 254.71 435.83 447.02 1254.41 2
AM10f12.38 25.47 83.28 255.18 435.43 446.17 1257.90 3
Simulation study (medium variance)NPE4.71 7.98 29.17 98.64 181.68 262.46 584.63 1
HB32.69 13.50 51.56 139.14 201.28 259.75 697.92 5
EI10354.06 30.54 31.62 100.54 184.71 265.91 967.38 7
AM108.61 14.12 34.26 105.27 202.16 296.57 661.00 4
AM2011.40 34.03 86.93 123.69 200.86 290.04 746.96 6
AM57.91 10.08 29.40 104.11 202.57 297.94 652.01 2
AM10f7.91 11.77 34.21 105.50 202.25 296.62 658.27 3
Simulation study (small variance)NPE1.12 4.08 5.18 10.96 21.54 147.22 190.10 1
HB13.31 7.73 27.22 40.17 236.39 157.86 482.68 6
EI10306.01 9.16 15.50 15.94 25.77 151.53 523.92 7
AM103.65 10.66 18.19 15.74 26.04 177.88 252.16 3
AM203.84 21.35 189.88 7.44 54.30 170.53 447.35 5
AM53.57 7.48 6.59 12.05 22.91 180.01 232.61 2
AM10f2.19 7.09 26.22 15.65 26.04 177.82 255.02 4
Mean disease severities
0.080.130.200.320.410.49
Based on estimates of pear scabNPE48.74 77.60 77.90 103.34 151.03 122.66 581.27 1
HB91.44 146.95 199.68 115.86 132.38 122.26 808.58 6
EI10111.89 89.24 70.20 116.81 160.47 129.65 678.25 2
AM1059.36 77.02 74.96 127.37 206.95 211.12 756.80 5
AM20106.09 117.79 110.30 133.63 199.47 189.06 856.32 7
AM549.55 76.98 77.99 123.25 208.75 201.55 738.07 3
AM10f57.02 77.07 74.62 127.37 206.95 211.12 754.16 4
1 Assessment methods: NPE (nearest percentage estimate); HB (Horsfall-Barratt scale); EI10 (equal interval 10% linear scale); AM10 (10% linear scale emphasizing severities ≤50% disease, and with additional intervals at severities <10%); AM20: Similar to the structure of AM10 (both scales have the same structure at ≤10%), this scale has wider ordinal scale intervals for disease severities 10 through 50% based on a 20% linear scale; AM5: Similar to the structure of AM10 (both scales have the same structure at ≤10%), this scale has narrower ordinal scale intervals for disease severities 10 through 50% based on a 5% linear scale; AM10f: Similar to the structure of AM10, this scale has 10% linear intervals at >10%, but has finer intervals at ≤10% disease.

Share and Cite

MDPI and ACS Style

Liu, H.I.; Tsai, J.R.; Chung, W.H.; Bock, C.H.; Chiang, K.S. Effects of Quantitative Ordinal Scale Design on the Accuracy of Estimates of Mean Disease Severity. Agronomy 2019, 9, 565. https://doi.org/10.3390/agronomy9090565

AMA Style

Liu HI, Tsai JR, Chung WH, Bock CH, Chiang KS. Effects of Quantitative Ordinal Scale Design on the Accuracy of Estimates of Mean Disease Severity. Agronomy. 2019; 9(9):565. https://doi.org/10.3390/agronomy9090565

Chicago/Turabian Style

Liu, Hung I., Jia Ren Tsai, Wen Hsin Chung, Clive H. Bock, and Kuo Szu Chiang. 2019. "Effects of Quantitative Ordinal Scale Design on the Accuracy of Estimates of Mean Disease Severity" Agronomy 9, no. 9: 565. https://doi.org/10.3390/agronomy9090565

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop