Skip to main content
ARS Home » Southeast Area » Byron, Georgia » Fruit and Tree Nut Research » Research » Publications at this Location » Publication #327561

Title: Plant disease severity assessment - How rater bias, assessment method and experimental design affect hypothesis testing and resource use efficiency

Author
item CHIANG, KUO-SZU - National Chung-Hsing University
item Bock, Clive
item LEE, I-HSUAN - National Chung-Hsing University
item EL JARROUDI, MOUSSA - University Of Liege
item DELFOSSE, PHILIPPE - Luxembourg Institute Of Science & Technology

Submitted to: Phytopathology
Publication Type: Peer Reviewed Journal
Publication Acceptance Date: 7/13/2016
Publication Date: 12/1/2016
Citation: Chiang, K., Bock, C.H., Lee, I., El Jarroudi, M., Delfosse, P. 2016. Plant disease severity assessment - How rater bias, assessment method and experimental design affect hypothesis testing and resource use efficiency. Phytopathology. 106(12):1451-1464.

Interpretive Summary: Rater bias and assessment method can affect hypothesis testing. This was studied for different experimental designs for plant disease assessment using balanced and unbalanced data sets. Data sets with the same number of replicate estimates for each of two treatments are termed ‘balanced’, and those with unequal numbers of replicate estimates are termed ‘unbalanced’. The three assessment methods considered were nearest percent estimates (NPEs), an amended 10% incremental scale, and the Horsfall-Barratt (H-B) scale. Estimates of severity of Septoria leaf blotch on leaves of winter wheat were used to develop distributions for a simulation model. As expected, at a given fixed number of observations, the balanced experimental designs invariably resulted in a higher power compared with the unbalanced designs at different disease severity means, mean differences, and variances. Based on these results the recommended number of replicate estimates taken per specimen is two, as this conserves resource. Results indicated that, regardless of experimental design or rater bias, an amended 10% incremental scale has slightly less power compared with NPEs, and that the H-B scale is more likely than the others to cause a Type II error. These results suggest that choice of assessment method, optimizing sample number and number of replicate estimates and using a balanced experimental design are important criteria to consider so as to maximize the power of hypothesis tests for comparing treatments using disease severity estimates.

Technical Abstract: The impact of rater bias and assessment method on hypothesis testing was studied for different experimental designs for plant disease assessment using balanced and unbalanced data sets. Data sets with the same number of replicate estimates for each of two treatments are termed ‘balanced’, and those with unequal numbers of replicate estimates are termed ‘unbalanced’. The three assessment methods considered were nearest percent estimates (NPEs), an amended 10% incremental scale, and the Horsfall-Barratt (H-B) scale. Estimates of severity of Septoria leaf blotch on leaves of winter wheat were used to develop distributions for a simulation model. The experimental designs are presented here in the context of simulation experiments which consider the optimal design for the number of specimens (individual units sampled) and the number of replicate estimates per specimen for a fixed total number of observations (total sample size for the treatments being compared). The criterion used to gauge each method was the power of the hypothesis test. As expected, at a given fixed number of observations, the balanced experimental designs invariably resulted in a higher power compared with the unbalanced designs at different disease severity means, mean differences, and variances. Based on these results the recommended number of replicate estimates taken per specimen is two, as this conserves resource. Furthermore, for biased estimates, an apparent difference in the power of the hypothesis test was observed between different assessment methods and between different experimental designs. Results indicated that, regardless of experimental design or rater bias, an amended 10% incremental scale has slightly less power compared with NPEs, and that the H-B scale is more likely than the others to cause a Type II error. These results suggest that choice of assessment method, optimizing sample number and number of replicate estimates and using a balanced experimental design are important criteria to consider so as to maximize the power of hypothesis tests for comparing treatments using disease severity estimates.