Skip to Content
SportsSports
  • Review
  • Open Access

7 May 2019

Recommendations for Measurement and Management of an Elite Athlete

,
,
,
,
,
,
and
1
High Performance Department, USA Ski and Snowboard Association, Victory Ln, Park City, UT 84060, USA
2
Head Sports Physiology and Research, Aspire Academy, Doha POB 22287, Qatar
3
Department of Physical Education Health and Recreation, Eastern Washington University, Cheney, WA 99004, USA
4
Physical Education Program, University of California, Berkeley, CA 94720, USA

Abstract

Athletes who merit the title ‘elite’ are rare and differ both quantitatively and qualitatively from athletes of lower qualifications. Serving and studying elite athletes may demand non-traditional approaches. Research involving elite athletes suffers because of the typical nomothetic requirements for large sample sizes and other statistical assumptions that do not apply to this population. Ideographic research uses single-athlete study designs, trend analyses, and statistical process control. Single-athlete designs seek to measure differences in repeated measurements under prescribed conditions, and trend analyses may permit systematic monitoring and prediction of future outcomes. Statistical process control uses control charting and other methods from management systems to assess and modify training processes in near real-time. These methods bring assessment and process control into the real world of elite athletics.

1. Introduction

Studies of elite athletes are not rare, too often because the label “elite athlete” is applied too generously. Since 1974, the PUBMED.GOV online database holds records of a total of 6661 studies averaging 155 studies per year through 2018 discoverable with the search term “elite athlete.” The majority of studies of “elite athletes” have occurred since the year 2000 with a total of 5934, averaging approximately 312 studies per year (PUBMED.GOV accessed 6 April 2019). Measurement and management of athletes, elite or otherwise, is not a simple problem of categorization but one of acquiring reliable and valid data that can benefit all the stakeholders.
Traditional approaches to studies of sport and exercise science are built upon assumptions that rarely hold for the elite athlete. We believe that reframing elite athlete research and management will serve all of the stakeholders that pursue the highest levels of performance. The methods described below may also benefit any investigator, coach, or scientist who is faced with the problem of ensuring that training and performance are optimized for any single athlete.

1.1. What Is an Elite Athlete?

An elite athlete is someone who trains and competes at the highest levels of his or her chosen sport. Elite athletes are rare by definition. An elite athlete differs both quantitatively and qualitatively from athletes of middle- and lower-level qualifications. Although one can argue that elite athletes continue to learn throughout their careers, the elite athlete is not considered developmental nor competes in a developmental path or system. The elite athlete’s training and competition capacities are relatively stable, and his or her skills have reached an exceptionally high performance level [1]. Previously, the United States Olympic Committee defined an elite athlete as an athlete ranked among the top eight in the world [2]. Lorenz and colleagues [3] described an elite athlete as someone who is:
  • “drafted or drafted in high rounds versus those undrafted or drafted in later rounds;
  • perceived as having greater performance ability than that of their peers in the same sport;
  • play at a higher level within a sport (division I vs II, professional vs amateur); and
  • for endurance, greater variables (e.g., running economy, AT, VO2max).” [3], p 542.
Elite athletes have backgrounds of great training volume, such as 7500 h in biathlon; 1.5 to 2 million ball strikes in golf; 150,000 to 600,000 arrows in archery; 2000 runs in bobsled, skeleton, and luge; and 250,000 to 500,000 elements per year in women’s gymnastics [4,5]. We will define an elite athlete as an athlete on a national team.
Too often, athletes described as elite are merely trained, highly trained, or included with a group of lower-level athletes. Studying elite athletes is difficult, and “estimates of enhancement of performance in laboratory or field tests in most previous studies may not apply to elite athletes in competitive events.” [6], p 472. In spite of recent controversies involving the role of practice in sports expertise [7,8,9,10,11,12]; the importance of many long hours of practice is still considered axiomatic to becoming an elite athlete. Lower-level and less-experienced athletes are comparatively deficient in training volume, experience, level of coaching, and talent [4,13].

1.2. The Covenant

Coaches and sport scientists working with elite athletes have a similar covenant to that of a physician and patient. While physicians are certainly directed by the results of experimental scientific research, documented trends, and other evidence-based-medicine; they must cure or heal the patient in front of them [14]. “Evidence based medicine (EBM) is the conscientious, explicit, judicious and reasonable use of modern, best-evidence in making decisions about the care of individual patients.” [15] p 219 (emphasis added). Physicians do not blindly apply a known method to a symptom profile and then consider their job complete. Physicians have to solve the individual patient’s problem, not some average patient. The sport scientist faces a similar covenant. The sport scientist must improve, stabilize, or repair the performance of the athlete he or she has, not the average athlete. The foregoing points lead to the use of individual athlete measurement and management (i.e., testing and analyzing a single athlete).

1.3. Research on Elite Athletes’ Demands for Individualization

In order to determine if a meaningful change has occurred in some aspect of training and performance, the investigator needs to know the background variation of the variable(s) in question [16,17,18]. In other words, to know if something has moved to a new place one must know where it began, its direction of change, how much variation came from the athlete (biological variability), measurement methods, environment, and a variety of other factors [18,19,20].
Single-subject research offers a viable, if not preferable, paradigm for analyses of the elite athlete. If the reader has never used or heard of single-subject research, what we will refer to as single athlete assessment (SAA), and is only familiar with the methods of R.A. Fisher, Karl Pearson, and other statistical pioneers [21], then the idea of single-athlete research may cause cognitive dissonance [22,23]. However, one should appreciate that there are a number of noted scientific results that were based on an ‘N of 1.’ For example, Pavlov instrumented one dog [24], Broca dissected one brain [24], B.F. Skinner studied operant conditioning with only one athlete (i.e., animal) at a time [25], and Ebbinghaus studied memory using only himself as the subject [26]. Research methods involving a single athlete would seem to be tailored perfectly for the understanding and evidence-based support of elite athlete preparation and performance.
An investigative question and a means of measurement in SAA approaches are usually determined before an experiment. SAAs are strong approaches to exploratory investigations often resulting in hypothesis generation rather than confirmation in the traditional research design sense [27,28]. One’s ability to be confident about measurements and what data might reveal are based on the use of multiple relevant measurements and similar results from different lines or types of inquiry. If multiple measurements of anything vary unreasonably, then one should be skeptical that whatever produced the measurements was inconsistent and may have arisen from multiple separate and independent or external causes rather than the treatment [29]. In group-based research, multiple measurements are conducted by virtue of doing the same measurements on many athletes. The variability of the measurements and their central tendency are characterized and then compared with the same measurements performed on a different group, the same group under different conditions, or both. Multiple measurements are performed with SAAs as well, but the measurements are performed on the same athlete. One could use SAA methods and simply accumulate a number of individual athletes and combine the data for typical nomothetic analyses, however such an approach may be more labor intensive and threaten internal validity. Concerns about the stability and repeatability of the measurements in both group and SAA investigations are shared [24,30].
Training monitoring approaches may share many common features with SAA investigations [31]. Monitoring training has become a field of sport science unto itself. Training diaries, whether by paper or computer, store long-term repeated measurements of an athlete. The dose–response information from training monitoring can be examined and can provide a historical context of the athlete’s development. Long-term collection of training data can serve as the raw material (i.e., archived data) for SAA methods. Both approaches intersect when measuring and managing the elite athlete. Monitoring and SAAs share repeated measurements and many aspects of analysis. However, monitoring may include group measurements and analyses, such as in team sports, and usually relies on collecting large datasets assuming that there is something worth measuring in the data. Data mining and “big data” methods are often employed in training and performance monitoring approaches [32,33]. SAAs share the same desires, but testable hypotheses are involved before data analyses.

1.4. Case Study vs SAA

Common in the medical field, case studies and SAA are not the same but share some common features [24]. Case studies are descriptive in nature with no repeated measures of a baseline period (BP) (control or non-treatment condition) or a treatment phase or condition (TP) [34]. Clinical case studies make little attempt to reduce and/or eliminate rival hypotheses that may have influenced the outcome [34]. Case studies are a type of one-and-done approach to investigations, but can still be a useful research method for elite athletes, in particular when reporting training loads and/or specific training interventions. There is a paucity of real examples of the preparation of elite athletes in the literature, often for reasons of proprietary and competitive advantage. Documentation of what was used for training and how the elite athlete responded may still be helpful information for those who deal with elite athletes.
SAA methods are valuable in naturalistic investigations such as those associated with case studies and quasi-experiments involving observation and analyses of real-time and archived data (i.e., athlete monitoring). Quasi-experiments only approximate the demands of true experiments [35], and often have some aspect of the study that is out of the control of the investigator [36]. For example, during the baseline phase or the treatment phase of an athlete assessment, the athlete may be called away for personal, school, or family reasons. Either or both phases may be disrupted. This interruption may be of sufficient duration to compromise the integrity of either phase. One who is familiar with the practical circumstances of elite athlete training and performance will probably agree that nearly all of the potential investigations of elite athletes are or will be quasi-experiments. Kazdin describes the need for quasi-experiments as the middle ground lying between case studies and the clean, crisp elegance of controlled laboratory experiments [36]. Quasi-experiments, while not incorporating rigid laboratory controls, can provide reasonable evidence which may prove useful for good decisions in a messy world.
Given that the scope of SAA and training management is limited to one athlete, the investigator and training manager should be especially sensitive to the poor generalizability of SAA approaches to other athletes and settings [37]. Many of us trained by the methods of Fisher and Pearson have trouble embracing the idea that SAAs do not pretend to be generalizable to other circumstances. The results of studies with single athletes should not be amplified, upgraded, or over-interpreted for use with groups of athletes [30,38]. Those involved with the actual preparation and performance of elite athletes are rarely concerned with other athletes and situations as described below.

1.5. Measurement for Management of the Elite Athlete–Why Nomothetic Approaches Can Fail

Nomothetic research designs (i.e., group-based analyses) are the most common method for assessing athletes of all levels by sports scientists and others. However, there is a problem in the application of typical scientific measurement and analysis practices with elite athletes. The rareness of elite athletes, unfortunately, makes finding groups of sufficient number that can be analyzed without violating numerous Fisherian statistical assumptions difficult or impossible [39,40]. A brief fictitious example may help emphasize the situation.
Ten female national team athletes, involved in a strength and power sport, participated in a study investigating the effects of a new strength and power program for enhancing lower extremity explosiveness. With the chosen measure of change being a countermovement vertical jump (CMJ), these athletes, acting as their own control, performed a pretest CMJ for height. After completing the strength and power program for eight weeks, the athletes were retested, with the pre-test and post-test data shown in Table 1.
Table 1. Comparison of pre- and post-test data for a countermovement vertical jump.
The results of a paired t-test were t(9) = −2.29, p = 0.048. It appears that the strength program worked. Everyone should be happy. The average improvement was about 8%. However, note that the athletes with the highest pretest jumps (athletes 2, 5, and 8) actually got worse. Moreover, the coach knows that these three athletes are the only current hopes for podium finishes. Should the coach consider the intervention a success? Experience with dozens of presentations of this scenario to coaches has shown that not a single coach has ever considered this scenario a success. The three athletes who jumped poorly may have been the victims of other extenuating circumstances that accounted for their poor performance, but this further amplifies the point that without knowing the context of each athlete individually, using means and standard deviations may obscure information that is pertinent to the interpretation of this test and similar tests and treatments.

1.6. Differences between the Baseline Phase and Treatment Phase

Repeated measurements are the foundation of SAAs. Use of a repeated series of BP measurements establishes an untreated control condition, followed by the same measurements imposed during and after the application of a TP condition. Thus, the athlete serves as his or her own control. SAA designs are often described as A-B designs, where A is the baseline phase, and B is the treatment phase. There are a variety of design approaches from A-B to A-B-A (baseline to treatment and then the treatment is removed for another baseline) to A-B-A-B (BP to TP to BP to TP), and so forth. If the treatment (B) measurements are sufficiently different from the baseline (A) measurements, then the treatment may have been responsible for the change. The protocols for SAAs may present data patterns such as those shown in Figure 1.
Figure 1. Single national team athlete analysis of scaled peak force (SPF) (N/NBW [Newtons per Newtons Body Weight]) change. The solid line indicates the SPF, the dashed lines are the means of the two phases (baseline phase, BP, and treatment phase, TP), and the dotted line is the value of −1.5 times the standard deviation of the BP below the mean of the BP.

1.7. Importance of the Baseline Phase

In order to detect a change, there must be an aspect of performance to which one can compare the new performance. A typical group study design assigns a treatment to one group of athletes but not the other group, and the comparison is performed between the two groups of athletes. In another group setting, a single group of athletes performs both the pre-test and the post-test using the pre-test scores as the baseline. Using the same athletes in each setting, pre-test, treatment, and post-test is known as a within-subjects (i.e., within-athletes) design. Within-athletes designs can be powerful because normal human variability is thought to be carried from the pre-test through the treatment and post-test, thus allowing the individual’s variability to be ‘subtracted out’ leaving only the influence of the treatment.
The importance of the BP in SAAs cannot be overstated. The BP serves as a descriptor of the current level of performance of the athlete. Baselines are the foundation from which predictions or comparisons are made, and therefore need to be stable [24,36]. The BP data should have relatively low variability, be without discernible trend, and be relatively free of error [36]. However, the requirements of baseline behavior are relative to the variable under assessment, the normal variability of the athlete, the timing of the baseline relative to treatment, and other factors [41,42].
Although having the BP prior to the TP is desirable, this is often not imperative. In some situations, the treatment may be the first phase needed immediately, leaving no time for observation to establish a baseline [36]. In other situations, the baseline behavior or performance may be obvious because the quality or magnitude of the treatment effects have never occurred before. In these circumstances, the design begins with the treatment or intervention in a B-A design. First, the treatment is applied (B), then the treatment is removed (i.e., shift to baseline, A). This approach may be most practical if the treatment is needed immediately to prevent injury or a serious learning or performance problem while still pursuing evidence-based assessment [36]. An example of this approach could be the first-time use of eyeshades or noise-abatement headphones on long plane flights with the outcomes being time and quality of sleep.

2. Analysis of SAAs

Two primary approaches are used to analyze SAA data [36]. The first is visual inspection [24,36,41], with the second being statistical or quantitative procedures [30,41,43,44,45,46,47]. Alternatively, one can consider these two methods as the order in which SAA data are analyzed—first visually and then quantitatively. Visual inspection suffers from various aspects of bias [48] while offering rapid interpretation and classification of the obvious [49]. Pragmatically, coaches and athletes rarely take the necessary steps for statistical verification when making a choice about a training intervention or modality. However, everyone involved can benefit from evidence that rises above opinion, educated guess, rumor, and folklore. When an investigator pursues greater levels of confidence in a training decision, invoking appropriate statistical procedures can provide a higher degree of confidence [36]. Moreover, the methods of SAA analyses are simple and understandable to any coach.
Referring to Figure 1, the SAA data presents a decline in scaled peak force (SPF) of a men’s national gymnastics team athlete who showed a decline in SPF during weekly drop jump tests. The data recorded displays the means of two drop jump trials collected from 5 May to 18 September, with BP and TP collected from 5 May to 7 July, and from 17 July to 18 September, respectively. In early July (summer break), the coach abandoned specific conditioning for explosive jumping in favor of a more relaxed training approach that focused on skill and technique training. This change is believed to have been the catalyst for this particular athlete’s decrease in his fitness level for explosive jumping. The research question was: Should we regard the decline in explosive ability as large enough to have exceeded a chance occurrence and thereby lend some credence to the idea that the shift in conditioning could be responsible? We can reasonably assume that if the decline exceeds chance variation, then the coach has made a mistake that needs remediation.
Visual Inspection. Analysis of SAAs involves principles and methods similar to those found in nomothetic studies. Athlete training and management may benefit from simply graphing the data, with visual inspection being sufficient for determining whether a worthwhile effect or difference occurred [24]. The data show an obvious decline in SPF during the latter half of the period of interest. However, was the decline large enough to be unlikely to have occurred by chance? Sadly, what may seem obvious could fool the investigator [50].
Quantitative Analysis. If the data are used for research purposes and/or there is a desire for increased confidence that the results are meaningful, then more rigorous analyses are merited. For example, a difference between BP and TP means of 1.5–2.0 times the BP standard deviation or the pooled standard deviation is considered indicative of a meaningful change based on typical error [51].
Following the approach described by Hopkins [52], the two periods shown in Figure 1 resulted in the following: The BP mean of SPF ± SD was 9.30 ± 0.62 N/NBW, CoV 6.7%, with the mean SPF value during the treatment phase being 8.20 ± 0.62 N/NBW, CoV = 7.6%. The difference between the BP and TP means was 1.14 N/NBW, exceeding −1.5 times the standard deviation of the BP and the pooled standard deviation 0.62 N/NBW. A value as large as or larger than −1.5 times the BP standard deviation or pooled standard deviation would only occur with a probability of approximately 7% based on normal probabilities. Since there are eight measurements in each phase, a paired t-test showed a statistically significant difference between the phases (t = 3.59(7), p = 0.009, 95% CI = 0.39 to 1.89). In this case, the paired t-test was a procedure of convenience used to confirm further that there could be a worthwhile difference. The data appear to reveal an important change in the lower extremity fitness of this athlete and the coach should be informed of a potential need for a change in the conditioning program.
Before leaving the SAA methods, there are a litany of other approaches that can be used to assess phase differences in single-athlete data. Additional methods for the analysis of SAA phases include:
  • percentage non-overlap of the data points in the BP compared to the TP (80% non-overlap is desired) [53],
  • counting points above and below a celeration line can be used to detect phase differences [24,36], and
  • confidence intervals and effect sizes continue to be valuable and important [54,55,56], and others.

4. Statistical Process Control (SPC)

Most individuals in athletics would agree that training is a process. Until recently, training management was largely a ‘seat-of-the pants’ endeavor, relying almost exclusively on the experience, knowledge, and personal philosophies of a coach or team of coaches. Statistical process control (SPC) uses continuous data sampling of a process to expose the patterns evident in the process. Originally, the idea was to measure how industrial and business processes of production could be measured and managed. SPC is a subset of Total Quality Management (TQM), which uses SPC and other management methods to measure and control nearly all aspects of a production process. SPC was implemented by W.E. Deming to assist Japan in industrial rebuilding after World War II [81,82]. SPC can also be used for research purposes [31,83,84], but its greatest strength lies in monitoring change and detecting outliers. SPC cannot identify the nature of the problem, only that something unusual and worthy of investigation has occurred or is occurring. Training is a process that could profit from many of the tenets of SPC. Continuous sampling of a single athlete’s training is synonymous with training monitoring.
Perhaps the most powerful technique of SPC that will find immediate use with elite athletes is the control chart [83,84,85,86]. Control charts are used to measure and monitor a process as it occurs. The primary information gained from control charts is the identification of trends, shifts, and outlier data. Outlier data in SPC refers to unusual data values that do not fit within the normal variability ranges of the process [58,59,87,88,89,90,91].
Control charts are the primary means of displaying the athlete’s training process over time. For example, if PPRH is collected systematically each day, one should expect some variability in the values obtained (Figure 5). Some variation is ‘normal’ (i.e., expected) and demands no explanation. Other variation can be considered abnormal or rare and because of the rarity merits further investigation. Athletes vary in their status and condition during all training durations, minute-to-minute, day-to-day, month-to-month, and year-to-year. By understanding an athlete’s performance and training variations, one gains the ability to determine when variation or change is too large to be expected by chance through normal variability. For example, if one plots the PPRH over time in conjunction with the mean and multiples of the standard deviation, the distribution of the PPRH values will vary predictably relative to the normal distribution. Figure 5 shows PPRH data from a former elite female gymnast over an entire collegiate season. Figure 5 is a control chart with information embellishments to help show how SPC works.
Figure 5. Control chart of a former elite gymnast’s PPRH for an entire collegiate season with associated mean, multiples of standard deviations, and normal curves.
A variable that is reflective of a process should indicate whether the process is ‘in’ or ‘out’ of control (i.e., varying normally or excessively). The process of athlete training should be in control, and variables reflective of that control should vary within relatively small ranges. As described above, human performance varies in many ways, but when plotted such as in Figure 5, the variability should remain within certain limits, defined by probability, in order to be considered in control. These limits are called ‘control limits,’ and although these limits can be calculated in a number of ways [81,82,83,84], assumptions of statistical normality are often used. As such, the distribution of data collected over time can be plotted and inspected horizontally. Figure 5 shows the normal distribution of the plotted data with normal curves (dotted lines) graphed on top of the longitudinal data for illustrative purposes. Calculated control limits (1.5 times and 2.0 times the SD) are also shown. Since the normal distribution is well understood, we can use the properties of the normal distribution to study the variability of PPRH in terms of statistical regularity or rareness.
The athlete began training on 11 September and was immediately confronted with high volume training after a summer of little physical training. The early PPRH data met or exceeded the upper control limits four times, and most of the remaining data values were above the mean. These data may indicate that the athlete was facing unaccustomed training loads and that she is being loaded above her accustomed norm. The outlier data may or may not indicate something worth investigating. In this case, the athlete’s response should be expected to show higher than normal resting heart rates.
Figure 5 shows four PPRH values above +2.0 × SD and four below −2.0 × SD relative to the control limits. In terms of the +1.5 × SD control limit, the percentage of data values expected between the +1.5 × SD and −1.5 × SD control limits is approximately 93.3%. Thus, the probability of a data value greater or less than the +1.5 × SD control limit is approximately 6.7%, still indicative of a rare result meriting investigation. The ±1.5 × SD value has been chosen based on experience. Two standard deviations is often too high, making the detection of outlier values too restrictive [31]. Since our interest lies in excessively high PPHR values, we can note that four values met or exceeded the +2.0 × SD control limit, and 11 PPHR values exceeded the 1.5 × SD control limit (including those greater than 2.0 × SD).

Combining Data Histories with SPC

With illness considered a threat to safe and productive training, Figure 6 references the assessment with the use of SPC and two data series recorded by a monitoring tool, daily PPRH and the sum of illness symptoms (ILL) for a former elite gymnast [31,69,87,91]. Although there may not be a sufficient number of records to meet the criteria described for detecting differences and trends as discussed above, SPC identified data outliers (unusual or rare values) compared to normal variation. These results provided the coach, scientist, and physician with a systematic approach regarding the potential relationship between PPRH and illness, possibly enabling prediction of the onset of illness.
Figure 6. SPC of PPRH and the sum of illness symptoms (ILL) across a complete collegiate gymnastics season. Note that when illness symptoms are high there is a corresponding increase in PPRH.
Although the ILL does not always reach a control limit for the number of illness symptoms, this analysis of ILL is limited by the use of a variable that can only assume a few values based on the presence or absence of illness symptoms, while resting heart rate can assume many values. The elegant simplicity of this approach for both variables remains helpful to the training and performance stakeholders in identifying outliers and their relationships. In this presentation, the number of variables is relatively small (i.e., two) and could be enhanced by including other variables from monitoring information.
SPC can lend considerable interpretative structure and value. The ability to identify illness periods and their relative influence on overall adaptation may serve as a window on the processes of change that the athlete is experiencing. Such information may serve as an evidence-based aid for enhanced decision making.
SPC combines all the techniques and methods described above and puts them to use in a process-monitoring format that can provide important information with minimum delay. The detection of outliers is important as “red flags” for coaches, scientists, and physicians. After all, if something is behaving normally, there is little need for concern or intervention. The need for concern arises when something gets “weird.” SPC uses probabilities to assess when something happens that deserves attention.

5. Conclusions

Studying and managing elite athletes can be extremely challenging because of their rareness in the athletic population and the sometimes suboptimal approach of applying traditional statistics to identify differences, trends, and outlier behavior when there are only one or a few elite athletes. Training and performance are staggeringly complex; the use of the methods described above helps reduce the universe of possible interpretations to a manageable level. Of course, the methods described here are not exclusively for elite athletes and can serve as a tool for any investigator who is interested in one or a few important individuals, or has the ability to process large amounts of individual athlete data. SPC in particular is a relatively simple method, intimately tied to training monitoring, that can help identify a process threat (i.e., training or performance), by indicating whether the threat is worth considering. SPC methods can be applied for near real-time measurement and management, unlike the most traditional approaches of waiting for a treatment’s influence to be observable or a trend to become manifest. It is hoped that this work can stimulate more sports scientists to understand, measure, and manage elite athletes. Considering the fact that “elite” sport is now a big industry, the impact of such research efforts could be large.

Author Contributions

M.C. was the inspiration for this paper. Following the conference on athlete monitoring in Doha, Qatar, we sought to provide more in-depth information on the analysis of elite athletes. M.C. applies elite athlete assessment on a daily basis with several sports. J.M. provided considerable editing of the article and extensive comments on the ideas and methods described. She also critiqued and analyzed the statistical procedures discussed. J.M. applies elite athlete assessment regularly with U.S.A Diving. S.M. examined the content for consistency and provided invaluable editing of the text. C.S. regularly applies the concepts described here in his teaching, research, and use with university athletes. J.R. provided extensive rewrites of the manuscript along with adding his expertise in the application of the ideas discussed. N.A. edited, commented, and rewrote many sections of the manuscript. M.S. and W.S. have shared a consistent and long-term interest in these concepts for decades. Together at the U.S. Olympic Training Center we applied the concepts here to U.S. national 0team athletes.

Funding

No external funding was involved with this manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Janelle, C.M.; Hillman, C.H. Expert Performance in Sports; Human Kinetics: Champaign, IL, USA, 2003. [Google Scholar]
  2. Kearney, J.T. Sport performance enhancement: Design and analysis of research. Med. Sci. Sports Exerc. 1999, 31, 755–756. [Google Scholar] [CrossRef]
  3. Lorenz, D.S.; Reiman, M.P.; Lehecka, B.J.; Naylor, A. What Performance Characteristics Determine Elite Versus Nonelite Athletes in the Same Sport? Sports Health 2013, 5, 542–547. [Google Scholar] [CrossRef]
  4. Balyi, I.; Way, R.; Higgs, C. Long-Term Athlete Development; Human Kinetics: Champaign, IL, USA, 2013; p. 286. [Google Scholar]
  5. Caine, D.J.; Lindner, K.J.; Mandelbaum, B.R.; Sands, W.A. Gymnastics. In Epidemiology of Sports Injuries; Caine, D.J., Caine, C.G., Lindner, K.J., Eds.; Human Kinetics: Champaign, IL, USA, 1996; pp. 213–246. [Google Scholar]
  6. Hopkins, W.G.; Hawley, J.A.; Burke, L.M. Design and analysis of research on sport performance enhancement. Med. Sci. Sports Exerc. 1999, 31, 472–485. [Google Scholar] [CrossRef]
  7. Baker, J.; Horton, S.; Robertson-Wilson, J.; Wall, M. Nurturing sport expertise: Factors influencing the development of elite athlete. J. Sports Sci. Med. 2003, 2, 1–9. [Google Scholar]
  8. Coughlan, E.K.; Williams, A.M.; McRobert, A.P.; Ford, P.R. How experts practice: A novel test of deliberate practice theory. J. Exp. Psychol. Learn. Mem. Cogn. 2014, 40, 449–458. [Google Scholar] [CrossRef]
  9. Ericsson, K.A. The road to excellence: The acquisition of expert performance in the arts and sciences, sports, and games. In The Acquisition of Expert Performance: An Introduction to Some of the Issues; Ericsson, K.A., Ed.; Erlbaum: Mahwah, NJ, USA, 1996; pp. 1–50. [Google Scholar]
  10. Ericsson, K.A. Development of elite performance and deliberate practive: An update from the perspective of the expert performance approach. In Expert Performance in Sport; Starkes, J.L., Ericsson, K.A., Eds.; Human Kinetics: Champaign, IL, USA, 2003; pp. 49–84. [Google Scholar]
  11. Ericsson, K.A. Training history, deliberate practice and elite sports performance: An analysis in response to Tucker and Collins review--what makes champions? Br. J. Sports Med. 2013, 47, 533–535. [Google Scholar] [CrossRef]
  12. Tucker, R.; Collins, M. What makes champions? A review of the relative contribution of genes and training to sporting success. Br. J. Sports Med. 2012, 46, 555–561. [Google Scholar] [CrossRef]
  13. Gulbin, J.; Weissensteiner, J.; Oldenziel, K.; Francoys Gagne, E. Patterns of performance development in elite athletes. Eur. J. Sport Sci. 2013, 13, 605–614. [Google Scholar] [CrossRef]
  14. Whitley, R.; Rousseau, C.; Carpenter-Song, E.; Kirmayer, L.J. Evidence-based medicine: Opportunities and challenges in a diverse society. Can. J. Psychiatry 2011, 56, 514–522. [Google Scholar] [CrossRef]
  15. Masic, I.; Miokovic, M.; Muhamedagic, B. Evidence Based Medicine—New Approaches and Challenges. Acta Inform. Med. 2008, 16, 219–225. [Google Scholar] [CrossRef]
  16. Bagger, M.; Petersen, P.H.; Pedersen, P.K. Biological variation in variables associated with exercise training. Int. J. Sports Med. 2003, 24, 433–440. [Google Scholar] [CrossRef]
  17. Katch, V.L.; Sady, S.S.; Freedson, P. Biological variability in maximum aerobic power. Med. Sci. Sports Exerc. 1982, 14, 21–25. [Google Scholar] [CrossRef]
  18. Macfarlane, D. Who do exercise physiologists test best - athletes or themselves. N. Z. J. Sports Med. 1991, 19, 13–15. [Google Scholar]
  19. Ojanen, M. Can the true effects of exercise on psychological variables be separated from placebo effects? Int. J. Sport Psychol. 1994, 25, 63–80. [Google Scholar]
  20. Olds, T. Measurement issues in exercise science. J. Sci. Med. Sport 2002, 5, 336–340. [Google Scholar] [CrossRef]
  21. Salsburg, D. The Lady Tasting Tea; Henry Holt and Company: New York, NY, USA, 2001; p. 340. [Google Scholar]
  22. Huberty, C.J. A History of Effect Size Indices. Educ. Psychol. Meas. 2002, 62, 227–240. [Google Scholar] [CrossRef]
  23. Porter, T.M. (Ed.) The Rise of Statistical Thinking 1820–1900; Princeton University Press: Princeton, NJ, USA, 1986. [Google Scholar]
  24. Barlow, D.H.; Hersen, M. Single Case Experimental Designs: Strategies for Studying Behavior Change; Pergamon Press: New York, NY, USA, 1984. [Google Scholar]
  25. Skinner, B.F. The Behavior of Organisms: An Experimental Analysis; B.F. Skinner Foundation: Cambridge, MA, USA, 1938. [Google Scholar]
  26. Ebbinghaus, H. Memory: A Contribution to Experimental Psychology; Dover: New York, NY, USA, 1885. [Google Scholar]
  27. Biesecker, L.G. Hypothesis-generating research and predictive medicine. Genome Res. 2013, 23, 1051–1053. [Google Scholar] [CrossRef]
  28. Huberty, C.J.; Morris, J.D. Multivariate analysis versus multiple univariate analyses. Psychol. Bull. 1989, 105, 302–308. [Google Scholar] [CrossRef]
  29. Susser, M. What is a cause and how do we know one? A grammar for pragmatic epidemiology. Am. J. Epidemiol. 1991, 133, 635–648. [Google Scholar]
  30. Bates, B.T.; James, C.R.; Dufek, J.S. Single-subject analysis. In Innovative Analyses of Human Movement; Stergiou, N., Ed.; Human Kinetics: Champaign, IL, USA, 2004; pp. 3–28. [Google Scholar]
  31. Sands, W.A.; Kavanaugh, A.A.; Murray, S.R.; McNeal, J.R.; Jemni, M. Modern Techniques and Technologies Applied to Training and Performance Monitoring. Int. J. Sports Physiol. Perform. 2017, 12, S2-63–S62-72. [Google Scholar] [CrossRef]
  32. Siegel, E. Predictive Analytics; Wiley and Sons: Hoboken, NJ, USA, 2016. [Google Scholar]
  33. Stein, M.; Janetzko, H.; Seebacher, D.; Jäger, A.; Nagel, M.; Hölsch, J.; Kosub, S.; Schreck, T.; Keim, D.; Grossniklaus, M. How to make sense of team sport data: From acquisition to data modeling and research aspects. Data 2017, 2, 2. [Google Scholar] [CrossRef]
  34. Backman, C.L.; Harris, S.R. Case studies, single-subject research, and N of 1 randomized trials: Comparisons and contrasts. Am. J. Phys. Med. Rehabil. 1999, 78, 170–176. [Google Scholar] [CrossRef]
  35. Campbell, D.T.; Stanley, J.C. Experimental and Quasi-Experimental Designs for Research; Houghton Mifflin: Boston, MA, USA, 1963. [Google Scholar]
  36. Kazdin, A.E. Single-Case Research Designs; Oxford University Press: New York, NY, USA, 2011; p. 452. [Google Scholar]
  37. Reboussin, D.M.; Morgan, T.M. Statistical considerations in the use and analysis of single-subject designs. Med. Sci. Sports Exerc. 1996, 28, 639–644. [Google Scholar]
  38. Hrycaiko, D.; Martin, G.L. Applied research studies with single-subject designs: Why so few? J. Appl. Sport Psychol. 1996, 8, 183–199. [Google Scholar] [CrossRef]
  39. Goodman, S.N. Aligning statistical and scientific reasoning. Science 2016, 352, 1180–1181. [Google Scholar] [CrossRef]
  40. Loken, E.; Gelman, A. Measurement error and the replication crisis. Science 2017, 355, 584–585. [Google Scholar] [CrossRef]
  41. Kinugasa, T.; Cerin, E.; Hooper, S. Single-Subject Research Designs and Data Analyses for Assessing Elite Athletes’ Conditioning. Sports Med. 2004, 34, 1035–1050. [Google Scholar] [CrossRef]
  42. Matson, J.L.; Turygin, N.C.; Beighley, J.; Matson, M.L. Status of single-case research designs for evidence-based practice. Res. Autism Spectr. Disord. 2012, 6, 931–938. [Google Scholar] [CrossRef]
  43. Bates, B.T. Single-subject methodology: An alternative approach. Med. Sci. Sports Exerc. 1996, 28, 631–638. [Google Scholar]
  44. Bates, B.T.; Dufek, J.S.; Davis, H.P. The effect of trial size on statistical power. Med. Sci. Sports Exerc. 1992, 24, 1059–1068. [Google Scholar] [CrossRef]
  45. Cicciarella, C.F. Runs analysis—A tutorial with application to sport. Int. Sports J. 2000, 4, 107–118. [Google Scholar]
  46. Kazdin, A.E. Single-Case Research Designs: Methods for Clinical and Applied Settings; Oxford University Press: New York, NY, USA, 1982. [Google Scholar]
  47. Nourbakhsh, M.R.; Ottenbacher, K.J. The statistical analysis of single-subject data: A comparative examination. Phys. Ther. 1994, 74, 768–776. [Google Scholar] [CrossRef]
  48. Tufte, E.R. The Visual Display of Quantitative Information; Graphics Press: Cheshire, CT, USA, 1983. [Google Scholar]
  49. Harrington, M.A. Comparing Visual and Statistical Analysis in Single-Subject Studies; University of Rhode Island: Kingston, RI, USA, 2013. [Google Scholar]
  50. Brossart, D.F.; Parker, R.I.; Olson, E.A.; Mahadevan, L. The relationship between visual analysis and five statistical analyses in a simple AB single-case research design. Behav. Modif. 2006, 30, 531–563. [Google Scholar] [CrossRef] [PubMed]
  51. Hopkins, W.G. How to interpret changes in an athletic performance test. Sport Sci. 2004, 8, 1–7. [Google Scholar]
  52. Hopkins, W.G. Measures of reliability in sports medicine and science. Sports Med. 2000, 30, 1–15. [Google Scholar] [CrossRef] [PubMed]
  53. Huberty, C.J.; Lowman, L.L. Group Overlap as a Basis for Effect Size. Educ. Psychol. Meas. 2000, 60, 543–563. [Google Scholar] [CrossRef]
  54. Kirk, R.E. Effect magnitude: A different focus. J. Stat. Plan. Inference 2007, 137, 1634–1646. [Google Scholar] [CrossRef]
  55. Odgaard, E.C.; Fowler, R.L. Confidence Intervals for Effect Sizes: Compliance and Clinical Significance in the Journal of Consulting and Clinical Psychology. J. Consult. Clin. Psychol. 2010, 78, 287–297. [Google Scholar] [CrossRef]
  56. Parker, R.I.; Hagan-Burke, S. Useful Effect Size Interpretations for Single Case Research. Behav. Ther. 2007, 38, 95–105. [Google Scholar] [CrossRef]
  57. Sands, W.A.; McNeal, J.R. Predicting athlete preparation and performance: A theoretical perspective. J. Sport Behav. 2000, 23, 1–22. [Google Scholar]
  58. Sands, W.A.; Stone, M.H. Are you progressing and how would you know? Olympic Coach 2006, 17, 4–10. [Google Scholar]
  59. Sands, W.A.; Stone, M.H. Monitoring the elite athlete. Olympic Coach 2006, 17, 4–12. [Google Scholar]
  60. Jackson, A.S. Application of regression analysis to exercise science. In Measurement Concepts in Physical Education and Exercise Science; Safrit, M.J., Wood, T.M., Eds.; Human Kinetics: Champaign, IL, USA, 1989; pp. 181–206. [Google Scholar]
  61. Sands, W.A. National women’s tracking program pt. 2—Response. Technique 1990, 10, 23–27. [Google Scholar]
  62. Halson, S.L. Monitoring training load to understand fatigue in athletes. Sports Med. 2014, 44 (Suppl. S2), S139–S147. [Google Scholar] [CrossRef]
  63. Galambos, S.A.; Terry, P.C.; Moyle, G.M.; Locke, S.A. Psychological predictors of injury among elite athletes. Br. J. Sports Med. 2005, 39, 351–354. [Google Scholar] [CrossRef]
  64. McEwen, B.S. Stress, Adaptation, and Disease: Allostasis and Allostatic Load. Ann. N. Y. Acad. Sci. 1998, 840, 33–44. [Google Scholar] [CrossRef]
  65. Burke, E.R. Precision Heart Rate Training; Human Kinetics: Champaign, IL, USA, 1998. [Google Scholar]
  66. Douglas, P.S.; O’Toole, M.L. Cardiovascular and hematologic alterations. In Overtraining in Sport; Kreider, R.B., Fry, A.C., O’Toole, M.L., Eds.; Human Kinetics: Champaign, IL, USA, 1998; pp. 131–143. [Google Scholar]
  67. Nye, S. Monitoring workouts with heart rate. Swim. Tech. 1987, 24, 25–29. [Google Scholar]
  68. Yushkov, O.P.; Repnevsky, S.M.; Serdyuk, V.P. Use of heart rate for control over training loads. Sov. Sports Rev. 1986, 21, 151–152. [Google Scholar]
  69. Sands, W.A.; Irvin, R.C.; Major, J.A. Women’s gymnastics: The time course of fitness acquisition. A 1-year study. J. Strength Cond. Res. 1995, 9, 110–115. [Google Scholar] [CrossRef]
  70. Traub, R. Classical Test Theory in Historical Perspective. Educ. Meas. Issues Pract. 1997, 16, 1745–3992. [Google Scholar] [CrossRef]
  71. Dressendorfer, R.H.; Wade, C.E.; Scaff, J.H. Increased morning heart rate in runners: A valid sign of overtraining? Phys. Sportsmed. 1985, 13, 77–86. [Google Scholar] [CrossRef] [PubMed]
  72. Borresen, J.; Lambert, M.I. The quantification of training load, the training response and the effect on performance. Sports Med. 2009, 39, 779–795. [Google Scholar] [CrossRef] [PubMed]
  73. Armstrong, L.E.; Maresh, C.M.; Castellini, J.W.; Bergeron, M.F.; Kenefick, R.W.; LaGasse, K.E.; Riebe, D. Urinary indices of hydration status. Int. J. Sport Nutr. 1994, 4, 265–279. [Google Scholar] [CrossRef] [PubMed]
  74. Chatfield, C. The Analysis of Time Series; Chapman and Hall: London, UK, 1975. [Google Scholar]
  75. Crosbie, J. Interrupted time-series analysis with brief single-subject data. J. Consult. Clin. Psychol. 1993, 61, 966–974. [Google Scholar] [CrossRef] [PubMed]
  76. Lendaris, G.G.; Fraser, A.M. Visual fitting and extrapolation. In Time Series Prediction; Weigend, A.S., Gershenfeld, N.A., Eds.; Addison-Wesley: Reading, MA, USA, 1993; pp. 319–322. [Google Scholar]
  77. Bompa, T.O.; Haff, G.G. Periodization, 4th ed.; Human Kinetics: Champaign, IL, USA, 2009. [Google Scholar]
  78. Harre, D. The dynamics of the training load. In A collection of European Sports Science Translations Part II; Jarver, J., Ed.; South Australian Sports Institute: Kidman Park, Australia, 1990; pp. 39–41. [Google Scholar]
  79. Stone, M.H.; Stone, M.E.; Sands, W.A. Principles and Practice of Resistance Training; Human Kinetics: Champaign, IL, USA, 2007; p. 374. [Google Scholar]
  80. Sands, W.A.; McNeal, J.R.; Stone, M.H. Plaudits and pitfalls in studying elite athletes. Percept. Motor Skills 2005, 100, 22–24. [Google Scholar] [CrossRef]
  81. Deming, W.E. Quality, Productivity, and Competitive Position; Massachusettes Institute of Technology, Center for Advanced Engineering Study: Cambridge, MA, USA, 1982; Volume 183, p. 373. [Google Scholar]
  82. Shewhart, W.A. Statistical Method from the Viewpoint of Quality Control; Dover: New York, NY, USA, 1986. [Google Scholar]
  83. Grant, E.L.; Leavenworth, R.S. Statistical Quality Control; McGraw-Hill: New York, NY, USA, 1988. [Google Scholar]
  84. Keller, P. Statistical Process Control Demystified; McGraw-Hill: New York, NY, USA, 2011; p. 302. [Google Scholar]
  85. Kiemele, M.J.; Schmidt, S.R. Basic Statistics; Air Academy Press: Colorado Springs, CO, USA, 1993. [Google Scholar]
  86. Pitt, H. SPC for the Rest of Us; K.W. Tunnel Publishing Group: King of Prussia, PA, USA, 1994. [Google Scholar]
  87. Sands, W.A. Monitoring the elite female gymnast. Nat. Strength Cond. Assoc. J. 1991, 13, 66–71. [Google Scholar] [CrossRef]
  88. Sands, W.A. AI and athletics. PC AI 1992, 6, 52–54. [Google Scholar]
  89. Sands, W.A.; Henschen, K.P.; Shultz, B.B. National women’s tracking program. Technique 1989, 9, 14–19. [Google Scholar]
  90. Stone, M.H.; Stone, M.E.; Sands, W.A. Monitoring resistance training. In Principles and Practice of Resistance Training; Human Kinetics: Champaign, IL, USA, 2007; pp. 181–199. [Google Scholar]
  91. Sands, W.A.; Shultz, B.B.; Newman, A.P. Women’s gymnastics injuries. A 5-year study. Am. J. Sports Med. 1993, 21, 271–276. [Google Scholar] [CrossRef]

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.