Athlete Health and Human Performance Will Not Improve Without Transdisciplinary Collaboration and Data Sharing in Elite Sport
There are 2 largely competing models for an athletics organization at both the collegiate and professional levels: the high-performance model and the medical model. The high-performance model (Figure 1) largely arises from an international football perspective that places a “performance director” at the center of teams consisting of athletic training staff, physical therapists, medical director, and strength and conditioning, sports nutrition, and sport psychology staff, and in direct communication with the sport coaches and operations staff.1 In contrast, the medical model separates off medical staff (athletic trainers [ATs] and physicians, predominantly) and emphasizes the autonomy of medical decisions, and is designed to minimize potential conflicts of interest that could adversely affect athlete health.2 Although the high-performance model is the predominate framework used internationally and in some professional organizations in the United States, the medical model has been strongly advocated by both the National Athletic Trainers’ Association and the National Collegiate Athletic Association.2,3 Although the medical model has legal benefits and potential benefits to AT caseload, it also creates a clear “medical silo” that has pitfalls both for the performance and health of individual athletes within the organization and for the athletic training profession as a whole.4,5



Citation: Journal of Athletic Training 59, 10; 10.4085/1062-6050-0580.23
Creating a medical silo means the athletic training staff may not have a holistic view of the physical and cognitive demands being imposed on their athletes or how these demands are being counterbalanced by nutritional or other modalities within their own organization. The high-profile acceptance of the high-performance model at the professional level means ATs have ceded considerable input in the areas of injury mitigation to sport scientists and sport-specific rehabilitation to strength and conditioning personnel who may have a lower level of formal medical training than ATs but a greater focus on sport-specific performance. A prime example of ceding medical expertise to nonmedical staff comes in the form of “workload management” to mitigate injury and enhance performance, but more specifically, the use of the acute to chronic workload ratio (ACWR) to decrease injury in sport.6–8 The ACWR posited that there was a fundamental ratio between acute workload and chronic workload and that a spike in acute work resulted in “injury,” but the measurement of workload was highly nonuniform (accelerometers, surveys, global positioning systems, etc) and “injury” seemed to mean any type of injury at all, depending on the study. The balance between training and overuse injuries has always been a responsibility shared across many practitioners in an organization, something the high-performance model overtly recognizes but the medical model does not, even if some organizations implementing the medical model have ancillary policies and procedures recognizing this shared responsibility. However, the high-performance model originated ACWR, which has now been widely discredited as atheoretical and lacking any causal evidence; it is reasonable to expect that an AT trained in the different processes of both bony remodeling and muscular physiology would not subscribe to the idea that some blanket metric from a survey or an accelerometer should be used as a guide for both stress fractures and muscle strains.9 The high-performance model seems to be more theoretically “correct” as a way to holistically understand athlete health, injury, and performance; however, its implementation seems critically flawed in such a way that it overemphasizes performance at the expense of medicine, which often requires a higher level of scientific evidence to make conclusive statements about patient care. In contrast, the medical model has left ATs in a medical silo (Figure 2), which limits our ability to holistically care for the individual athletes and limits our wider impact in the field of athlete health and injury mitigation. We argue that the medical model is theoretically consistent with the high-performance model only if we reject the notion that the “performance director” is an administrative person and instead conceptualize this as a “health and performance information hub” that facilitates transdisciplinary collaboration.



Citation: Journal of Athletic Training 59, 10; 10.4085/1062-6050-0580.23
THE DIFFERENCE BETWEEN TRANSDISCIPLINARY AND INTERDISCIPLINARY
At best, collaboration in most athletic organizations is interdisciplinary, meaning a group of individuals with different expertise work together to address a common problem, each from their own disciplinary perspectives.10 In other words: “Stay in your lane, I’ll stay in mine, and together we’ll have a team that isn’t injured and wins games/matches.” In reality, ATs know a lot about sports medicine, but also aspects of human performance and technology. The sport scientist may know a lot about technology, but also have ideas about injuries specific to that sport and increasing performance particular to that sport. The strength coach obviously has expertise in strength and conditioning, but their regular exercise regimens with athletes make them uniquely capable to identify potential issues before they’re seen in the athletic training facility. In 1992, Rosenfield coined the term transdisciplinary as an organizational method emphasizing that different disciplines are working from a shared conceptual framework to solve a problem and not just working together from their own disciplinary frameworks (ie, interdisciplinary).10 In the sports context, this means all practitioners should not just be working together in their singular areas of expertise; they should be working jointly, using a shared conceptual framework informed by all of their disciplines to enhance performance and decrease injuries. This does not mean ATs should be expected to “do more sport science and conditioning,” which would be a multidisciplinary approach (one person or group using many disciplines). We also do not suggest that individual disciplines should have “more meetings,” but rather that information and data flows should be standardized and structured around a shared vision to increase athlete health and performance. At the professional and collegiate levels, there is often access to higher-level data infrastructure and analytical talent that can facilitate this type of transdisciplinary work within an athletics organization (Figure 3). In this way, we increase the quality of athlete care while potentially decreasing the administrative burden on the practitioner via the effective implementation of technology.



Citation: Journal of Athletic Training 59, 10; 10.4085/1062-6050-0580.23
To achieve transdisciplinary collaboration, it is necessary for all groups to “speak the same language,” share data, and have an agreed-upon framework from which to review those data that are particular to their specialty. Although this may sound like an insurmountable challenge, it is actually relatively easy to achieve with a data-broker system. With this arrangement, each group agrees to record its data (eg, injury record, nutrition data, body composition data, psychological testing, strength training, travel information) in a system that flows in a centralized data ecosystem. This does NOT require that ATs and nutritionists (for example) use the same vendor or system (eg, Kitman Labs, Smartabase Athlete app, Athletic Trainer System), simply that each of their individual systems can transmit the data into a central location. Once these data are in a central location, the groups simply need to agree upon “rules for who sees what” from a data perspective and, if higher-level analytics are performed, what sort of predictive or inferential models they want built into their system. The data-governance rules can be regulatory in nature (eg, the Family Educational Rights and Privacy Act [FERPA] or the Health Insurance Portability and Accountability Act [HIPAA]), a need-to-know basis instituted locally, or any other rule-based system deemed appropriate by the organization. The data can then be displayed via a real-time dashboard to the appropriate practitioners with the right context (ie, with full data from other disciplines, simplified data from other disciplines [eg, rank within position group or team-level measure], or a prediction taking into account information from other disciplines without providing the underlying data).
At this point, it is valuable to return to our earlier example of workload management. I (M.S.T.) was at a recent conference discussing load management practices with a number of National Basketball Association (NBA) sport science personnel and asked what work had been done to validate that their load management practices were valid in reducing injury, and was surprised to find that the sport scientists were physically unable to pair their “load” data with the injury tracking system. Although the sport scientists described it as “We can’t get the data out of the system,” my later discussions with NBA personnel made it clear that the issue was not technical but political and some sort of internal “turf battle.” Indeed, after years of NBA personnel stating that the science supports “load management,” they have now publicly stated that the data “… just doesn’t show that resting, sitting guys out correlates with lack of injuries, or fatigue, or anything like that.”11 Although organizational politics and turf battles are often inherent in any athletic organization, these roadblocks are often not apparent to upper management (eg, athletic directors, general managers, ownership) who have decisional authority to end unnecessary and unhelpful disputes. The process of setting up a data-governance structure, approved and championed by upper management, makes this discord apparent and able to be adjudicated appropriately.
Clearly, the artificial data and information partition between practitioners in the same organization is not leading to the best health care for the athlete, nor is it likely leading to the best team performance. Within a transdisciplinary organization (Figure 3), it may be agreed that all clinicians want to mitigate overtraining syndrome (a medical condition) and so preemptively identifying athletes in the nonfunctional overreaching stage is of vital importance (ie, identifying a common framework). Each practitioner then notes the below information in their respective systems:
-
Sport nutrition indicates athlete has had decreased calorie consumption coinciding with increased body fat and decreased muscle mass.
-
Sport psychology indicates steadily increasing scores on weekly Beck Depression Inventory and State Anxiety.
-
Strength and conditioning records decreasing sprint speeds, decreasing jump height, and inability to maintain existing standards on the bench press.
-
Sports medicine records athlete has received ice bags regularly after practice.
-
The group handling game dynamics, research and development, notes this athlete’s in-game velocity of travel has decreased and their apparent reaction time to ball movement on the court is longer.
All of these above data points individually are not necessarily flags from the onset that any one group would identify; however, when all of the data are placed together in a central repository, it paints a very clear picture of an athlete in the early stages of overtraining syndrome and someone with high likelihood of sustaining a stress fracture due to relative energy deficiency in sport.12,13 Injury prevention is a team effort, requiring structured and trusted data sharing. It is not necessary for all clinicians to see the individual athlete’s psychological testing data, which would likely be inappropriate. Nor is it helpful or necessary for the nutritionist or strength coach to see every nuance of the athlete’s injury history; however, a central aggregation of data and analytical processing of those joint data sources can provide key information to the appropriate practitioner for intervention on either the mitigation side of the equation or the treatment side. Not only does this transdisciplinary approach allow for the best health care of our individual athletes, it also has the ability to put ATs, and sports medicine as a whole, on much firmer footing to generate evidence that our clinical practice is effective for athlete care.
It’s (Still) Time for Evidence
More than 15 years ago, the Journal of Athletic Training published an editorial entitled “It’s Time for Evidence.”14 The editorial emphasized that ATs were being marketed and using devices and products that were being held to no standard of efficacy or accountability. This stands in stark contrast to other fields of medicine, such as pharmaceuticals, which undergo numerous levels of preclinical, clinical, and postauthorization trials to confirm that the benefit-to-harm ratio is appropriate.15 Many aspects of our industry have not changed since 2006. The devices and products marketed to and used by practitioners across the spectrum of sports medicine (orthopaedic surgery, physical therapy, athletic training, sports nutrition, etc) are still largely unregulated and lack high-quality evidence.
The primary solution proposed in the commentary was to use randomized controlled trials (RCTs) to provide evidence that a product mitigated injury or enhanced return to play. Although RCTs remain the gold standard for determining causal effects of an intervention, RCTs can be expensive to implement and challenging in the field of sport, where you have numerous stakeholders (eg, ATs, sport coaches, team physicians, and athletic directors). Everyone wants “results now,” and it is hard to find a stakeholder who is willing to wait for an RCT to show if an intervention works, even if it is equally likely that said intervention adversely affects injury likelihood or return to play. In recognition of these challenges to implementing RCTs in sports medicine, we would like to introduce our colleagues to another option that allows us to provide a high level of evidence that our clinical interventions cause the changes they purport: real-world evidence (RWE) causal inference methods.16
RWE in Athletics
The US Food and Drug Administration defines real-world data as “data relating to patient health status and/or the delivery of health care routinely collected from a variety of sources,” and in 2018 it defined a framework for using RWE for various types of regulatory review.17 The growing acceptance and development of analytical methodologies to extract causal inferences from RWE can be a valuable tool to validate interventions used by ATs if we embrace and facilitate this work. Turning real-world data into RWE requires collaborating with experts in the field of biostatistics and epidemiology who have a particular skill set in RWE causal inference. Many large universities have 1 or 2 (or even none) of these experts, and a professional clinical research organization will have many. These experts will often need direct access to the centralized data hub in Figure 3, so it is important to consult with and act on the advice of your general counsel (attorney) to ensure that the appropriate guidelines for transmission of identifiable data are being followed (eg, FERPA or HIPAA) and that any appropriate memorandums of understanding are in place. There is often an internal process of setting aside biases or assumptions (eg, “What if my outcomes aren’t as good as I think they are?” “What if that treatment we’ve been doing for 3 years isn’t increasing return to play like I’ve been telling my AD [athletic director] it has?”) that needs to occur because having someone critically evaluate records can create a feeling of a loss of control. Clearly, there needs to be a culture at your local organization that learning what doesn’t work is just as important as learning what works. We also need to understand that those RWE experts aren’t doing this complicated analytical work for “free,” but are often expecting to publish the results (in academia) or want remuneration (in a clinical research organization context). All expectations should be included in any agreements to mitigate potential issues.
Finally, RWE causal inference is a defined skill. It is not as easy as saying, “We had 22 athletes get injured wearing this protective device and 78 athletes injured not wearing the device; therefore, the device works.” Many statisticians will follow the potential outcomes framework to derive causal information from real-world data. Much of this framework has the goal of “making the observational data as much like an RCT as possible,” so there are often complex weighting or matching processes involved, to say nothing of the resulting analyses. Athletic trainers should expect to work with RWE experts because those experts are often not knowledgeable about sport and even less knowledgeable about how the data were recorded. It is a collaborative task to extract evidence from the data we already collect, but well worth it. Ultimately, we have 3 options: (1) continue to practice athletic training with a suboptimal evidence base; (2) work with clinical research experts to develop potentially expensive and time-consuming RCTs, which remain the gold standard of evidence; or (3) leverage the data we already collect and collaborate with RWE analytical experts, potentially already at our institutions, to generate evidence supporting or refuting athletic training practices.
CONCLUSIONS
The medical model is not broken, but it does require reconceptualization. The medical model administrative structure is necessary to limit conflicts of interest that may harm the athlete, but in administratively partitioning off the sports medicine staff from other groups in an organization, we have inadvertently decreased the quality of care that can be provided to the athletes and also decreased the impact that ATs can have in their organization and on the medical community as a whole. There is an opportunity to adopt the philosophical ideas of the high-performance model, while avoiding that model’s pitfall of overemphasizing performance ideas at the expense of high-quality evidence supporting either performance or medical outcomes. Any organizational structure can fail due to poor execution, but the technical and legal solutions exist for us to judiciously share data and information across our organization that enhance athlete care and demonstrate our value to leadership; we just need to champion these efforts and support the formation of a transdisciplinary organizational structure.

Diagram of high-performance model as described by Smith and Smolianov.1

Diagram of typical athletic organization under the medical model.

Diagram of the data and information flow that occurs in a transdisciplinary organization using an automated data broker with real-time information delivery.
Contributor Notes