Next Article in Journal
A Self-Adaptive Strip Pooling Network for Segmenting the Kidney Glomerular Basement Membrane
Next Article in Special Issue
Quantifying the Effects of Detraining on Female Basketball Players Using Physical Fitness Assessment Sensors
Previous Article in Journal
Resilient Communication for Software Defined Radio: Machine Reasoning and Electromagnetic Spectrum Evaluation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Continuous Monitoring of Recruits During Military Basic Training to Mitigate Attrition

by
Robbe Decorte
1,*,
Jelle Vanhaeverbeke
1,
Sarah VanDen Berghe
2,
Maarten Slembrouck
1 and
Steven Verstockt
1
1
IDLab, Ghent University-Imec, Technologiepark-Zwijnaarde 122, 9052 Ghent, Belgium
2
Department of Rehabilitation Sciences, Ghent University, C. Heymanslaan 10, 9000 Ghent, Belgium
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(6), 1828; https://doi.org/10.3390/s25061828
Submission received: 13 February 2025 / Revised: 12 March 2025 / Accepted: 13 March 2025 / Published: 14 March 2025

Abstract

:
This paper explores the use of wearable technology (Garmin Fenix 7) to monitor physiological and psychological factors contributing to attrition during basic military training. Attrition, or the voluntary departure of recruits from the military, often results from physical and psychological challenges, such as fatigue, injury, and stress, which lead to significant costs for the military. To better understand and mitigate attrition, we designed and implemented a comprehensive and continuous data-capturing methodology to monitor 63 recruits during their basic infantry training. It’s optimized for military use by being minimally invasive (for both recruits and operators), preventing data leakage, and being built for scale. We analysed data collected from two test phases, focusing on seven key psychometric and physical features derived from baseline questionnaires and physiological measurements from wearable devices. The preliminary results revealed that recruits at risk of attrition tend to cluster in specific areas of the feature space in both Linear Discriminant Analysis (LDA) and Principal Component Analysis (PCA). Key indicators of attrition included low motivation, low resilience, and a stress mindset. Furthermore, we developed a predictive model using physiological data, such as sleep scores and step counts from Garmin devices, achieving a macro mean absolute error (MAE) of 0.74. This model suggests the potential to reduce the burden of daily wellness questionnaires by relying on continuous, unobtrusive monitoring.

1. Introduction

In a military context, training programs usually start with Basic Training. The candidates, who are called recruits, in this context must undergo an extensive and demanding training period which usually lasts for around six months.
During training, the candidates have both theoretical and practical sessions to help them learn the principles of being a soldier. These demanding sessions often cause attrition, a term used to indicate quitting the army and returning to civil life. Through the formation, recruits quit the army because they cannot successfully complete the training period. Recruits quit because of several reasons related to the build-up of fatigue, chronic overuse, injuries, or mental wear down [1,2,3]. Attrition costs the military a lot of money, money that could be invested towards further developing soldiers that remain in the army for a longer time. Because the recruits’ bodies and minds are often pushed to their limits, the program must balance testing the recruits to their fullest potential (in preparation of tough military operations) and preventing drop-out and overtraining.
The use of wearable technology (e.g., smart watches, step counters, smartphones) for daily wellness monitoring has been heavily discussed in the scientific literature. In healthcare, for instance, most studies agree that the emerging and ever improving wearable technology can not only partly automate health monitoring, but also make it continuous and unobtrusive [4,5] as compared to the alternative: questionnaires. However, an important aspect of our manuscript is how a large group could be monitored with minimal interference in their day-to-day operations. For example, in military training, it is not feasible to have dedicated ‘laboratory days’ due to the time-constrained schedules they operate under, nor can there be someone on-site 24/7, as everything must be independently operated. Consequently, for the sensor, we had to find a balance between measurement validity and practicality—selecting a device that was compact enough to capture multiple parameters while still being robust enough to withstand the harsh military environment.
In this paper, we will study the physiological and psychological factors contributing to attrition in the recruitment phase at the infantry division of the Belgian military. We will discuss the data processing architecture and provide preliminary data science insights on centrally collected data from two test phases. In total, 63 recruits were monitored during their basic infantry training in the military base at Arlon, Belgium.

2. Related Work

Attrition is a phenomenon that has already been studied in the literature. On the mental side, a study by Wolfe et al. [6] in the U.S. military concluded that people with a premilitary interpersonal trauma were 1.5x more likely to drop out of the recruit training. Further, Molloy et al. [7] focused more on the physical side and stated that today’s youth are less prepared for entry-level physical training compared to their predecessors. For physical testing in the U.S. military, they observed the first-time failure rate increasing from 4% in 2003 to 34% in 2009. In Belgium, where this study was performed, more than 40% of recruits leave the army early. Detecting those that are at risk is of the utmost importance.
In the Netherlands, the attrition rate is just as problematic as in Belgium. Furthermore, Huijzer et al. [8], who collaborate with the Dutch Commando Corps, focus on predicting special forces dropout via explainable machine learning on a set of physical and psychological tests performed prior to the training. They found that both physical and psychological variables were related to dropout. More specifically, a higher score on the 2800 m time, connectedness, and skin folds were most strongly associated with dropping out. A similar baseline testing and analysis is also performed in our study and further discussed in Section 5.1. Compared to their study, which only focused on the baseline testing, we also use these insights in the continuous monitoring during Basic Training, as is further explained in the paper. In their follow-up study, den Hartigh et al. [9] extended the baseline testing with a weekly follow-up of experienced psychological and physical stress, recovery, self-efficacy, and motivation. Their results show that low levels of self-efficacy and motivation are significantly associated with dropout and that the dropout could often be predicted multiple weeks in advance based on the weekly follow-up of this data. In our study, we use similar wellness questionnaires, which were collected on a daily basis. Outlier detection on this data is also used to flag recruits that are at risk.
Baseline testing and wellness questionnaires, however, are only part of the puzzle. Combining them with the continuous monitoring of the recruits’ internal/external load, their sleep, and stress gives us much richer information to build models with. A recent study on U.S. Navy populations demonstrated the feasibility of using commercial wearable devices, such as the Oura Ring and ReadiBand, to monitor sleep–wake patterns in operational environments. With over 10,000 person-days of data collected from 845 sailors across multiple ship cohorts (with a respective prevalence rate of 69% and 71% for the ring and band), the study supports the integration of wearables into fatigue management systems, reinforcing the value of continuous monitoring in high-performance settings [10].
For daily health and fitness monitoring of people, the sports and medical world is usually a good example for effective strategies, mechanisms and hardware. A study by Abuwarda et al. [11] researched the possible crossovers of wearable monitoring in healthcare to construction workers. They defined four specific hazards that construction workers are exposed to: slips and falls, collisions with others or materials, stress and fatigue and other factors (e.g., fire and noise). These hazards can be closely monitored by wearables using the sensors equipped on the devices (e.g., accelerometers, body temperature, heart rate, respiratory and location sensors). One must also consider how invasive all these sensors are and what the impact is when multiple of them are worn. Take for example heart rate, which is typically measured with a chest strap. Alternatively, a photoplethysmograph sensor could be utilized. It is a common principle that measures the light absorption of local tissue over time. The principle is often used in wearables to measure heart rate as the amount of light absorption is related to the heartbeat pattern. Heart rate and heart rate variability (HRV) are often used to monitor performance in sports and other physical activities (as reviewed by Stephenson et al. [12] in military context, or by Addleman et al. [13] in sports related fields).
In a publication of Kutilek et al. [14], a review study of the possibilities of wearable technology for the health monitoring of soldiers was performed. In their work, the authors summarized wearable technology that is suited for military use. They conclude that heart rate, electrocardiogram (ECG), and respiration rate are the basic indicators for overall condition monitoring.
In the U.S. Marines, Saxon et al. [15] experimented with Apple iPhones and Watches to continuously monitor the specialized training phase of the elite military units to study how and when attrition takes place. They concluded that most of the withdrawals (23.5%) already took place before day seven of the training. Furthermore, they divided the attrition into different subcategories (i.e., Drop-On-Request, Medical, Safety, Performance). Drop-on-request is the most common one, in which the failure is trainee-initiated. The study used continuous monitoring to better isolate the attributing factors of attrition in this category, showing that continuously monitoring mental and physical status during military training is possible.
Although there are already existing studies that use continuous monitoring in the analysis and prediction of attrition, as discussed above, none of them focus on how to efficiently and effectively implement it in a challenging defence context. This exactly is what we have investigated. In what follows, we further explain the need for a scalable, secure monitoring set-up and describe the technological building blocks of our solution and the challenges related to it.

3. Methodology

As mentioned in the introduction, the main goal of this research project is to prevent and understand attrition in the military recruitment phase of the Belgian army. To achieve this goal, a preliminary device and technology were studied. For the device, the main consideration was the need for an unobtrusive device with minimal impact on the daily schedules and sessions of the recruits. Furthermore, the battery life of the device was also an important factor. During training sessions, troops often go on-field and perform orienteering, survival, scouting, and protection tasks in remote areas for multiple days, where no reception or mains power is available. This means that the recording device not only needs to have outstanding battery life but also that the offloading of data needs to be as quick and unobtrusive as possible.
For the data capturing, a central storage solution was developed. This allows for collaboration between researchers and military personnel and enables quick and clean data analysis from a central data source. A schematic overview of the overall data capturing and storage architecture is presented in Figure 1.
The data collection protocol consists of three main elements: the wearable watch capturing health-related monitoring data, a smartphone application to wirelessly collect data from the watches, and a central server to collect all the data from a smartphone.

3.1. Wearable Selection

As mentioned in the related work section, Saxon and colleagues proposed a good starting point for monitoring with wearable devices. They used Apple iPhones and Watches to collect the data. This choice of hardware implies frequent device recharging, a task that was assigned to dedicated personnel for their study. For our study, we had the explicit requirement to be as unobtrusive and low impact as possible. This implicated that device autonomy was an important aspect in our final choice. Table 1 shows comparative information between a line-up of two consumer-focused smartwatches and three wearable sensors, which are more often found in heavy-duty industries such as firefighting, military operations, or patient recovery. A big advantage of these consumer-focused smartwatches is that they are more widely applicable for the population and thus more cost-friendly compared to specialized products. Another consideration to make is that with a single device, we can measure a range of physiological parameters, while others (such as the eqLifemonitor and the Axioma Padis) either require multiple attachment points or measure a single parameter. Although these multi-component setups have proven to have a high validity compared to gold standard measurements [16,17], they are still much harder to implement on a large scale with a minimal impact approach (i.e., somebody needs to wear/charge/care for multiple devices).
For our study, we selected a recent iteration of Garmin’s Fenix lineup, the Fenix 7, since it is a balanced combination of an interesting feature set for daily wellness monitoring, user-friendliness, and device sturdiness for a military context (i.e., some watches were severely scratched and damaged on the display after the first testing phase). The device has a claimed 18 days of battery life, with GPS disabled for typical use. For continuous and fine grained location tracking (using GPS, Galileo and GLONASS), the watch can last 40 h, or 57 h when only GPS is used [23]. So, this is a reasonable battery life for expeditions during the training phase when recruits can manually pause location tracking when stationary for prolonged periods of time (e.g., when sleeping or during briefings). The watch is also equipped with useful sensors such as a gyroscope, accelerometer, pulse oximeter, and a wrist-based heart rate monitor. Another very important factor for selecting this device is that for partner institutions, the Garmin ecosystem offers a device management and data extraction Software Development Kit (SDK) for both Android and iOS, called the Health SDK (currently on version 4.2.7). One potential drawback of using a smartwatch as a sensing device is that, in this context, its screen is visible when viewed through night vision goggles. While this issue can be mitigated during training by covering the screen (which also prevents normal watch usage), this may not be a practical solution in operational environments. If the system were to transition to an operational deployment, it might be more appropriate to use the smartwatch primarily for assessing readiness-to-perform rather than having it worn during sensitive night-time missions.

3.2. Wearable Management and Data Synchronization

3.2.1. Integrating the Garmin Health SDK

Through the Health SDK, it is possible to take control of the Garmin wearable used in the study through a self created smartphone app that integrates the functionality of the SDK. There are two versions available: the Standard SDK [24] is used to access all health and fitness data through your own services, while the Companion SDK further extends this to access real-time sensor streams. In this research, the Standard SDK was sufficient, as real-time access was not required. So, through our custom app, it was possible to synchronize the watch data while maintaining full control of the data. This is an attractive mechanism in the military since external servers (commonly used to process data, e.g., Garmin Connect servers) can be completely circumvented by use a self provided aggregation/processing server instead. This implies that it can also be deployed in a closed and safe environment, since every aspect is self managed and can even work offline if installed in a local network. In a military context, it is important that all data remain on military network infrastructure so that critical information is not leaked publicly. An extra reason for this choice is that multiple watch connections can be managed on a single phone with the Standard SDK. This means that rather than providing each recruit with a dedicated single syncing device, the data synchronization can be managed centrally and controlled by the study personnel or by one of the commanding officers, which also further reduces the required workload by the military. This principle is further referenced as an operator sync. One downside of the watches is that they are rather costly for use on a large scale. The Garmin ecosystem, however, has lower priced watches (e.g., Venu 2 or Forerunner 245), but often with less battery capacity, a smaller subset of supported features, or lower build quality compared to the Fenix line. Although these watches could be used without modifications to the data syncing and storage strategies, whether all required features are supported should be checked with the Health SDK documentation, for example, checking if on-device sleep analysis is supported. If not, then the SDK will send anonymized data to a Garmin server. To conclude the wearable selection discussion, we collected a user experience questionnaire from the recruits who did wear the watches 24/7 in test phase 1. They reported that they were largely satisfied with the perceived comfort of the chosen watch and no complaints were received during test phase 2.
The watch itself has two important operation modes. Activity mode is equal to that of a watch not managed by the Health SDK and collects relevant data on the selected sport supplemented by location information, etc. This mode is actively initiated and stopped by the recruit or on command by the supervisors. Activities recorded in this mode were based on their itinerary and would often include exercises like orienteering, scouting, and physical education, where the added location information and calculated derivatives are meaningful. On the contrary, when no activity was being recorded, the watch was in monitoring mode. In this case, the watch passively monitored and recorded physiological parameters, such as heart rate, stress, respiration rate, blood oxygen saturation, beat-to-beat intervals, and movement data like the amount of steps and raw accelerometer and gyroscope data. It also collected sleep quality measurements and time spent in sleep stages. Using the management functions of the Health SDK, these functions can be activated on the watch, in addition to the frequency rate at which it should record (for most parameters between 1–3600 s, accelerometer between 1–24 Hz, and gyroscope between 1–32 Hz). Note that the amount spent in activity mode and the amount of functions activated in monitoring mode may have a significant impact on the battery charge and the duration of the synchronization procedure.

3.2.2. Offloading and Processing of the Wearable Data

The offloading of the captured data on the watches to the processing server was performed in two steps. First, the data were collected by the smartphone using the operator sync procedure. Based on the model and the Bluetooth antenna of the smartphone, a fixed number of Bluetooth connections to watches can be maintained. On a Xiaomi Redmi Note 9 Pro and Samsung Galaxy A14, the hardware permits a maximum of 10 open Bluetooth connections be maintained, but a maximum of 5 watches is advised to not saturate the Bluetooth manager (although this can be increased non-concurrently, as further elaborated in Section 5.3). So, using this system, we can achieve at least a 1:5 ratio between the watches and the collection devices, which can be extended to 1:10 through advanced connection management. This is a significant cost reduction compared to a 1:1 sync model. Another option would be to reuse recruits’ personal devices, but this would be much harder to manage. Some control concerning the confidentiality is lost and it relies on the goodwill of the recruits to use everything correctly. It could also create issues about who is responsible for damage to personal devices if they are used for work-related purposes. As an extension, we have also created a parser for watches that are backed up over USB, in which case the data are automatically uploaded to the processing server (and the section below is not applicable).
After the initial pairing of a watch to the smartphone (which only happens once), the watch synchronizes its data to the smartphone. The app and SDK support concurrent synchronization and automatically initiate the data transfer when devices are in range. Whenever new data are received, the app will extract the relevant information out of the transferred binary files before saving to the app’s embedded database. After all the watches are synced to the phone, the operator can initiate the upload process. This will save chunks of data from the internal database to files on the phone’s file system. When a connection to the processing server is available, a request is sent for each created chunk file. This is performed as as to not transfer too many data at the same time (and to respect the http message limits imposed by the server). As a connection is not always stable or may be unavailable due to poor data reception on the field, the app has to contain a failover strategy. So, if no connection is available or it suddenly drops, the chunk files remain on the phone’s file system. The next time an upload attempt is made with a connection, these files will be picked up for transfer first. Moving the data out of the internal database is a necessary step. Based on which measurements are collected (and their sample rate), the app may have to process large amounts of data. By transferring it to outside storage, the queries performed on its database remain fast and thus this keeps the app more responsive. This also makes it less susceptible to memory issues.
During testing, the data were usually uploaded to servers once a week either by study personnel or by one of the commanders of the platoon. The overall syncing strategy can be executed by the military personnel as it is a straightforward process that can be easily performed during, for instance, lunch or a theory lecture. The watches have an advantageous feature in that can sync and charge simultaneously. This can be used to further limit the time spent off a recruit’s wrist. For syncing and charging, we utilized a self-made charging station consisting of a USB charging hub to place and charge nine watches and an attachment point for the syncing smartphone (see Figure 2).

3.3. Data Processing

The data on the server were stored in a Postgres SQL (v17.0) database with TimescaleDB (v2.17.0) [25] extensions to allow for efficient date time and timestamp operations. The data were pushed from the smartphones to the server through an Application Programming Interface (API) built with ASP.NET Core (v9.0). This API was also used to manage partner interactions built on top of the collected data. These are outside the scope of this publication, but some examples are as follows: monitoring of the metabolic energy balance [26], analyses of structured running workouts [27] during PE sessions, body scans [28], and the integration of sweat-lactate measurements [29] for selected exercises. Authorized partners can query the information relevant to them and also add their own results in the centralized database. The data on the server were roughly subdivided into five categories: monitoring data, activity data, sleep data, injury reports, and wellness questionnaire data. Monitoring data, activity data, and sleep data are pushed from the watches to the database through the Android application. The other data (i.e., wellness questionnaires and injury reports) can be inserted via a web-based front-end application by the medical examiner (for injury reports) or by the recruits or study personnel for wellness questionnaires (when they are filled out on paper while in a connectionless environment). This workflow allows for the further integration of other measurement tools and information with minimal modification to the base code (e.g., inclusion of blood sample results, clinical refractometer, or nutrition data).
To conclude the discussion on the hardware and software setup, we should mention that the architecture was built with the information criticality and sensitivity of a military context in mind. Firstly, the choice for the Garmin Standard SDK made it possible to keep the data on military-owned hardware (i.e., watches and smartphones), rather than sending the data through Garmin servers (i.e., the Garmin Connect ecosystem). Secondly, the server that the data are ultimately pushed to is also in a protected environment and is managed by the study and military personnel. Both choices imply that with the correct system and hardware security precautions, the data will never leave the military network and thus will not be pushed to the internet or public servers.

4. Dataset

Research was conducted during two test phases during 2023 and 2024. A summary of the two studies (test phase 1 and 2) is shown in Table 2a. Test phase 1 lasted 8 weeks, while test phase 2 lasted 17 weeks. The number of recruits, aged between 18 and 27, that were monitored with watches also increased significantly from 17 to 46. The attrition number increased both in absolute values as well as relatively: 18% and 24% for test phase 1 and 2, respectively. Within the attrition numbers, we also investigated reasons for attrition, as reported in the study of Saxon et al. [15] in Table 2b. During the test phases, all attritions were registered under Drop-On-Request or Medical, with the majority being drop-on-request.
Each recruit was equipped with a Garmin Fenix 7 which recorded various types of data. Ranging from full activities with GPS coordinates to passive sensor measurements such as heart rate, steps, stress, body battery, respiration rate, energy expenditure, comprehensive sleep assessments (scores, stages, duration), and beat-to-beat intervals during sleep. Before and during the test phases, additional data were also obtained by means of questionnaires. These questionnaires were divided into two three categories: a series of baseline questionnaires at the beginning of each test phase, a weekly questionnaire, and a daily questionnaire. In test phase 1, these questionnaires were collected on paper using a dedicated mailbox while in test phase 2 used a smartphone app was used to notify recruits to fill in the daily and weekly questionnaires. This change to an electronic collection mechanism was part of an effort to create a more independent and efficient system (compared to pen and paper) and to possibly improve compliance with a more user-friendly interface. Although these changes decreased the workload for the study personnel, we observed a downward trend where the compliance dropped gradually as weeks went by. Efforts by researchers and staff sparked compliance rates for short amounts of time but remained a sticking point unless they were constantly reminded. One of the challenges of this research was to cope with missing data. As later discussed in Section 5.2, we could try to predict missing values using data collected through the watch.
Besides the wearable data and questionnaires, we also collected context metadata from the staff, which gave more insights into the particular courses and sessions the recruits performed. It allowed us to filter the data streams by session and highlighted other information such as the intensity of each session (estimated by the instructor), if they were equipped with heavy equipment or weapons, if it was coupled to an evaluation or other stress-inducing factors, and if it was performed outside. A physical baseline was also obtained for each recruit at the beginning of the test phase, with follow-ups during each phase.

5. Results and Discussion

As mentioned in the introduction, the data capturing setup had the goal of finding contributing factors towards attrition and ultimately detecting recruits that might be prone to dropping out the formation in the near future. The results in this paper will mainly focus on two parts: the baseline questionnaire data and the 24/7 monitoring data of the Garmin watches.

5.1. Baseline Classification

Mental and Physical Baseline

A first analysis was conducted on answers from the questionnaires measuring the psychological characteristics and physical test results of the recruits. Table 3 lists the different questionnaires that were filled in by the recruits. Four published questionnaires were used and a custom domain-specific questionnaire was also conducted. Additionally, the results of the 2400 m running test were also included. The 400 m sprint and 5000 m running test were also measured, but the number of recruits that executed all tests was limited. Therefore, we decided to only include the 2400 m test because the results of the three tests (for the recruits that did all of them) strongly correlated.
Based on eight features (seven psychometric features and one physical feature), we were able to include 56 recruits in the analysis, of which 15 underwent attrition during the course of the Basic Training program. We performed two standard methods: Linear Discriminant Analysis (LDA) and Principle Component Analysis (PCA). LDA is a supervised method that takes into account the classification (attrition or no attrition), while PCA is unsupervised. While LDA performs better, the difference with PCA is small.
Can we identify recruits that are at risk of attrition by solely considering the responses to these questions? In order to find out, we performed supervised dimensionality reduction, where the classification (attrition and no attrition) was used as class labels.
In Figure 3a and Figure 4, we observe that most recruits who experienced attrition are positioned on the right side of the feature space. While distinct clusters are not immediately apparent, this analysis can help identify recruits who may be more susceptible to dropping out. The supervised PCA weights in Figure 3b provide insight into why recruits are positioned on the right side of the graph (PC1 = blue). Both the weights and feature values can be either positive or negative, and to move to the right, their multiplication must yield a positive result. This means that the values must either be both positive or both negative.
By knowing the weights for each feature, we can determine the values that would lead to a recruit moving to the right side of the graph. For instance, a negative weight for motivation means that recruits with low motivation at the start of Basic Training (i.e., a negative value) will push them further along the PC1 axis. Similarly, low scores in resilience (CD Resilience), stress mindset (SMM Stress, which reflects a debilitating stress mindset), consistency (SGS Consistency, which indicates low consistency of interest), and perseverance (SGS Perseverance) will also contribute to an increase along the PC1 axis. On the other hand, CHQ ME (morningness-eveningness) and CHQ AM (amplitude) have positive weights, meaning that higher scores on these traits (which indicate more significant changes in circadian rhythm) will increase the PC1 value. The cumulative explained variance of the principal components is shown in Figure 5.
These results can be used to pay closer attention to recruits with PC1 > 0 and to divide the groups more equally.

5.2. Predicting Self-Reported Sleep Scores

Previous studies conducted in different branches of the military of several countries have shown that questionnaire responses can be used to identify at-risk personnel during basic military training [8,34,35]. Although useful, it still causes a disruption when they need to be filled in on a daily basis. Electronic-based systems (e.g., through a smartphone app) require that every recruit has access to such a device and is responsible for it to be charged at the time when the questionnaire should be filled in, which may not always be the case on multi-day operations. In the rural areas where a training camp may be organized, a stable internet connection cannot always be provided. Another approach that uses paper and pen does not suffer from these issues but requires substantially more work and is not scalable since somebody has to input all the forms in the platform, and it is more susceptible to be forgotten since no automated notification can be provided. It also seems to be a step in the different direction when it comes to modernizing data collection protocols in the military. Therefore, it would be useful to develop a method to predict these responses using physiological measurements taken by the Garmin smartwatch. This could alleviate the workload on the recruits by reducing the length of the daily/weekly wellness questionnaires, and perhaps in the future, might eliminate the need to complete the questionnaires altogether. For some cases, it could also be used as a data-imputation strategy, such that missing values in the collected data could be filled in using a predicted value instead of removing that instance.
During the second testing phase, recruits filled in the daily wellness questionnaire (DWQ) every morning during breakfast. It consisted of questions for each of the following categories, which were scored between 1 and 5: sleep quality, motivation, mental recovery, muscle soreness, energy intake, and physical recovery. The total amount of DWQ entries in this test phase was 2543. In the remainder of this section, we focus on developing a predictive scoring technique, focusing mainly on the design for sleep scores with a technique that is transferable to other parts of the questionnaire as well.
We chose to analyse the self-reported sleep scores in greater depth, as this factor is often linked to attrition, particularly in its relationship to stress, fatigue, and motivation. For instance, Taylor et al. [36] found that U.S. Air Force trainees experiencing frequent sleep difficulties were 2.7 times more likely to be discharged compared to those without such difficulties. Similarly, a study by Bulmer et al. [37] revealed that perceived sleep quality had the strongest connection to changes in stress, recovery, and post-sleep fatigue, while the average sleep duration (6.3 h) could negatively affect training outcomes during basic military training (BMT).
The distribution of each measured sleep score in the function of the reported score is shown in Figure 6. The Spearman correlation coefficient between both values is 0.51. Since this is only a moderate correlation, it is inappropriate to use the reported sleep score as a drop-in replacement for the self-reported sleep score. The (sometimes large) difference could be attributed to the inability to cope with the harsh military environment, which is significantly different compared to the living circumstances of the general population. Some of the error could also be explained by the measurement accuracy of the device. The validity of the hardware itself is not further discussed in this publication, but studies of consumer-focused wearables are still regularly performed [16,17].
As each questionnaire entry was filled in during breakfast, we took the measurements from the day before as the input. Feature extraction of the nightly BBI data was processed using tsflex [38]. Experiments were performed using both strided-window and full-night statistics, but for the technique presented later in this section, the full night measurements were used. From the BBI data, we calculated the following measures: mean, standard deviation, minimum, maximum, maximal difference, and the skew. These are combined with the reported sleep scores of the Garmin of that sleep interval (see Table 4). The final feature used represents the physical exertion level of the day before, represented as the amount of steps taken.
An initial look into the data using Linear Discriminant Analysis (LDA), as shown in Figure 7, indicates that the different scores are somewhat separable using the underlying data. Higher scores of 5 mostly adhere to 0 L D 1 , while lower scores are more concentrated at L D 1 1.5 . As this visualization technique only considers a linear relationship, we expect to further refine this with a non-linear modelling technique. However, an important consideration for choosing an appropriate modelling technique is that it has to handle class imbalance well. As shown in Figure 8, the availability of lower scores on the entry distribution is noticeably lower than values in the middle/higher part of the spectrum. This is to be expected due to the fact that most people that will probably feel better more than they feel bad, and those that consistently feel bad are more likely to drop out and no longer provide low values. But since we are mostly interested in these values (as they are the indicators for at-risk recruits), we should take some preventative measures in the design and evaluation of the model to avoid class bias. Therefore, a Balanced Random Forest Classifier (BRFC) [44] was selected for further experiments. The main difference with a traditional random forest is that it takes a bootstrap sample from the minority classes and sample she same number of samples from the majority classes with replacement. This down-sampling technique of the majority class artificially alters the class distribution such that each class is represented equally in each tree.
Evaluation parameters of the model are calculated using 5-fold cross validation. Folds are homogeneous with respect to the recruit, which means that data from a recruit can’t be in the training and the testing set simultaneously. Folds are created by sampling recruits until they contain at least 15% of the recruits. As this is cross-validation, there is no replacement in sampling the recruits. The distribution of each fold is shown in Table 5.
The mean residuals of the predicted scores of the best performing model are shown in Table 6. The macro mean absolute error (MAE) is calculated by averaging the MAEs per class, meaning that the under-represented classes in the splits are not dominated in the overall result. From the different experiments (sets of variables), we concluded that the combination of Garmin scores (from Table 4) and the amount of steps performs best, with an overall macro MAE of 0.74. For comparison, when transforming the steps based on the heart rate zone, the macro MAE was 0.77, and with only Garmin sleep scores it was 0.76; with the sleep scores, BBI derivatives, and steps, respectively, transformed steps, values of 1.05 and 1.11 were found. The feature importance of the BRFC indicates that the sleep duration score (0.21), overall sleep score (0.17), and number of steps (0.13) had the most impact on the predictions. From these results, we conclude that a non-linear transformation on only the Garmin-provided sleep scores can be utilized to estimate self-reported sleep scores (with an average error of 0.74) in harsh military environments.
Table 6 presents the mean residuals of the predicted scores for the best-performing model. The macro mean absolute error (MAE) was computed by averaging the MAEs for each class, ensuring that under-represented classes in the data splits are not dominated in the overall results. Based on the different experiments (using different variable sets), we found that the best performance was achieved using a combination of Garmin sleep scores (from Table 4) and step count, yielding an overall macro MAE of 0.74. For comparison, transforming steps based on heart rate zones resulted in a slightly higher macro MAE of 0.77. Using only Garmin sleep scores produced a macro MAE of 0.76, while incorporating sleep scores, BBI derivatives, and steps (or transformed steps) led to values of 1.05 and 1.11, respectively.

5.3. Garmin Fenix 7 Synchronization Benchmark

To ensure smooth integration into the military workflow without adding delays to the already time-constrained schedules of basic training camps, we must focus and optimize two important aspects of the system: simultaneous action execution and concurrent synchronization. The first aspect is already handled by providing the mobile charging stations, such that the data on the watches can be unloaded while charging at an appropriate time (i.e., during a theory or lunch session, such that a minimal amount of activity data are lost). While concurrency is something we can manage through the operator synchronization mechanism. From an operation standpoint, the concurrency factor is mainly limited by the amount of simultaneous connections that can be managed by the BLE manager and the operating system. Through empirical testing and discussions with a Garmin representative, we have learned to limit the number of connected watches to five. While some phones, like the Xiaomi Redmi Note 9 Pro and Samsung Galaxy A14, can support up to 10 connections, exceeding 5 can overwhelm/saturate the BLE manager. This can lead to issues within the Health SDK, resulting in gaps in the transferred data.
From SDK version 4.2.7 onward, new functionality can be leveraged to manage the BLE connections made by the SDK. This means that we are able to programmatically cycle through the paired devices by activating and performing a concurrent sync with a few watches (an insight into how many is given below) before selecting a new set of watches. Since the supported amount of paired but inactive devices is higher than the amount that can be concurrently synced, this approach scales by 2-3x without needing extra phones.
The procedure followed to obtain the benchmark results in Table 7 was as follows. The amount of watches needed for each iteration were loaded with the exact same amount of data, which consisted of two weeks worth of monitoring data. The watches are then paired to a freshly installed app, for instance on a Samsung Galaxy A14. When the app started, it automatically connected with the watches and started the synchronization process. The timings of the start and end signals for each watch were collected through the relevant action handlers provided through the Health SDK. This process was repeated five times for every iteration of the specified amount of watches. Each iteration round in the results table is represented by two measurements, in seconds. The mean individual completion time of a watch (Ind.) is independent of the other watches in the round and the total time required to complete all watches in the round, or the aggregated completion time (Agg.).
This table shows that the variation in completion times for singular devices can be quite large, since the BLE manager takes liberty in the scheduling process of the data transfer and the OS requires extra time/memory for the garbage collector to run. But, compared to the variation in the completion times of the full rounds, it is much lower and thus more consistent. Even though the scheduling behaviour of the BLE manager may seem unpredictable, it has only a limited impact on the total synchronization time. At first glance, without the Bluetooth hopping option, it might appear more efficient to perform one-to-one synchronization five times, as managing multiple concurrent connections adds overheads to the BLE manager. However, scaling this approach comes with significant drawbacks. Either the cost of additional hardware would be paid, or a large amount of manual effort would be needed to ensure only one active connection at a time, such as manually disabling BLE on all other devices and cycling through them. Therefore, the added time required for a more autonomous synchronization process is justified by avoiding these complications. However, most of these problems are irrelevant whenever this connection cycling is performed automatically. In this situation, the only extra delay that must be factored in is the time between sending the reconnection signal and when it actual connects. Typically, this is only 10–20 s, but outliers of 1–2 min have been observed during testing when the Garmin watch prefers battery savings rather than continuously polling for a Bluetooth connection.
Note that the described upper limits of the system may not always be the most performant for every case. By increasing the amount of watches a phone has to cycle through, the required time also increases. In some cases, horizontal scaling may be more appropriate by adding more phones. From our experience with cycling between 10 watches per phone, we conclude that combining the Bluetooth hopping with 2–3 concurrent devices works best.
This gives a general idea of the applicability and time slots in which it should be scheduled in. There are of course other factors that have an impact during deployment and affect synchronization times, such as which activities are recorded (some of which the duration will vary due to it being tied to the effectiveness of the recruit) and if the watch was consistently worn.

6. Conclusions and Future Work

In this publication, we present a comprehensive data-capturing methodology designed to monitor recruits during their training phase, alongside preliminary insights from two testing phases. While the current dataset is not yet sufficient to develop early-stage attrition detectors, our analysis identified several potential factors contributing to attrition, including low motivation, resilience, and stress mindset. These findings offer valuable insights for improving the monitoring of at-risk recruits and optimizing group divisions during Basic Training. Follow-up studies are scheduled for late 2024, where we plan to refine both the hardware and software, optimize the monitoring setup, and conduct further analysis into the patterns leading to attrition.
Additionally, we introduced the design of a large-scale, minimally invasive 24/7 monitoring solution aimed at reducing manual intervention during field use. The methodology is still being actively researched to achieve a balance between minimizing the required time and expertise of military personnel and reducing the need for research staff involvement. The long-term goal is to enable autonomous deployment of the monitoring system without research intervention. Future phases will also explore the integration of the wearable device for wellness questionnaires, with the potential to eliminate the need for additional smartphones or paper forms, streamlining the data collection process.
Our analysis of mental and physical characteristics based on eight key features from baseline questionnaires reveals that recruits at risk of attrition tend to cluster on the right side of the feature space in both LDA and PCA. While these clusters have some overlap close to the threshold boundary, they help in identifying recruits more likely to drop out. These insights can better guide support systems for at-risk recruits. Furthermore, our predictive model for self-reported sleep scores, leveraging physiological data from Garmin devices combined with step count, produced promising results, with a macro MAE of 0.74. This method shows potential for reducing the need for and duration of daily wellness questionnaires, particularly in challenging military environments, while still offering reliable insights into recruits’ well-being. Another potential use-case could also be predicting future reports, allowing for the early flagging of potential risk recruits who may require an intervention from the staff.
Future work will focus on refining this predictive approach to further reduce the burden on recruits and to improve accuracy across various wellness categories. Additionally, there is potential to extend this method as a data imputation strategy for filling in missing intervals in the collected dataset, enhancing the robustness of the monitoring system.

Author Contributions

Conceptualization, R.D., J.V., S.V.B., M.S. and S.V.; Data curation, R.D. and M.S.; Formal analysis, R.D., J.V. and M.S.; Funding acquisition, S.V.; Investigation, R.D., S.V.B., M.S. and S.V.; Methodology, R.D., M.S. and S.V.; Project administration, R.D., S.V.B., M.S. and S.V.; Resources, S.V.; Software, R.D., J.V. and M.S.; Supervision, M.S. and S.V.; Validation, R.D., J.V. and M.S.; Visualization, R.D. and M.S.; Writing—original draft, R.D., M.S. and S.V.; Writing—review & editing, R.D., J.V., S.V.B., M.S. and S.V. All authors have read and agreed to the published version of the manuscript.

Funding

The research leading to these results was funded by the Belgian Ministry of Defence under contract number 22DEFRA002.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Review Board (or Ethics Committee) of Ghent University (id 12407, protocol code ONZ-2023-0273, approved 28 June 2023).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data are not publicly available but may be available from the corresponding author subject to further clarification and subsequent approval from the Belgian Ministry of Defence.

Acknowledgments

The authors would like to express their gratitude to the recruits and staff on the military base in Arlon for their cooperation and for allowing the monitoring phases to take place during an already time-pressured training period.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Vrijkotte, S.; Roelands, B.; Pattyn, N.; Meeusen, R. The overtraining syndrome in soldiers: Insights from the sports domain. Mil. Med. 2019, 184, e192–e200. [Google Scholar] [CrossRef] [PubMed]
  2. Schram, B.; Pope, R.; Orr, R. Injuries in Australian Army full-time and part-time personnel undertaking basic training. BMC Musculoskelet. Disord. 2019, 20, 6. [Google Scholar] [CrossRef] [PubMed]
  3. Hauschild, V.D.; Lee, T.; Barnes, S.; Forrest, L.; Hauret, K.; Jones, B.H. The etiology of injuries in US Army initial entry training. US Army Med. Dep. J. 2018, 22–29. [Google Scholar]
  4. Jat, A.S.; Grønli, T.M. Smart Watch for Smart Health Monitoring: A Literature Review. In Proceedings of the Bioinformatics and Biomedical Engineering; Rojas, I., Valenzuela, O., Rojas, F., Herrera, L.J., Ortuño, F., Eds.; Springer: Cham, Switzerland, 2022; pp. 256–268. [Google Scholar]
  5. Krizea, M.; Gialelis, J.; Protopsaltis, G.; Mountzouris, C.; Theodorou, G. Empowering People with a User-Friendly Wearable Platform for Unobtrusive Monitoring of Vital Physiological Parameters. Sensors 2022, 22, 5226. [Google Scholar] [CrossRef]
  6. Wolfe, J.; Turner, K.; Caulfield, M.; Newton, T.L.; Melia, K.; Martin, J.; Goldstein, J. Gender and Trauma as Predictors of Military Attrition: A Study of Marine Corps Recruits. Mil. Med. 2005, 170, 1037–1043. [Google Scholar] [CrossRef]
  7. Molloy, J.M.; Feltwell, D.N.; Scott, S.J.; Niebuhr, D.W. Physical training injuries and interventions for military recruits. Mil. Med. 2012, 177, 553–558. [Google Scholar] [CrossRef]
  8. Huijzer, R.; de Jonge, P.; Blaauw, F.J.; Baatenburg de Jong, M.; de Wit, A.; Den Hartigh, R.J.R. Predicting special forces dropout via explainable machine learning. Eur. J. Sport Sci. 2024, 11, 1564–1572. [Google Scholar] [CrossRef]
  9. den Hartigh, R.J.R.; Huijzer, R.; Blaauw, F.J.; de Wit, A.; de Jonge, P. Early identification of dropouts during the special forces selection program. Sci. Rep. 2025, 15, 3242. [Google Scholar] [CrossRef]
  10. Kubala, A.G.; Roma, P.G.; Jameson, J.T.; Sessoms, P.H.; Chinoy, E.D.; Rosado, L.R.; Viboch, T.B.; Schrom, B.J.; Rizeq, H.N.; Gordy, P.S.; et al. Advancing a U.S. navy shipboard infrastructure for sleep monitoring with wearable technology. Appl. Ergon. 2024, 117, 104225. [Google Scholar] [CrossRef]
  11. Abuwarda, Z.; Hegazy, T.; Oetomo, A.; Morita, P.P. Using wearables to monitor and mitigate workers’ fatigue. In Proceedings of the Canadian Society of Civil Engineering Annual Conference; Springer: Berlin/Heidelberg, Germany, 2021; pp. 587–597. [Google Scholar]
  12. Stephenson, M.D.; Thompson, A.G.; Merrigan, J.J.; Stone, J.D.; Hagen, J.A. Applying heart rate variability to monitor health and performance in tactical personnel: A narrative review. Int. J. Environ. Res. Public Health 2021, 18, 8143. [Google Scholar] [CrossRef]
  13. Addleman, J.S.; Lackey, N.S.; DeBlauw, J.A.; Hajduczok, A.G. Heart Rate Variability Applications in Strength and Conditioning: A Narrative Review. J. Funct. Morphol. Kinesiol. 2024, 9, 93. [Google Scholar] [CrossRef] [PubMed]
  14. Kutilek, P.; Volf, P.; Viteckova, S.; Smrcka, P.; Krivanek, V.; Lhotska, L.; Hana, K.; Doskocil, R.; Navratil, L.; Hon, Z.; et al. Wearable systems for monitoring the health condition of soldiers: Review and application. In Proceedings of the 2017 International Conference on Military Technologies (ICMT), Brno, Czech Republic, 31 May–2 June 2017; pp. 748–752. [Google Scholar]
  15. Saxon, L.; DiPaula, B.; Fox, G.R.; Ebert, R.; Duhaime, J.; Nocera, L.; Tran, L.; Sobhani, M. Continuous Measurement of Reconnaissance Marines in Training With Custom Smartphone App and Watch: Observational Cohort Study. JMIR Mhealth Uhealth 2020, 8, e14116. [Google Scholar] [CrossRef] [PubMed]
  16. Grandner, M.A.; Lujan, M.R.; Ghani, S.B. Sleep-tracking technology in scientific research: Looking to the future. Sleep 2021, 44, zsab071. [Google Scholar] [CrossRef]
  17. Chinoy, E.D.; Cuellar, J.A.; Huwa, K.E.; Jameson, J.T.; Watson, C.H.; Bessman, S.C.; Hirsch, D.A.; Cooper, A.D.; Drummond, S.P.A.; Markwald, R.R. Performance of seven consumer sleep-tracking devices compared with polysomnography. Sleep 2020, 44, zsaa291. [Google Scholar] [CrossRef]
  18. Garmin. Fenix 7 Standard/Solar/Pro Series Owner’s Manual. Available online: https://www8.garmin.com/manuals/webhelp/GUID-C001C335-A8EC-4A41-AB0E-BAC434259F92/EN-US/fenix_7_Series_OM_EN-US.pdf (accessed on 22 August 2024).
  19. Fitbit. Charge 5 User Manual. Available online: https://help.fitbit.com/manuals/charge_5/Content/manuals/Topics/Basics/Adjust%20settings.htm (accessed on 22 August 2024).
  20. Movesense. MOVESENSE MD User Guide. Available online: https://www.movesense.com/wp-content/uploads/2021/02/Movesense-MD-User-Guide-R78-Public.pdf (accessed on 22 August 2024).
  21. Equivital. Eqlifemonitor: World-Class Physiological Data Monitoring That Makes a Difference. Available online: https://equivital.com/products/eq02-lifemonitor (accessed on 22 August 2024).
  22. Axiamo. Axiamo PADIS 2.0 Long Term Activity Monitoring for Soldiers. Available online: https://www.axiamo.com/axiamo-padis-2-0/ (accessed on 12 March 2025).
  23. Garmin. Fenix 7 Standard/Solar/Pro Series Owner’s Manual-Battery Information. Available online: https://www8.garmin.com/manuals/webhelp/GUID-C001C335-A8EC-4A41-AB0E-BAC434259F92/EN-US/GUID-694C4E14-D875-479F-AFB1-2A6A582FF506.html (accessed on 22 August 2024).
  24. Garmin. Take Control of Garmin Wearables and Stream Real-Time Data. Available online: https://developer.garmin.com/health-sdk/overview/ (accessed on 12 March 2025).
  25. Timescale Inc. TimescaleDB: SQL Made Scalable for Time-Series Data. 2017. Available online: https://caxapa.ru/thumbs/880838/timescaledb.pdf (accessed on 12 June 2024).
  26. Fernández, A.P.; Leenders, C.; Aerts, J.M.; Berckmans, D. Emotional States versus Mental Heart Rate Component Monitored via Wearables. Appl. Sci. 2023, 13, 807. [Google Scholar] [CrossRef]
  27. Bourgois, J.G.; Bourgois, G.; Boone, J. Perspectives and Determinants for Training-Intensity Distribution in Elite Endurance Athletes. Int. J. Sport. Physiol. Perform. 2019, 14, 1151–1156. [Google Scholar] [CrossRef]
  28. Mullie, P. Impact of a Special Operations Forces Basic Training on Body Composition and Energy Expenditure. Mil. Med. 2024, 189, e1588–e1592. [Google Scholar] [CrossRef]
  29. Xuan, X.; Pérez-Ràfols, C.; Chen, C.; Cuartero, M.; Crespo, G.A. Lactate Biosensing for Reliable On-Body Sweat Analysis. ACS Sens. 2021, 6, 2763–2771. [Google Scholar] [CrossRef]
  30. Connor, K.M.; Davidson, J.R. Development of a new resilience scale: The Connor-Davidson resilience scale (CD-RISC). Depress. Anxiety 2003, 18, 76–82. [Google Scholar] [CrossRef]
  31. Crum, A.J.; Salovey, P.; Achor, S. Rethinking stress: The role of mindsets in determining the stress response. J. Personal. Soc. Psychol. 2013, 104, 716. [Google Scholar] [CrossRef]
  32. Oginska, H.; Mojsa-Kaja, J.; Mairesse, O. Chronotype description: In search of a solid subjective amplitude scale. Chronobiol. Int. 2017, 34, 1388–1400. [Google Scholar] [CrossRef] [PubMed]
  33. Duckworth, A.L.; Quinn, P.D. Development and validation of the Short Grit Scale (GRIT–S). J. Personal. Assess. 2009, 91, 166–174. [Google Scholar] [CrossRef] [PubMed]
  34. Niebuhr, D.W.; Gubata, M.E.; Oetting, A.A.; Weber, N.S.; Feng, X.; Cowan, D.N. Personality Assessment Questionnaire as a pre-accession screen for risk of mental disorders and early attrition in U. S. Army recruits. Psychol. Serv. 2013, 10, 378–385. [Google Scholar] [CrossRef] [PubMed]
  35. Mobley, W.; Hand, H.H.; Baker, R.M.L.; Meglino, B.M. Conceptual and empirical analysis of military recruit training attrition. J. Appl. Psychol. 1979, 64, 10–18. [Google Scholar] [CrossRef]
  36. Taylor, D.J.; Straud, C.; Hale, W.; Grieser, E.; Garb, L.; Garb, H. Sleep difficulties as a predictor of attrition in United States Air Force recruits. Sleep Health 2020, 6, 338–343. [Google Scholar] [CrossRef]
  37. Bulmer, S.; Aisbett, B.; Drain, J.R.; Roberts, S.; Gastin, P.B.; Tait, J.; Main, L.C. Sleep of recruits throughout basic military training and its relationships with stress, recovery, and fatigue. Int. Arch. Occup. Environ. Health 2022, 95, 1331–1342. [Google Scholar] [CrossRef]
  38. Van Der Donckt, J.; Van Der Donckt, J.; Deprost, E.; Van Hoecke, S. tsflex: Flexible time series processing & feature extraction. SoftwareX 2021. [Google Scholar] [CrossRef]
  39. Garmin. What is the Sleep Score and Insights Feature on My Garmin® Watch? | Garmin Customer Support. Available online: https://support.garmin.com/en-US/?faq=DWcdBazhr097VgqFufsTk8 (accessed on 12 June 2024).
  40. Garmin. What Is Advanced Sleep Monitoring in Garmin Connect? | Garmin Customer Support. Available online: https://support.garmin.com/en-US/?faq=mBRMf4ks7XAQ03qtsbI8J6 (accessed on 12 June 2024).
  41. Garmin. Garmin Technology: Sleep Tracking. Available online: https://www.garmin.com/en-US/garmin-technology/health-science/sleep-tracking/ (accessed on 12 June 2024).
  42. Hirshkowitz, M.; Whiton, K.; Albert, S.M.; Alessi, C.; Bruni, O.; DonCarlos, L.; Hazen, N.; Herman, J.; Adams Hillard, P.J.; Katz, E.S.; et al. National Sleep Foundation’s updated sleep duration recommendations: Final report. Sleep Health 2015, 1, 233–243. [Google Scholar] [CrossRef]
  43. Ohayon, M.; Wickwire, E.M.; Hirshkowitz, M.; Albert, S.M.; Avidan, A.; Daly, F.J.; Dauvilliers, Y.; Ferri, R.; Fung, C.; Gozal, D.; et al. National Sleep Foundation’s sleep quality recommendations: First report. Sleep Health 2017, 3, 6–19. [Google Scholar] [CrossRef]
  44. Chen, C.; Liaw, A.; Breiman, L. Using Random Forest to Learn Imbalanced Data. Department of Statistics, UC Berkeley, July 1, 2004. Available online: https://statistics.berkeley.edu/sites/default/files/tech-reports/666.pdf (accessed on 12 March 2025).
Figure 1. Schematic overview of the data synchronization architecture. Note that an online data store is not a hard requirement as this setup works fully offline as well.
Figure 1. Schematic overview of the data synchronization architecture. Note that an online data store is not a hard requirement as this setup works fully offline as well.
Sensors 25 01828 g001
Figure 2. The suitcases used to perform simultaneous charging and synchronization with the smartphones (attached to the suitcases).
Figure 2. The suitcases used to perform simultaneous charging and synchronization with the smartphones (attached to the suitcases).
Sensors 25 01828 g002
Figure 3. (a) PCA based on the answers of the baseline questionnaires (first 2 components). We notice that most of the recruits that underwent attrition are on the right side of the scatter plot. (b) Weights of the PCA analysis on the results of the mental baseline questionnaires from Table 3.
Figure 3. (a) PCA based on the answers of the baseline questionnaires (first 2 components). We notice that most of the recruits that underwent attrition are on the right side of the scatter plot. (b) Weights of the PCA analysis on the results of the mental baseline questionnaires from Table 3.
Sensors 25 01828 g003
Figure 4. LDA based on the answers of the baseline questionnaires. We notice that most of the recruits that underwent attrition are more frequent for LD1 > 0.
Figure 4. LDA based on the answers of the baseline questionnaires. We notice that most of the recruits that underwent attrition are more frequent for LD1 > 0.
Sensors 25 01828 g004
Figure 5. Graph of the cumulative explained variance of the principal components, based on the answers of the seven variables of the mental baseline questionnaire.
Figure 5. Graph of the cumulative explained variance of the principal components, based on the answers of the seven variables of the mental baseline questionnaire.
Sensors 25 01828 g005
Figure 6. Visualization of the self-reported sleep scores on the questionnaires vs. the score provided by the Garmin wearable. This relationship has a Spearman correlation coefficient of 0.51.
Figure 6. Visualization of the self-reported sleep scores on the questionnaires vs. the score provided by the Garmin wearable. This relationship has a Spearman correlation coefficient of 0.51.
Sensors 25 01828 g006
Figure 7. Projecting self-reported sleep scores to the two most discriminative directions using Linear Discriminant Analysis.
Figure 7. Projecting self-reported sleep scores to the two most discriminative directions using Linear Discriminant Analysis.
Sensors 25 01828 g007
Figure 8. Visualization of the score distributions of the wellness questionnaire entries.
Figure 8. Visualization of the score distributions of the wellness questionnaire entries.
Sensors 25 01828 g008
Table 1. Comparison between 5 consumer available wearables. Information taken from each respective product/owner brief.
Table 1. Comparison between 5 consumer available wearables. Information taken from each respective product/owner brief.
Garmin Fenix 7 [18]Fitbit Charge 5 [19]Movesense [20]eqLifemonitor [21]Axioma Padis [22]
Device type SmartwatchSmartwatchWearable sensorWearable sensorWearable sensor
Body position WristWristUpper arm/chestChestWrist, belt, backpack
PriceEUR 599.99EUR 150EUR 104EUR 1650EUR 800
Battery life18 days4 days16 days HR
7 days ECG
2 days7 days
Charging typeCableCableReplacable batteryCharging caseCharging case
Heart rateYesYesYesYesYes
Raw accel.Yes *NoYesYesYes
Raw HRVYes *Only during sleepYesYesNo
ECGNoExternal appYesYesNo
GPS supportYesYesNoYesNo
SynchronizationWiFi/ANT+/BLE/USBBLE/USBBLE BLE/USBBLE/USB
Cloud storageYesYesNoYes Yes
On-premiseYes *NoNoYesYes
* Through usage of the Garmin Health SDK. Can only offload in real time due to the very limited internal storage. Only available with an extra service charge.
Table 2. Summary of the two on-site test phases. The second test phase lasted longer and monitored more recruits. Attrition is classified into 4 categories: Drop-On-Request, Medical, Safety and Performance.
Table 2. Summary of the two on-site test phases. The second test phase lasted longer and monitored more recruits. Attrition is classified into 4 categories: Drop-On-Request, Medical, Safety and Performance.
(a)
Study Start date End date # recruits
Test phase 131 August 202327 October 202317
Test phase 230 January 202431 May 202446
(b)
ReasonTest phase 1Test phase 2
Drop-On-Request37
Medical04
Safety00
Performance00
Total311
Table 3. Questionnaires used during test phase 1 and 2. Most questionnaires focused on the mental state of the recruits.
Table 3. Questionnaires used during test phase 1 and 2. Most questionnaires focused on the mental state of the recruits.
VariableQuestionnaireReference
ResilienceCD-RISC-25[30]
Stress mindsetStress mindset measure[31]
SleepCHR-NL 1–SIF–SIC[32]
GritShort Grit Scale[33]
Motivation, self-reported healthCustom
injury, and sports history
Table 4. Overview of the reported sleep scores of the Garmin device. Calculated based on sleep times and sleep stages identified using a combination of heart rate, heart rate variability, pulse oximetry, respiration, and body movement data. Information from this table is directly sourced from different Garmin blogposts [39,40,41].
Table 4. Overview of the reported sleep scores of the Garmin device. Calculated based on sleep times and sleep stages identified using a combination of heart rate, heart rate variability, pulse oximetry, respiration, and body movement data. Information from this table is directly sourced from different Garmin blogposts [39,40,41].
MeasurementDescription
Sleep duration scoreHow long the recruit slept compared to globally accepted age-based recommendations [42].
Awakenings count scoreA high score corresponds to continuous sleep through the night, with few to no stretches of awake time.
Awake time scoreScore based on the total time spent awake during the recorded sleep interval.
Interruptions scoreScore based on the number of times you are awake for longer than 5 min.
Light sleep scoreScore based on time spent in the first stage of sleep. Eye movements and muscle activity slow during light sleep as your body gets ready for deep sleep.
Deep sleep scoreScore based on time spent in deep sleep stage. Eye and muscle movements stop completely. Your heart rate and breathing slow. This stage can be referred to as restoration mode, where the body will recover, building bone and muscle, and boosting your immune system.
REM sleep scoreScore based on time spent in REM sleep stage. Brain activity is almost as active as when you are awake.
Sleep quality scoreQuality aspects of the sleep score come from a combination of sleep architecture, stress data, interruptions during the night and other factors [43].
Sleep recovery scoreN/A, no Garmin provided description of this score available.
Sleep restlessness scoreThis feature indicates sudden movement, typically detected in light sleep.
Overall sleep scoreCalculated based on a blend of how long you slept, how well you slept, and evidence of recovery activity occurring in your autonomic nervous system derived from heart rate variability data. This score is calculated on a scale of 0–100 (Excellent: 90–100, Good: 80–89, Fair: 60–79, Poor: Below 60).
Table 5. Overview of the data partitioning into 5 unique folds.
Table 5. Overview of the data partitioning into 5 unique folds.
Score 1Score 2Score 3Score 4Score 5Recruits
Train Test Train Test Train Test Train Test Train Test Train Test
Split 13651521939547452101259123267
Split 229121234833011242213132854267
Split 33381264530413843411933448267
Split 43110131403667643811530577267
Split 535615219373874668730280285
Table 6. Residuals of each score over all splits. Macro MAE calculated by averaging all class results in a split. Classwise MAE calculated per score over all splits.
Table 6. Residuals of each score over all splits. Macro MAE calculated by averaging all class results in a split. Classwise MAE calculated per score over all splits.
Score 1Score 2Score 3Score 4Score 5Macro MAE
Split 10.000.531.001.011.210.75 ± 0.44
Split 20.670.940.920.760.610.79 ± 0.13
Split 30.751.020.930.790.480.79 ± 0.18
Split 40.500.900.870.740.730.75 ± 0.14
Split 50.170.900.870.760.530.64 ± 0.27
Classwise MAE0.42 ± 0.290.86 ± 0.170.92 ± 0.050.81 ± 0.10.71 ± 0.26
Table 7. Synchronization benchmark results for each round of the specified amount of concurrent watches (# Watches). Measured in seconds using the Standard SDK 4.2.3.
Table 7. Synchronization benchmark results for each round of the specified amount of concurrent watches (# Watches). Measured in seconds using the Standard SDK 4.2.3.
# WatchesRound 1Round 2Round 3Round 4Round 5Mean Agg.
Ind. Agg. Ind. Agg. Ind. Agg. Ind. Agg. Ind. Agg.
1109-128-156-125-123-128 ± 15
2318 ± 7324343 ± 15354217 ± 6234271 ± 12290276 ± 36302300 ± 40
3346 ± 52391459 ± 31492217 ± 69369414 ± 18454401 ± 41458425 ± 45
4476 ± 181631549 ± 31600434 ± 184579451 ± 162585521 ± 51598599 ± 18
5514 ± 216730528 ± 200719557 ± 151701579 ± 290818615 ± 161755744 ± 41
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Decorte, R.; Vanhaeverbeke, J.; VanDen Berghe, S.; Slembrouck, M.; Verstockt, S. Continuous Monitoring of Recruits During Military Basic Training to Mitigate Attrition. Sensors 2025, 25, 1828. https://doi.org/10.3390/s25061828

AMA Style

Decorte R, Vanhaeverbeke J, VanDen Berghe S, Slembrouck M, Verstockt S. Continuous Monitoring of Recruits During Military Basic Training to Mitigate Attrition. Sensors. 2025; 25(6):1828. https://doi.org/10.3390/s25061828

Chicago/Turabian Style

Decorte, Robbe, Jelle Vanhaeverbeke, Sarah VanDen Berghe, Maarten Slembrouck, and Steven Verstockt. 2025. "Continuous Monitoring of Recruits During Military Basic Training to Mitigate Attrition" Sensors 25, no. 6: 1828. https://doi.org/10.3390/s25061828

APA Style

Decorte, R., Vanhaeverbeke, J., VanDen Berghe, S., Slembrouck, M., & Verstockt, S. (2025). Continuous Monitoring of Recruits During Military Basic Training to Mitigate Attrition. Sensors, 25(6), 1828. https://doi.org/10.3390/s25061828

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop