Abstract
Background: Advances in data science and technology have transformed lifestyle research by enabling the integration of multimodal information and the generation of large-scale datasets. Despite the growing interest in machine learning (ML) within health behavior research, significant methodological gaps remain.
Objective: The study aims to systematically review the applications of supervised ML algorithms in the analysis of healthy lifestyle data, with a particular focus on the methodological approaches used. The specific objectives are to explore the types and sources of data used for health outcomes, examine the ML processes used, including explainable artificial intelligence (XAI) methods, and review the software tools used. Additionally, this review aims to provide practical guidelines to enhance the quality and transparency of future ML research in health.
Methods: Following the PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews) recommendations, the search was conducted across PubMed, PsycINFO, and Web of Science, yielding 65 studies that met the inclusion criteria.
Results: Most studies (48/65, 74%) integrated multidomain data from physical activity, diet, sleep, and stress. Data sources were split between self-acquired data (33/65, 51%) and health repositories (32/65, 49%). Single-item measurements were common, particularly for physical activity, diet, and sleep. Although 40 of 65 studies used a multimodel approach, random forest was the most frequently applied algorithm. To improve explainability, 22 of 65 (33.84%) studies incorporated specific XAI methods, with 21 using Shapley Additive Explanation values and 1 using local interpretable model-agnostic explanations. R (R Core Team) and Python (Python Software Foundation) were the most widely used software tools, with variation in the libraries used.
Conclusions: This review highlights methodological gaps in the application of supervised ML to healthy lifestyle data. The ML workflow should span from data acquisition to explainability, using iterative steps to improve methodological rigor. Although multidomain data collection enhances the understanding of health issues related to lifestyle, representativeness remains limited due to methodological shortcomings in data acquisition. While random forest was the most commonly used algorithm, a multimodel approach is recommended for a comprehensive comparison. Lifestyle components consistently ranked among the top features in studies integrating XAI. Incorporating XAI methods into the ML pipeline can support personalized interventions, provided data collection is accurate. The R metapackage (tidymodels; Max Kuhn and Hadley Wickham) facilitates process evaluation through unified syntax, improving replicability. Methodological and reporting guidelines and a checklist are provided to enhance transparency and replicability in multidisciplinary ML research.
International Registered Report Identifier (IRRID): RR2-10.37766/inplasy2023.3.0065
doi:10.2196/78648
Keywords
Introduction
There is a growing interest in understanding the effects of synergistic relationships among lifestyle behaviors and their effect on health outcomes [,]. Traditionally, healthy lifestyle (HL) research has primarily focused on physical activity and diet. However, recent studies increasingly include sleep and stress management as critical components of lifestyle [,]. For instance, stress has been shown to negatively influence physical activity, sleep, and dietary habits [], which in turn have an overall impact on health and well-being. This multidimensional perspective has gained attention in public health under the concept of lifestyle medicine, which incorporates physical activity, diet, sleep, and stress management as cost-effective interventions to prevent noncommunicable diseases, such as cardiovascular and metabolic diseases [-].
Technological advances, including wearable devices and lifelogging processes, have significantly enhanced the capability to collect multimodal, high-frequency, and ecological lifestyle data [,]. This wealth of data provides valuable contextual information and insights for researchers and users []. However, the vast amount and complexity of behavioral and physiological data expose significant analytical challenges. Traditional statistical models often struggle with the high dimensionality, heterogeneity, and nonlinearity typical of lifestyle studies. Recent progress in computational power and artificial intelligence (AI), particularly machine learning (ML), has contributed to addressing these limitations [].
ML models are capable of analyzing complex data types and generating insights and knowledge to improve decision-making [,]. Furthermore, ML algorithms can flexibly handle nonlinear relationships among features and outcomes. While the boundary between classical statistics and ML is not clear, ML algorithms are recognized for their flexible data-driven approach, avoiding the imposition of a predetermined relational structure between variables [-]. Additionally, prioritizing algorithms that maximize generalizability to new data, often referred to as scalability in the big data context, is crucial to face new health challenges [,]. These characteristics make ML analysis a suitable methodology for predictive modeling and feature extraction in health-related lifestyle research.
ML models are broadly classified into supervised learning (SL) and unsupervised learning (UL). In SL, the model is trained with labeled data, where each observation has an associated response measurement, to predict known outcomes such as disease risk or behavioral adherence []. The goal of SL is to fit a model that can predict the response when applied to new data. When the response value is continuous, this is known as a “regression problem”; when the response is categorical, it is known as a “classification problem.” In contrast, in UL models, the goal is to discover patterns rather than predict outcomes, since there is no associated response to the input, and the model seeks relationships and similarities between observations. In the health domain, where diagnosis and detection are key focuses, SL, and particularly classification tasks, are more prevalent due to their ability to evaluate these predictions [,]. Clinical applications of SL include triage systems, prognosis prediction, and disease classification using rapid testing []. Consequently, SL methods are standard in epidemiology to enhance clinical decisions based on input-output relationships []. Since prediction and explainability are central concerns in health research, this scoping review focuses specifically on SL methods.
Despite the growing attention to ML in health behavior research, there remain significant methodological gaps. Prior reviews have focused primarily on outcome effectiveness or AI chatbot interventions, often providing limited detail about the ML process involved []. A recent scoping review on ML methods used in health promotion and behavioral change found that the main interventions studied are those related to physical activity, while other crucial aspects of HL were overlooked, revealing an imbalance in this literature []. Similarly, Lai et al [] reviewed the applications of large language models in exercise recommendations and physical activity, highlighting methodological limitations associated with these AI models. In sum, these studies underscore the need for a more comprehensive review to include a holistic concept of HL. Furthermore, methodological details such as data preprocessing, model evaluation, and explainability are often underreported, hindering transparency, reproducibility, and interdisciplinary collaboration.
To address the lack of explainability, explainable artificial intelligence (XAI) has emerged, which focuses on understanding AI algorithms and making them more transparent. XAI aims to provide human-understandable explanations for the decisions made by ML models []. In HL research, XAI can be directed to identify the set of behaviors that significantly influence health, thereby enhancing transparency and trust in AI. It is important to distinguish between interpretability and explainability in the AI context. While interpretability refers to understanding the influence of each feature in the original model, explainability involves deriving actionable human insights from the model’s predictions []. Interpretability enables AI developers to delve into the model’s decision-making to comprehend how algorithms reach their decisions. Conversely, explainability refers to the process for creating common meaning from model decisions and therefore provides human-readable explanations []. Therefore, reporting the explainability method used in ML projects is crucial not only to enhance the decision-making process of the end user but also to understand how lifestyle factors interact with health outcomes.
Therefore, this study aims to systematically review the applications of supervised ML algorithms in analyzing HL data, with a specific focus on the methodological aspects used in these studies, rather than their results. The specific objectives are to explore (1) the specific lifestyle data used in health outcome studies; (2) the sources and types of data subjected to analysis; (3) the characteristics of the ML models, including XAI methods; and (4) the programs and libraries used for ML implementation. Additionally, based on the findings of this scoping review, we aim to provide practical guidelines to enhance the quality and transparency of future ML research in lifestyle science. A scoping review is the type of systematized review (ie, systematic, transparent, and replicable) most appropriate for addressing these objectives [].
Method
Overview
To maximize the reporting quality of this scoping review, we followed the PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews) recommendations [] (checklist provided in ). The protocol for this scoping review was registered with the International Platform of Registered Systematic Review and Meta-Analysis Protocols (INPLASY) []. All data generated in this review are provided in and are accessible in the institutional repository [].
Search Strategy
In this scoping review, we searched for primary studies in the 3 principal health databases: PubMed (National Center for Biotechnology Information), PsycINFO (ProQuest), and Web of Science (Clarivate). The search was restricted to medical and psychological databases to capture studies directly relevant to health outcomes. Consequently, studies primarily published in engineering or computer science, which may focus on algorithm development or sensor-based data processing, were not included. The search strategy followed the PRESS (Peer Review of Electronic Search Strategies) [] and PRISMA-S (Preferred Reporting Items for Systematic Reviews and Meta-Analyses–Search extension) guidelines [] and consisted of 2 groups of search terms referring to (1) HL and (2) ML. We also added a third group of terms preceded by the Boolean operator “NOT” to improve the specificity of the search strategy.
This scoping review adopts a health-focused perspective, in which HL is treated as a multidimensional construct rather than the sum of isolated behaviors [,,]. Therefore, the umbrella term HL was combined using the operator “OR” with an interaction block including (1) physical activity, (2) diet, (3) sleep, and (4) stress. This block is aligned with the multiple health behavior change and lifestyle medicine frameworks, in which the interaction between behaviors is a central construct [,].
The search strategy was adapted to the specific syntax of each database (Table S1 in ). The search was conducted on October 10, 2025, with language restrictions (English and Spanish) but without limitations on publication years.
Study Selection
Inclusion and Exclusion Criteria
Studies were included in or excluded from the review according to the following criteria provided in .
Inclusion criteria:
- Used supervised machine learning (ML) models for analyzing lifestyle data.
- Analyzed lifestyle behaviors as either inputs or outputs of the ML models.
- Used data from real individuals (not simulations).
- Published in English or Spanish.
Exclusion criteria:
- Focused on unsupervised learning (UL) without connection to supervised learning (SL) modeling.
- Focused on mathematical formulation or guidelines for implementing ML models in health.
- Used simulated data or aimed to develop a chatbot or app based on ML.
- Primarily addressed substance abuse, such as alcohol intake or smoking cessation.
- Focused exclusively on classical statistical regression algorithms, such as linear or logistic regression, which were not considered ML on their own in this review.
Justification of Exclusions
UL algorithms were excluded because they do not have an associated response to inputs, thereby lacking performance evaluation. Classical statistical regression algorithms, such as linear or logistic regression, were not considered in this review. While the boundary between classical statistics and ML is not clear, ML algorithms are recognized for their flexible data-driven approach, avoiding the imposition of a predetermined relational structure between variables [-]. Additionally, prioritizing algorithms that maximize generalizability to new data, often referred to as scalability in the big data context, is crucial to address new health challenges [,]. Consequently, studies focusing exclusively on this type of statistical algorithm were excluded. However, we acknowledge that the use of a model ensemble approach allows for the inclusion of these statistical algorithms to assess the performance of different algorithms during the evaluation step. Studies on substance abuse disorders were excluded as they involve distinct behavioral and neurobiological mechanisms that differ substantially from the domains of physical activity, diet, sleep, and stress, which are the core components of HL behaviors as defined in this review. In addition, substance abuse is categorized within the risk avoidance cluster, which is conceptually distinct from the other 4 behaviors examined in this review []. This distinction is well established in multiple health behavior theories, which differentiate behaviors that enhance health from those that reduce risk through avoidance. Therefore, its exclusion preserves the applicability of results to a multiple health behavior framework, as the selected behaviors are interrelated through shared psychological resources [,].
Two reviewers (TE and CA) independently screened titles and abstracts in the first phase and full texts in the second phase. Discrepancies were resolved by consensus, with the participation of a third reviewer (JML) when necessary. Agreement between reviewers during the selection process was analyzed by calculating Cohen κ.
Data Management
Mendeley was used as reference management software; the results of the search strategy were entered, and duplicates were merged or removed. An ad hoc checklist was used to extract information from the included papers. The checklist was divided into 5 sections:
- General information: authors, title, year, and country of affiliation.
- Methodological data: type of study, aim, year of data collection, form of data acquisition, sample, and countries represented in the data.
- Study variables: health issues, lifestyle features, and the model’s input and output variables.
- Software: statistical programming language, libraries, and packages.
- Model aspects: type of problem, stages of ML analysis, ML methods, model evaluation, evaluation metrics, and XAI methods.
Strategy of Data Synthesis
The review was presented as a narrative synthesis, and the information was summarized in tables and figures. The information extracted from the studies was divided into 3 blocks: type of data, ML process, and software. For data extraction, we focused on lifestyle components, health outcomes, data sources, acquisition methods, and data typology. Regarding the ML process, we focused on the whole process, consisting of preprocessing, modeling, validation, evaluation, and XAI methods. For model evaluation, only the procedures used for final performance assessment were extracted. When studies reported cross-validation, we classified it as the final performance estimation method unless authors explicitly stated its use for hyperparameter optimization. To identify the top-ranked lifestyle components in the XAI analysis, we systematically examined the figures and tables reported in each study. In this review, the term ranked refers to features that were highlighted by the XAI algorithm. In contrast, unranked indicates that appeared in the model but were not reported in the XAI visualization, while not available denotes features that were not included in the ML model. Software used in each study was also recorded.
Results
Overview
A total of 2249 papers were retrieved from the databases, and 65 studies met the eligibility criteria and were included in this scoping review (refer to ). There was very good agreement between reviewers during the selection process: 96% (n=52; κ=0.84, 95% CI 0.61-1.0) in the title and abstract screening and 94% (n=35; κ=0.88, 95% CI 0.71-1.0) in the full-text screening.

From this point forward, the “Results” section is structured following the ML workflow depicted in , which illustrates the 5 key steps in the ML pipeline. The process begins with data acquisition, followed by preprocessing to prepare the data. Then, SL algorithms are applied and evaluated to determine their effectiveness. Finally, explainability techniques are used to understand the models. The dashed lines indicate that modeling, evaluation, and explainability can improve earlier stages, making the process iterative. Each stage of the process corresponds to a subsection. Finally, we examined the software used throughout the entire process in the included studies.

Data Acquisition: Collection Modes, Data Typology, Lifestyle Variables, and Health Outcomes
The 65 papers included in this review were published between 2004 and 2025, with 57 (87.7%) published since 2019. shows the annual productivity output stratified by lifestyle components. The studies were carried out in several geographical regions across 4 continents (Table S2 and Figure S1 in ). The mean sample size was 29,905.14 participants, with the smallest study including 8 participants and the largest including 470,778 participants.

Four lifestyle domains were identified in the review: physical activity, diet, sleep, and stress. Most studies (48/65, 74%) integrated data from more than 1 lifestyle domain (refer to ). The most studied component was physical activity, featured in 61 of 65 studies. Diet and sleep appeared in 33 of 65 (51%) and 34 of 65 (52%) studies, respectively, while stress appeared in 15 of 65 (23%) studies (refer to ). Only 17 of 65 (26%) studies focused exclusively on a single domain (15 on physical activity [-] and 2 on sleep [,]).
Regarding data sources, 33 of 65 (51%) studies relied on self-acquired data, while 32 of 65 (49%) studies used either private or public health datasets, such as UK Biobank []. Among the studies using self-acquired data, the mean sample size was 10,503.41 participants. Six studies focused exclusively on females [,,-], and 27 studies reported a female proportion ranging from 41% to 89.4%. Studies using health repositories exhibited a greater mean sample size of 80,406.38 participants. Nineteen studies reported a female proportion ranging from 41.33% to 70%, 12 did not report sex distribution, and 1 study focused exclusively on females.
summarizes the different data collection modes used for each of the 4 lifestyle domains identified (Table S3 in provides a detailed description of each measure). Single-item measures were used in 40 of 65 (61.5%) studies assessing physical activity, 18 of 65 (27.7%) studies assessing diet, and 22 of 65 (33.8%) studies assessing sleep. Despite the heterogeneity of these items, distinct categories emerged within each lifestyle domain.
For physical activity, the emerging categories included engagement in physical activities [,-,,,-], intensity [,,,,-], frequency [,,,-], and environmental factors [,]. Within the diet domain, categories included frequency of consumption [,,,,], types of products [,,,], environmental factors [], and consumption habits [,,,,,,,]. Regarding sleep, categories included sleep duration [,,,-,,,-,,,,,], perceived sleep quality [,], and sleep-related problems [,]. Finally, within the stress domain, the emerging category focused primarily on stress level [,,,,,,].
Questionnaires were used to collect physical activity data in 13 studies. The standardized questionnaires included the Global Physical Activity Questionnaire (GPAQ) [,,], the International Physical Activity Questionnaire (IPAQ) [], the Lifetime Total Physical Activity Questionnaire (exercise and sport subscale) [], the Nutritional and Social Healthy Habits (NutSo-HH) scale [], the Indian Migration Study Physical Activity Questionnaire (IMS-PAQ) [], physical fitness test [], the Short Questionnaire to Assess Health-Enhancing Physical Activity (SQUASH) [,], the physical activity scale from the Active Living Index [], the Pregnancy Physical Activity Questionnaire [], and the Physical Activity Scale for the Elderly []. For diet assessment, 12 studies used the Food Frequency Questionnaire (FFQ) [,,,,,,,], the PrimeScreen questionnaire [], the Mini Nutritional Assessment [], the NutSo-HH scale [], and a nonstandardized questionnaire consisting of items from different questionnaires []. The standardized questionnaires used to measure sleep were the Pittsburgh Sleep Quality Index [,,,,,,,], the NutSo-HH scale [], the Munich Chronotype questionnaire and Sleep Disturbance Scale for Children [], and the Epworth Sleepiness Scale []. To measure stress, the stress subscale of the Depression Anxiety Stress Scale (DASS) [,], the INTERHEART stress questionnaire [], the Psychosocial Well-being Index-Short Form [], the Perceived Stress Scale [,,], and the Profile of Mood States [] were used. Regarding data collection through sensors, most studies used wearable devices. One study used a smartphone to obtain points of interest related to physical activity and diet [], and 1 sleep study used polysomnography []. Finally, 2 studies used words related to physical activity and diet, 1 derived from Google Trends [] and the other from Twitter (Twitter, Inc) [].
Concerning the modeled inputs, 56 of 65 (86.1%) studies used multimodal data. The input modalities were lifestyle (60/65, 92.3%), sociodemographic (49/65, 75.3%), clinical (29/65, 44.6%), anthropometric (14/65, 21.5%), psychological (20/65, 30.7%), physical (3/65, 4.62%), environmental (10/65, 15.38%), physiological (3/65, 4.62%), and behavioral (2/65, 3.07%). The model outcomes included lifestyle domains in 14 (22% studies; 5 physical activity [,-,], 5 sleep [,,,,], 3 diet [,,], and 1 stress []) and other health outcomes in 51 (75% studies; with mental health, cancer, cardiovascular diseases, and diabetes being the most frequent categories; refer to ).
Cross-sectional data were acquired in 36 (55%) studies [-,-,-,-,,,-,-,,,,,,,], longitudinal data in 18 (28%) [,,,,,,,-,,,] studies, time-series data in 7 (11%) [,,,,,,] studies, combined longitudinal and time-series data in 1 study [], textual data in 2 studies [,], and combined cross-sectional and geographical data in 1 study [].
| Study | Physical activity | Diet | Sleep | Stress | Health outcome | Preprocess | ML algorithm | Model evaluation | XAI |
| Abdul Rahman et al [] | Questionnaire (standardized) and single items (frequency) | Single items (frequency) | Single items (sleep hours) | — | Mental health | Missing imputation, Resampling, and dimensionality reduction | RF, ANN, NB, and KNN | Hold-out test set | — |
| Afrash et al [] | Single items (engagement) | Single items (frequency) | — | — | Cancer | Transformation, missing imputation, and dimensionality reduction | DT, MLPNN, RBFNN, FNN, PNN, and KNN | 10-fold cross-validation (final evaluation) | — |
| Ai et al [] | Questionnaire (standardized) and single items (frequency) | Questionnaire (nonstandardized) | — | Questionnaire (standardized) | Alzheimer disease | Transformation, missing imputation, and dimensionality reduction | RF and SVM | Nested cross-validation | — |
| Allen [] | Single items (environment) | Single items (environment) | Single items (sleep hours) | — | Obesity | Missing imputation and dimensionality reduction | RF and DT | 2-fold cross-validation (final evaluation) | LIME |
| Alshuraf et al [] | Sensor (wearable) | Single items (habits) | — | Questionnaire (standardized) | Cardiovascular disease | Transformation, missing imputation, and dimensionality reduction | RF, DT, KNN, and NB | Leave-one-out cross-validation (LOOCV) | — |
| Birk et al [] | Questionnaire (standardized) | Questionnaire (standardized) | — | — | Diabetes | Resampling and dimensionality reduction | RF | Hold-out test set | — |
| Bôto et al [] | Single items (engagement) | Single items (habits) | Single items (sleep hours) | — | Lifestyle (diet) | Transformation and dimensionality reduction | DT | Not reported | — |
| Butkevičiūtė et al [] | Sensor (wearable) | — | — | — | Lifestyle (physical activity) | Transformation | RF | 5-fold cross-validation (final evaluation) | — |
| Cai et al [] | Questionnaire (standardized) | — | — | — | Successful aging | Transformation and dimensionality reduction | RF, GBM, and ANN | 10-fold cross-validation (final evaluation) | — |
| Cheung et al [] | Sensor (wearable) | — | — | Single items (stress level) | Lifestyle (physical activity) | Dimensionality reduction | RF and DT | Not reported | — |
| Chiang and Dey [] | Sensor (wearable) | — | Sensor (wearable) | — | Blood pressure | Transformation, missing imputation, and dimensionality reduction | RF, GBM, MLPNN, LSTM-RNN, and SVM | 5-fold cross-validation (final evaluation) and online weighted-resampling | — |
| Cortés-Ibañez et al [] | Questionnaire (standardized) and single items (frequency) | Questionnaire (standardized) | — | — | Cancer | Transformation, missing imputation, resampling, and dimensionality reduction | RF and SVM | Hold-out test set | — |
| Cortés-Ibañez et al [] | Questionnaire (standardized) and single items (frequency) | Questionnaire (standardized) | — | — | Cancer | Transformation, resampling, and dimensionality reduction | RF, GBM, and SVM | 5-fold cross-validation (final evaluation) | — |
| Dianati-Nasab et al [] | Single items (engagement) | — | — | — | Cancer | Missing imputation and dimensionality reduction | RF, DT, XGBoost, and ANN | 10-fold cross-validation (final evaluation) | — |
| Faruqui et al [] | Single items (frequency) | Single items (type of products) | — | — | Diabetes | Transformation and missing imputation | LSTM-RNN, ANN, and KNN | Hold-out test set | — |
| Gu et al [] | Single items (frequency) | Single items (frequency) | Single items (sleep hours) | — | Infertility risk in women | Dimensionality reduction | RF, DT, BoostDT, LightGBM, and AdaBoost | Hold-out test set | SHAP values |
| Guthrie et al [] | Single items (frequency) | Single items (type of products) | — | — | Cardiometabolic disease | — | RF | Leave-one-out cross-validation (LOOCV) | SHAP values |
| Hu et al [] | Single items (frequency) | — | Single items (sleep hours) | — | Cardiovascular disease | Missing imputation and dimensionality reduction | RF and BART | Not reported | — |
| Hu et al [] | Single items (frequency) | — | Single items (sleep hours) | — | Cardiovascular disease | Missing imputation and dimensionality reduction | RF, XGBoost, and BART | 5-fold cross-validation (final evaluation) | — |
| Huang et al [] | Single items (engagement) | — | Single items (sleep quality) | — | Cognitive function | Missing imputation and resampling | RF, BoostDT, XGBoost, and LSTM-RNN | Hold-out test set | SHAP values |
| Jin and Halili [] | Single items (intensity) | — | Single items (sleep hours) | — | Mental health | Transformation, missing imputation, resampling, and dimensionality reduction | RF, DT, XGBoost, LightGBM, CatBoost, Bagging, HistGBM, SVM, and MLPNN | Hold-out test set | SHAP values |
| Kim et al [] | Questionnaire (standardized) | Questionnaire (standardized) | Questionnaire (standardized) and Single items (Sleep hours) | Questionnaire (standardized) | Quality of life | Transformation, Resampling | RF, DT, XGBoost, SVM, NB, KNN | 6-fold cross-validation (final evaluation) | SHAP values |
| Kimura et al [] | Sensor (wearable) | — | Sensor (wearable) | — | Alzheimer disease | Dimensionality reduction | SVM | 5-fold cross-validation (final evaluation) | — |
| Kiss et al [] | Single items (engagement) | — | Questionnaire (standardized) and single items (sleep hours) | Questionnaire (standardized) | Mental health | Transformation, missing imputation, and dimensionality reduction | XGBoost | Nested cross-validation | SHAP values |
| Li and Song [] | Single items (engagement) | — | — | — | Cognitive function | Transformation | CNN, Transformer, LSTM-RNN, GRU-Attention, WaveNet, and RNN | 10-fold cross-validation (final evaluation) | SHAP values |
| Lim et al [] | Single items (intensity) | — | — | — | Osteoarthritis | Transformation, missing imputation, resampling, and dimensionality reduction | FFNN | Hold-out test set | — |
| Lim et al [] | Single items (intensity) | — | Questionnaire (standardized) and single items (sleep quality) | Single items (Stress level) | Lifestyle (sleep) | Transformation, missing imputation, and dimensionality reduction | RF and DT | Hold-out test set | — |
| Lin et al [] | Single items (engagement) | — | — | — | Loneliness | Transformation, Missing imputation, Dimensionality reduction | RF, DT, SVM, MLP, and KNN | 10-fold cross-validation (final evaluation) | SHAP values |
| Liu et al [] | Questionnaire (standardized) | — | — | — | Cardiovascular disease | Transformation | RSF | Hold-out test set | — |
| Luo et al [] | Single items (frequency) | — | Single items (sleep hours) | Single items (stress level) | Social network addiction risk | Transformation and dimensionality reduction | RF | Hold-out test set | — |
| Luo et al [] | Single items (intensity) | Questionnaire (standardized) | Single items (sleep hours) | Single items (stress level) | Chronic kidney disease | Missing imputation and dimensionality reduction | GBM | Hold-out test set | — |
| Luo et al [] | Single items (intensity) | Single items (habits) | Single items (sleep hours) | — | Frailty | Missing imputation | XGBoost | 10-fold cross-validation (final evaluation) | SHAP values |
| Majcherek et al [] | Single items (frequency) | Single items (type of products) | Single items (sleep problems) | — | Mental health | Missing imputation | XGBoost | Not reported | SHAP values |
| Majcherek et al [] | Single items (engagement) | Single items (habits) | — | — | Diabetes | Resampling | RF, DT, AdaBoost, CatBoost, HistGBM, LightGBM, XGBoost, KNN, NB, and Nearest Centroid | Hold-out test set | SHAP values |
| Matta et al [] | Sensor (wearable) | — | — | — | Lifestyle (sleep) | Transformation | MLP | Hold-out test set | — |
| Moon and Woo [] | — | Single items (habits) | Single items (sleep quality) | Single items (stress level) | Mental health | Transformation, missing imputation, resampling, and dimensionality reduction | RF and ANN | Not reported | — |
| Morris et al [] | Questionnaire (standardized) and single items (environment) | Questionnaire (standardized) and single items (environment) | — | Questionnaire (standardized) | Cardiovascular disease | Missing imputation | RF and ANN | 10-fold cross-validation (final evaluation) | SHAP values |
| Mousavi et al [] | Single items (engagement) | Single items (frequency) | Single items (sleep hours) | — | Lifestyle (diet) | Dimensionality reduction | FFNN | Hold-out test set | — |
| Mun and Geng [] | Single items (frequency) | — | Single items (sleep hours) | Questionnaire (standardized) and sensor (wearable) | Fatigue | Transformation, missing imputation, and dimensionality reduction | RF | 10-fold cross-validation (final evaluation) | — |
| Nichols et al [] | Questionnaire (standardized) | Questionnaire (standardized) | Questionnaire (standardized) | Questionnaire (standardized) | Lifestyle (stress) | Transformation, missing imputation, resampling, and dimensionality reduction | SVM | Hold-out test set | — |
| Oladeji et al [] | Words (Google Trends) | Words (Google Trends) | — | — | Obesity | Dimensionality reduction | RF, GBM, and SVM | Out-of-sample | — |
| Park et al [] | Single items (intensity) | — | — | — | Adverse health event | Dimensionality reduction | XGBoost | Hold-out test set | — |
| Park and Edington [] | Single items (frequency) | Single items (type of products) | Single items (sleep hours) | Single items (stress level) | Diabetes | Missing imputation and resampling | MLPNN | Hold-out test set | — |
| Park [] | Single items (frequency) | Questionnaire (standardized) | — | — | Visceral fat | Transformation, missing imputation, and dimensionality reduction | RF, XGBoost, and ANN | Hold-out test set | SHAP values |
| Pereira et al [] | Questionnaire (standardized) | Questionnaire (standardized) | Questionnaire (standardized) | Questionnaire (standardized) | Mental health | Transformation and missing imputation | RF, XGBoost, and SVM | 10-fold cross-validation (final evaluation) | — |
| Puterman et al [] | Single items (intensity) | — | Single items (sleep problems) | — | Mortality | Transformation, missing imputation, and dimensionality reduction | RSF | Hold-out test set | — |
| Qasrawi et al [] | Single items (intensity) | Single items (habits) | Single items (sleep hours) | — | Mental health | Missing imputation and dimensionality reduction | RF, DT, XGBoost, SVM, ANN, KNN | 10-fold cross-validation (final evaluation) | — |
| Recenti et al [] | Single items (frequency) | — | — | — | Lifestyle (physical activity) | Missing imputation and resampling | RF, GBM, and AdaBoost | 10-fold cross-validation (final evaluation) | — |
| Recenti et al [] | Single items (frequency) | — | — | — | Diabetes | Missing imputation and resampling | RF, GBM, and AdaBoost | 10-fold cross-validation (final evaluation) | — |
| Ren et al [] | Single items (engagement) | — | — | — | Cognitive function | Missing imputation, Resampling | RF, XGBoost, SVM | Hold-out test set | SHAP values |
| Ruiz et al [] | Single items (engagement) | — | — | — | Depression | Not reported | DT | Not reported | — |
| Sandri et al [] | Questionnaire (standardized) | Questionnaire (standardized) | Questionnaire (standardized) | — | Lifestyle (diet) | Transformation and resampling | RF, DT, XGBoost, CatBoost, HistGBM, and FFNN | Hold-out test set | SHAP values |
| Sathyanarayana et al [] | Sensor (wearable) | — | Sensor (wearable) | — | Lifestyle (sleep) | Missing imputation | MLPNN, CNN, SETRNN, and LSTM-RNN | Hold-out test set | — |
| Shi et al [] | Single items (frequency) | Single items (frequency) | Single items (sleep hours) | — | Osteoporosis | Transformation, missing imputation, resampling, and dimensionality reduction | RF, DT, SVM, and KNN | Hold-out test set | SHAP values |
| Staudenmayer et al [] | Sensor (wearable) | — | — | — | Lifestyle (physical activity) | Transformation | RF, DT, ANN, and SVM | Leave-one-out cross-validation (LOOCV) | — |
| Stemmer et al [] | Words (Twitter) | Words (Twitter) | — | — | Inflammatory bowel disease | — | RF, GBM, AdaBoost, and SVM | Hold-out test set | — |
| Su et al [] | — | — | Questionnaire (standardized) | — | Resilience | Dimensionality reduction | RF, DT, and XGBoost | Not reported | SHAP values |
| Wallace et al [] | Questionnaire (standardized) | — | Questionnaire (standardized) and single items (sleep hours) | — | Mortality | — | RSF | Not reported | — |
| Wallace et al [] | — | — | Single items (sleep hours) and sensor (polysomnography) | — | Mortality | Missing imputation and dimensionality reduction | RF | External dataset | — |
| Wang et al [] | — | Questionnaire (standardized) | Questionnaire (standardized) | — | Lifestyle (sleep) | Missing imputation, resampling, and dimensionality reduction | GBM, LightGBM, SVM, MLPNN, and KNN | 10-fold cross-validation (final evaluation) | SHAP values |
| Xin and Ren [] | Single items (frequency) | — | Single items (Sleep hours) | — | Mental health | Dimensionality reduction | RF | Hold-out test set | SHAP values |
| Zhang et al [] | Single items (frequency) | Single items (habits) | Questionnaire (standardized) | Single items (stress level) | Lifestyle (sleep) | Resampling | RF, DT, XGBoost, SVM, ANN, and KNN | External dataset | SHAP values |
| Zhou et al [] | Sensor (wearable) | — | — | — | Lifestyle (physical activity) | Transformation | SVM | Out-of-sample | — |
| Zhou et al [] | Sensor (phone) | Sensor (phone) | — | — | Obesity | Missing imputation and dimensionality reduction | RF, GRF, and ANN | 10-fold cross-validation (final evaluation) | — |
| Zhou et al [] | Single items (intensity) | Single items (habits) | — | — | Psoriasis | Resampling | XGBoost | Hold-out test set | SHAP values |
aML: machine learning.
bXAI: explainable artificial intelligence.
cRF: random forest.
dANN: artificial neural network.
eNB: naive Bayes.
fKNN: k-nearest neighbor.
gDT: decision tree.
hMLPNN: multilayer perceptron neural network.
iRBFNN: radial basis function neural network.
jFNN: fuzzy neural network.
kPNN: probabilistic neural network.
lSVM: support vector machine.
mLIME: local interpretable model-agnostic explanations.
nGBM: gradient boosting.
oLST-RNN: long short-term memory recurrent neural network.
pXGBoost: extreme gradient boosting.
qBoostDT: boost decision tree.
rLightGBM: light gradient boosting machine.
sAdaBoost: adaptive boosting.
tSHAP: Shapley Additive Explanations.
uBART: Bayesian additive regression trees.
vCatBoost:categorical boosting.
wBagging: bootstrap aggregating.
xHistGBM: histogram-based gradient boosting machine.
yCNN: convolutional neural network.
zGRU-Attention: gated recurrent unit with attention.
aaRNN: recurrent neural network.
abFFNN: feed-forward neural network.
acRSF: random survival forest.
adSETRNN: simple Elman-type recurrent neural network.
aeGRF: generalized random forest.

Preprocessing
The preprocessing phase was divided into variable transformation, missing imputation, resampling, and dimensionality reduction. At least one of these preprocessing phases was reported by 59 (90.76%) studies (refer to and Table S4 in for more details).
There were 26 (40%) studies that reported normalization or other arithmetic or statistical transformations of variables before the modeling phase [,,,,,,,,,,,,,-,,,,,,-,]. Six (9.23%) studies recoded categorical variables into quantitative variables, 5 used one-hot encoding [,,,,], and 1 used principal component analysis with quantile transformer scaler [].
Missing data imputation was reported in 37 (56.92%) studies. Twelve papers simply removed cases with missing data [,-,,,,,,,,], while others applied cutoff percentages for missing values (eg, 10% [], 30 % [], or >50% [,]), and 1 study removed observations with missing values in the output []. Techniques included single imputation (mean, median, or mode) [,,,,,,,], multiple imputation by chained equations [,,,], k-nearest neighbor [,], regression-based algorithm [,], random forest (RF)–based multiple imputation [,,], the MissForest algorithm [,], imputation based on peers with similar health profile group [], imputation using training data [], and replacement of missing values with the last available data [].
Resampling techniques were reported in 21 (32.30%) papers. Eighteen studies balanced the datasets using methods such as using the minority class as a reference, undersampling the majority class [,], or the synthetic minority oversampling technique (SMOTE) [,,,,,,,,,,,,,,]. One paper compared the results of SMOTE against the adaptive synthetic algorithm []. Finally, 1 study stabilized variations in underrepresented outcome classes using bootstrap resampling []. Regarding cancer studies where cases were fewer than controls, 2 different strategies were applied to the same dataset: sample-size equalization by randomly grouping cancer-free participants based on the number of cancer survivors [], while another study matched cases and controls by sex, age, and education level, then selected a random sample resulting in 50% cases and 50% controls [].
Dimensionality reduction was used in 38 (58.46%) studies using 3 approaches. The first approach involved assessing the relationship between features and outcomes by removing redundant information [,,,,,,,,,]. Other methods included factor analysis [] and principal component analysis [,,]. The second approach optimized models to achieve lower prediction error [,,]. The third approach involved automatic selection of predictors during model training [,,,,-,,,,,,]. For more information, refer to Table S4 in .
SL Models
Models Overview
In this section, the core components of ML models are described, beginning with problem formulation and algorithm families, followed by evaluation components. Depending on the purpose of the ML analysis, papers were grouped as classification or regression when the objective was prediction, and as feature selection when the goal was explanation []. Most studies (46/65, 70.77%) focused on classification, 9 (13.83%) on regression, 2 (3.1%) on both classification and regression, and 8 (12.30%) on feature selection.
Six families of algorithms emerged from the studies: tree-based, deep learning, support vector machines, k-nearest neighbors, naïve Bayes, and nearest centroid. The approach adopted across 40 studies was a multimodel approach, with 29 of them incorporating algorithms from diverse algorithmic families. provides a comprehensive taxonomy of the specific algorithms implemented. The following subsections describe the application of specific algorithms within the 3 most used families (tree-based, deep learning, and support vector machines) in relation to the type of data used.

Tree-Based Algorithms
Tree-based algorithms were applied in 55 (84.61%) studies, covering all data types. RF was used in 45 (69.23%) out of 55 studies. Specifically, RF was implemented in 27 cross-sectional studies [,,,,-,,,,,-,-,-,,-], 11 longitudinal studies [,,,,,,,,,,], 3 time-series studies [,,], 2 textual studies [,], 1 study with both cross-sectional and geographical data [], and 1 study with both longitudinal and time-series data [].
Different versions of the gradient boosting algorithm were performed in 28 studies, including gradient boosting machines, extreme gradient boosting, adaptive boosting, and light gradient boosting machine [,,,,,,,,,-,,,,,,,,,,,,,,,,]. Finally, decision tree algorithms were implemented in 19 studies [,,,,,,-,,,,,,,,,,], and Bayesian additive regression trees in 2 studies [,] (; ).
Deep Learning Algorithms
Neural networks (NNs) are considered the cornerstone of deep learning algorithms. Various NN architectures were applied in 24 (36.92%) of the reviewed studies. A multilayer perceptron neural network was used in cross-sectional [,,], longitudinal [,], and time-series data [,,]. Long short-term memory recurrent neural network was applied to longitudinal [,] and time-series data [,,]. Feed-forward neural networks were used for cross-sectional data [,,]. Convolutional neural networks were used to analyze time-series [] and cross-sectional data []. Simple Elman-type recurrent neural networks were applied in a time-series study []. Radial basis function neural networks, fuzzy neural networks (FNNs), and probabilistic neural networks were used in a cross-sectional study []. In addition, Transformer, gated recurrent unit with attention, WaveNet, and RNNs were used to analyze longitudinal data []. In contrast, 11 studies did not specify the artificial neural network architecture used [,,,,,,,,,,] (; ).
Support Vector Machine Algorithms
Support vector machine (SVM) algorithms were used in 19 (27.08%) studies, applied across various data types. SVM was implemented in 11 cross-sectional studies, 2 longitudinal studies, 3 time-series studies, and 2 studies that analyzed textual data. Configurations included support vector machine with a radial basis function [,,,] or a support vector machine with linear kernel [,,,]. Eleven studies did not report the type of kernel used [,,,,,,,,,,] (; ).
Evaluation
Final model evaluation procedures were explicitly reported in 57 (87.69%) reviewed studies, with the hold-out test set being the most applied strategy. These included a hold-out test set [,,,,,,,,,,-,,-,,,-,,,,], k-fold cross-validation (final evaluation) [,,,,,,,,,,,,,,,,,,,], nested cross-validation [,], leave-one-out cross-validation (for small datasets with n<150) [,,]. Two studies used external datasets for model performance assessment [,]. For time-series data, 2 studies divided the dataset based on the time of acquisition, keeping an out-of-sample dataset for model evaluation [,], while 1 study [] used 5-fold cross-validation in an offline setting followed by an online weighted resampling methodology to address drift.
Among the 11 studies that addressed regression problems, the reported evaluation metrics included mean absolute error [,,,], mean squared error [,,], root-mean-square error [,,,,,,], mean absolute percentage error [], and coefficient of determination (R2) [,,,,,].
For classification problems, reported evaluation metrics included specificity and sensitivity (recall) [-,,,,,,,,,,,,,,], precision and recall [-,,,,,,,], the confusion matrix [,,], error rate as the proportion of misclassified observations (1 – accuracy) [,,,], Cohen κ [,], F1-score [,,,,,,,,,,], and model training time []. The most frequently used metrics were accuracy [,-,-,,,,,,,,,,,,,,,,] and area under the receiver operating characteristic curve [-,-,,,,,-,,,,,,,,,,] (Table S5 in ). Additionally, 1 epidemiologic study used the Brier score to assess cardiovascular mortality [].
Explainability
To enhance explainability, 22 (33.84%) studies implemented specific XAI methods to clarify the contribution of each predictor and the direction of the relationship with the outcome. Of these, 21 studies used Shapley Additive Explanations (SHAP) values, while 1 used local interpretable model-agnostic explanations (LIME). As shown in , XAI methods, particularly SHAP values, have been applied across different health domains, showing a remarkable increase in use in 2025. The most common visualization methods were beeswarm and bar plots, with the contribution of each feature ranked along the y-axis, placing the most important feature at the top of the plot. The number of features plotted ranges from 6 to 50, with 20 being the most common (refer to ).

| Study | Health outcome | Total number of features (features ranked) | Healthy lifestyle components (category, position, features ranked) | XAI (visualization) |
| Allen [] | Obesity | 64 (10) | Physical activity (physical inactivity, 1, 10); diet (food insecurity, unranked, n/a); sleep (sleep hours, unranked, n/a); stress (n/ac, n/a, n/a). | LIME (waterfall plot) |
| Gu et al [] | Infertility risk in women | 39 (10) | Physical activity (physical activity health score, unranked, n/a); diet (diet health score, 10, 10); sleep (sleep health score, unranked, n/a); stress (n/ac, n/a, n/a). | SHAP values (beeswarm plot, bar plot, and dependency plot) |
| Guthrie et al [] | Cardiometabolic disease | 13 (13) | Physical activity (minutes of physical activity, 8, 13); diet (plant-based meal, 6, 13); sleep (n/ac, n/a, n/a); stress (n/ac, n/a, n/a). | SHAP values (beeswarm plot) |
| Huang et al [] | Cognitive function | 20 (20) | Physical activity (exercise, 10, 20); diet (n/ac, n/a, n/a); sleep (sleep quality, 8, 20); stress (n/ac, n/a, n/a). | SHAP values (beeswarm plot and bar plot) |
| Jin and Halili [] | Mental health | 21 (21) | Physical activity (intensity, unranked, n/a); diet (n/ac, n/a, n/a); sleep (sleep hours, 1, 21); stress (n/ac, n/a, n/a). | SHAP values (bar plot) |
| Kim et al [] | Quality of life | 21 (20) | Physical activity (transport-related, 4, 20; physical activity score, 7, 20); diet (eating index, 15, 20); sleep (sleep quality, 2, 20; sleep hours, 6, 20); stress (stress level, 1, 20). | SHAP values (beeswarm plot) |
| Kiss et al [] | Mental health | SHAP values (bar plot) | ||
| Positive affect: 231 (20) | Physical activity (doing outdoor activities, 3, 20; duration of sitting, 6, 20; frequency of walking, 7, 20); diet (n/ac, n/a, n/a); sleep (sleep disorder, 8, 20); stress (coping strategies, 9, 20). | |||
| Perceived stress: 228 (20) | Physical activity (engagement, 12, 20); diet (n/ac, n/a, n/a); sleep (sleep hours, 7, 20); stress (coping strategies, 5, 20). | |||
| Anxiety: 228 (20) | Physical activity (engagement, unranked, n/a); diet (n/ac, n/a, n/a); sleep (sleep hours, 15, 20); stress (coping strategies, 10, 20). | |||
| Depressive symptoms: 240 (20) | Physical activity (engagement, 14, 20); diet (n/ac, n/a, n/a); sleep (sleep hours, 16, 20); stress (coping strategies, 9, 20). | |||
| Li and Song [] | Cognitive function | 20 (20) | Physical activity (sport social capital index, 3, 20); diet (n/ac, n/a, n/a); sleep (n/ac, n/a, n/a); stress (n/ac, n/a, n/a). | SHAP values (beeswarm plot, heat map, temporal analysis, and dependency plot) |
| Lin et al [] | Loneliness | 15 (15) | Physical activity (exercise, 7, 15); diet (n/ac, n/a, n/a); sleep (n/ac, n/a, n/a); stress (n/ac, n/a, n/a). | SHAP values (beeswarm plot) |
| Luo et al [] | Frailty | SHAP values (beeswarm plot and dependency plot) | ||
| US cohort: 121 (20) | Physical activity (play sports and exercise, 8, 20; moderate physical activity, 16, 20); diet (n/ac, n/a, n/a); sleep (sleep problems, 4, 20); stress (n/ac, n/a, n/a). | |||
| UK cohort: 125 (20) | Physical activity (vigorous physical activity, 4, 20; moderate physical activity, 5, 20); diet (fruit consumption, 19, 20); sleep (sleep problems, 2, 20; sleep duration, 6, 20); stress (n/ac, n/a, n/a). | |||
| China cohort: 94 (20) | Physical activity (n/ac, n/a, n/a); diet (n/ac, n/a, n/a); sleep (sleep problems, 5, 20; sleep duration, 8, 20); stress (n/ac, n/a, n/a). | |||
| Majcherek et al [] | Mental health | 26 (24) | Physical activity (sport, 6, 24; walking time, 11, 24); diet (vegetable portion, 13, 24; fruit portion, 14, 24); sleep (n/ac, n/a, n/a); stress (n/ac, n/a, n/a). | SHAP values (dependency plot) |
| Majcherek et al [] | Diabetes | 22 (6) | Physical activity (regular physical activity, unranked, n/a); diet (habits, unranked, n/a); sleep (n/ac, n/a, n/a); stress (n/ac, n/a, n/a). | SHAP values (dependency plot) |
| Morris et al [] | Cardiovascular disease | 50 (50) | Physical activity (availability of outdoor activities, 5, 50); diet (available favorable food stores, 14, 50); sleep (n/ac, n/a, n/a); stress (global stress, 18, 50). | SHAP values (bar plot) |
| Park et al [] | Visceral fat | 32 (20) | Physical activity (frequency, unranked, n/a); diet (high rice consumption, 2, 20; Asian-style balanced diet, 6, 20); sleep (n/ac, n/a, n/a); stress (n/ac, n/a, n/a). | SHAP values (beeswarm plot and bar plot) |
| Ren et al [] | Cognitive function | 39 (20) | Physical activity (exercise, 16, 20); diet (n/ac, n/a, n/a); sleep (n/ac, n/a, n/a); stress (n/ac, n/a, n/a). | SHAP values (beeswarm plot, bar plot, and force plot) |
| Sandri et al [] | Lifestyle (diet) | SHAP values (beeswarm plot) | ||
| Mediterranean diet: 41 (20) | Physical activity (sport, 8, 20); diet (fish consumption, 1, 20); sleep (sleep quality, unranked, n/a); stress (n/ac, n/a, n/a). | |||
| Intermittent fasting: 41 (20) | Physical activity (sport, 3, 20); diet (fish consumption, 1, 20); sleep (sleep quality, 14, 20); stress (n/ac, n/a, n/a). | |||
| Vegan diet: 41 (20) | Physical activity (sport, 5, 20); diet (fish consumption, 1, 20); sleep (sleep quality, 9, 20); stress (n/ac, n/a, n/a). | |||
| Vegetarian diet: 41 (20) | Physical activity (sport, 5, 20); diet (fish consumption, 1, 20); sleep (sleep quality, unranked, n/a); stress (n/ac, n/a, n/a). | |||
| Shi et al [] | Osteoporosis | 45 (20) | Physical activity (physical activity health score, 16, 20); diet (diet health score, unranked, n/a); sleep (sleep health score, 5, 20); stress (n/ac, n/a, n/a). | SHAP values (beeswarm plot, waterfall plot, and force plot) |
| Su et al [] | Resilience | 12 (4) | Physical activity (n/ac, n/a, n/a); diet (n/ac, n/a, n/a); sleep (sleep disturbance, 3, 4); stress (n/ac, n/a, n/a). | SHAP values (beeswarm plot, bar plot, and dependency plot) |
| Wang et al [] | Lifestyle (sleep) | 7 (7) | Physical activity (n/ac, n/a, n/a); diet (nutritional status, 3, 7); sleep (outcome, n/a, n/a); stress (n/ac, n/a, n/a). | SHAP values (beeswarm plot, bar plot, and waterfall plot) |
| Xin and Ren [] | Mental health | SHAP values (beeswarm plot and bar plot) | ||
| Rural older adults: 55 (20) | Physical activity (exercise, 19, 20); diet (n/ac, n/a, n/a); sleep (sleep hours, 13, 20); stress (n/ac, n/a, n/a). | |||
| Urban older adults: 55 (16) | Physical activity (exercise, unranked, n/a); diet (n/ac, n/a, n/a); sleep (sleep hours, 6, 16); stress (n/ac, n/a, n/a). | |||
| Zhang et al [] | Lifestyle (sleep) | 20 (10) | Physical activity (sedentary time, 5, 10); diet (vegetable consumption, 3, 10); sleep (outcome, n/a, n/a); stress (stress score, 1, 10). | SHAP values (beeswarm plot and dependency plot) |
| Zhou et al [] | Psoriasis | 150 (20) | Physical activity (intensity, 17, 20); diet (habits, unranked, n/a); sleep (n/ac, n/a, n/a); stress (n/ac, n/a, n/a). | SHAP values (beeswarm plot) |
aXAI: explainable artificial intelligence.
bn/ac: not applicable.
cn/a: not available.
dLIME: local interpretable model-agnostic explanations.
eSHAP: Shapley Additive Explanations.
Because the nonlifestyle features included in each model vary substantially across studies, the ranking reported in should be interpreted in relative terms. When a lifestyle behavior appears among the top-ranked features, this indicates that it contributed more strongly to the model than nonlifestyle variables included in the analysis. Conversely, lifestyle components appearing in lower positions acted as secondary predictors.
In a study on cardiometabolic disease [], the ML solution was first explained at the individual participant level to provide specific behavioral feedback, and then at the group level to reveal the ranking of features for succeeding in behavioral changes. In both models, physical activity and diet variables were among the top contributors. However, in another study [], neither regular physical activity nor diet habits were among the top 6 variables for predicting diabetes in adults.
In 4 mental health studies [,,,], SHAP values were used to rank the contribution of each feature. In 3 of the 4 studies, lifestyle variables were among the top contributors. In [], physical activity and fruit and vegetable consumption; in [], sleep variables were the top predictors of stress in young adolescents during the COVID-19 pandemic; and in [], sleep duration was identified as important for predicting depression. However, in [], the intensity of physical activity was unranked because it was excluded from the XAI analysis during a feature selection step prior to modeling. In the same study, sleep hours were the most important variable for predicting depression among adults. Regarding mental health studies, psychological resilience was assessed among medical students [], with sleep disturbance being a key factor affecting their resilience.
Three studies focusing on older adults in China had cognitive function as their outcome [,,]. In these studies, physical activity variables were ranked as top predictors. Additionally, in 1 of these studies [], sleep quality was ranked as the eighth top feature out of 20. Furthermore, when loneliness was assessed among older adults from China [], exercise was ranked as the seventh variable out of 15.
In a study predicting quality of life [], stress, sleep quality, and physical activity emerged as the strongest predictors, with the eating index appearing in the top 15 variables.
Two studies focused on sleep as the specific outcome. In [], the SHAP values ranked stress score, vegetable consumption, and sedentary time among the top 5 variables for predicting sleep disturbance. In [], nutritional status was the third most important variable for predicting the risk of sleep disorders in older adults.
In [], “favorable” food stores and global stress were identified as the top variables for predicting incidence of cardiovascular disease, with the availability of outdoor activities ranking in the top 5. In a longitudinal study investigating the association of diet with long-term reduction in waist circumference, SHAP values highlighted the importance of high-quality components in reducing visceral fat []. This study also measured exercise with a single item of frequency, which was not included in the top ranking. Regarding a diet study [], the adoption of different diets in the Spanish population was assessed, with fish consumption positioned as the most important variable for all diets studied (Mediterranean, intermittent fasting, vegan, and vegetarian). In the same study, practicing sport was ranked among the top variables across the different diets. However, sleep quality was only included in the ranking for intermittent fasting and the vegan diet.
Physical inactivity emerged as the most important feature in explaining county-level obesity using LIME []. In this study, the food environment and insufficient sleep, both measured as single items, were not included as top predictors of obesity prevalence.
In a cross-national study assessing frailty [], sleep variables were among the 20 most important variables across all the cohorts studied. Physical activity was included in the US and UK cohorts and was also ranked in the top 20 variables. In the UK cohort, fruit consumption was the 19th out of 20 key contributors to frailty.
One study on osteoporosis [] used the Life’s Essential 8 scores for physical activity, diet, and sleep. Sleep and physical activity scores were among the 20 most important variables, but the diet health score was unranked. Similarly, the Life’s Essential 8 scores were used to determine infertility risk in women [], with the diet score being the only one ranked among the top 10 variables.
Finally, 1 study combined lifestyle factors with metabolites associated with psoriasis [], with physical activity intensity ranked among the 20 key factors out of 150 variables, whereas dietary habits were not ranked.
Software to Implement ML Models
Of the included papers in the review, 23 (35.38%) used R software (R Core Team) [] for data analysis (). The R packages used were data.table for data manipulation, tidyverse [] as a general package for data science, Multivariate Imputation by Chained Equations [] for missing data imputation, missMDA for performing multiple imputation with principal component analysis, FactoMineR for exploratory data analysis and principal component analysis, Boruta [] for feature selection through a wrapper algorithm, caret (Classification and Regression Training) [] for creating models, randomForest for RF analysis, randomForestSCR for RF for survival, regression, and classification analysis, rpart for recursive partitioning and regression trees, xgboost for extreme gradient boosting, bartMachine for Bayesian additive regression trees, kernlab and e1071 for support vector machines, survival for survival analysis, lime for local interpretable model-agnostic explanations, and finally, SuperLearner [] to choose the optimal learner for a given prediction problem with a k-fold cross-validation algorithm.
| Software used | Number of studies | Study references |
| R (R Core Team) | n=23 | [,-,,,,,-,,,,,,,-,,,] |
| Python (Python Software Foundation) | n=24 | [,,,,,,,,,,,,,,,,,,,,,,,] |
| R and Python | n=4 | [,,,] |
| SPSS (IBM Corp) | n=2 | [,] |
| MATLAB (The MathWorks Inc) | n=2 | [,] |
| KNIME [] | n=2 | [,] |
In contrast, 24 (36.92%) studies developed the models in Python (Python Software Foundation), using the following libraries: Scikit-learn, used in all studies for predictive data analysis; pandas, for manipulating tabular data; NumPy, for mathematical functions; Keras and TensorFlow, for implementing deep learning; lightGBM, for performing light gradient boosting machine; SHAP, to explain ML solutions; creme, for online ML; Bayesian optimization, as a global optimization package to find the maximum value of an unknown function in as few iterations as possible; imbalanced-learn, to combine either undersample or oversample methods; and TextBlob, emoji, nltk, and profanity, for processing and analyzing textual data.
Finally, 4 studies used both Python and R, and 6 studies used other software programs such as SPSS, MATLAB, and KNIME. Eight papers did not report the software used [,,,,,,,].
Discussion
Overview
This scoping review of 65 studies provides the current state of the application of supervised ML algorithms for the analysis of lifestyle data. The increase in studies in this field since 2019 indicates that it is a noteworthy area of study. The diversity in the sample origin, alongside the accessibility to new AI tools and novel methods for monitoring health outcomes (eg, wearables), denotes global attention to lifestyle. This section addresses the methodological shortcomings found in the reviewed studies.
About Data Acquisition
In relation to lifestyle data, we found that most studies adopted a multidomain approach, integrating more than just a single component. This strategy enhances and facilitates a more comprehensive understanding of health problems related to the 4 lifestyle domains considered in this review: physical activity, diet, sleep, and stress. The distribution of lifestyle domains identified here was similar to that reported in a previous scoping meta-review [], although we observe that sleep has gained prominence in recent years, now reaching a level comparable to diet. Although these results highlight the growing recognition of the interrelated nature of lifestyle behaviors, the imbalance in the distribution of these factors limits the capacity of current studies to fully model and understand the interaction among the 4 lifestyle components and their combined effects on health.
Concerning the data acquisition process, over half of the studies acquired their own data. This acquisition process implies control over variables and reduces the time required for cleaning []. Interestingly, both self-acquired datasets and those sourced from private or public health repositories demonstrated gender parity in the analyzed datasets. However, we detected a major limitation in this part of the process regarding the data acquisition methodology. In most studies, data were collected through single items, such as regular physical activity (response “yes” or “no”) [] or usual time of waking up and going to bed [], resulting in low representativeness of the construct being measured. The result of this acquisition method is high heterogeneity in measures, which hinders their generalizability. Therefore, the quality of data must be one of the challenges to be addressed, and specifically, the consistency in measures [,].
Nevertheless, the current accessibility and precision of health sensors such as wearables [] and the Internet of Things [] may contribute to transferability and actionability in the population []. The growth in technology allows the integration of different data forms as well as more objective measures of lifestyle, substantially reducing the impact of retrospective bias by tracking real-time data in an ecological situation []. Therefore, merging questionnaires and sensor data may be the key to identifying relationships between lifestyle measurements and personalizing interventions or changes in specific behaviors. This integration would include physiological, psychological, and behavioral factors, which are the most common analysis types in the ML community to extract clinical insights [].
About Characteristics of ML Models
Regarding the analysis of ML, 2 different approaches emerged in the reviewed studies: 1 focused on prediction through classification and regression problems, and the other focused on interpretability through feature selection. The first is already an acknowledged approach, while the second typically constitutes an important component of the ML process, specifically during the preprocessing stage. However, feature selection studies do not use model evaluation metrics, which can limit their statistical validity and the generalizability of results. Remarkably, the family of ML algorithms most closely related to feature selection is tree-based because it provides indices of the importance of each variable. Although most papers in this scoping review combined different families of algorithms and compared their results, the most common model family was tree-based, which was applied for each data typology identified. Specifically, RF is the most used algorithm, which may be due to its robustness in handling missing values, the consideration of complex interaction in the data [], and its lower sensitivity to variable scales []. Despite the benefits of RF, the underusage of DL algorithms represents a critical missed opportunity for robustly analyzing complex and multimodal data. In this review, DL algorithms were underused for lifestyle data. This result may reflect a gap in expertise or access to computational resources among lifestyle researchers, potentially limiting the application of more complex models. With ongoing advances in computational power and algorithmic efficiency, it is expected that the use of DL algorithms will become more widespread in the near future [].
Regarding the preprocessing stage, most studies detailed some phases of the process, but there is no consensus on the description of this stage of ML. Variable transformation is a crucial step for certain algorithms, particularly for SVM and specific architectures of DL that exhibit sensitivity to the raw form of the variables. In this review, 12 out of 19 studies that performed the SVM algorithm, and 13 out of 26 studies focused on DL, reported variable transformation. Additionally, it is worth noting that these algorithms cannot handle missing values, requiring imputation before the modeling phase. Among the SVM studies, 11 out of 19 reported techniques for missing imputation, and 17 out of 26 DL studies explicitly addressed this. In contrast, tree-based algorithms are less sensitive to variable scales and missing values, yet incorporating these feature engineering steps could enhance model performance []. It is noteworthy that the preprocessing steps, specifically how missing values are addressed, have been identified as a potential concern for transparency. This procedural aspect could introduce sampling biases, thereby influencing the generalizability and comprehension of the dataset context [].
Resampling techniques, aiming to balance the dataset, are commonly implemented in classification problems. SMOTE has been the most widely used technique in this review, especially because it achieves better results than a simple undersampling of the majority class. In the health domain, imbalanced datasets are common, and SMOTE oversamples the minority class with synthetic examples and randomly undersamples the majority class to balance the dataset [].
Finally, dimensionality reduction enables capturing the most relevant information for the outcome while eliminating noise and redundant information. In this review, dimensionality reduction was the most frequent preprocessing step, appearing in 38 studies. Notably, not only could SVM and DL models benefit from removing irrelevant predictors, but also in tree-based algorithms, dimensionality reduction minimizes model complexity, resource consumption, and data acquisition costs [].
The division of the original dataset is an essential step for assessing the performance of the ML solution. In this review, 8 studies did not report how they split their data to assess the model, which denotes a lack of generalization of their results. This omission represents a common issue in ML research that should be carefully addressed to minimize bias []. Train and test division, also known as “hold out,” is a method with considerable variability due to the use of a unique random data distribution []. Therefore, other methods might be more suitable. For example, leave-one-out cross-validation, which trains the model on “n – 1” observations and makes predictions on the remaining one. Although effective for small datasets, it is computationally intensive with large datasets []. K-fold cross-validation involves randomly dividing the original dataset into k groups. K-fold cross-validation not only offers computational advantage over leave-one-out, but also gives more accurate estimations due to the bias-variance trade-off []. In time-series data, only 1 paper [] introduced a different form of data split, considering the dependencies of the entire series. Ideally, this type of data should be treated with a method called rolling forecast origin resampling, which estimates the model with historical data and evaluates it with the most recent data []. In other words, the training set should ideally comprise observations that occurred before those in the test set; however, this method was not found in this review.
Regarding the evaluation of ML models, the choice of metrics depends on the nature of the problem, whether regression or classification. Specific evaluation metrics tailored to each problem are crucial for correct evaluation, aligning with the priorities and needs of each field. For instance, in medical studies where the cost of treatment in terms of health is high, it becomes crucial to identify true patients over false positives. In contrast, if the treatment has minimal side effects and has demonstrated benefits, sensitivity might not be as important as specificity. It is worth highlighting that in the field of data science, precision and recall are more commonly used, whereas in medical fields, specificity and sensitivity are more prevalent []. These differences may cause misunderstandings between the 2 domains.
About Explainability Methods
Regarding model explainability, 22 studies incorporated a dedicated step in the ML process for explainability. SHAP values and LIME were the only XAI methods applied to lifestyle data, and these 2 methods were the most common in a recent systematic review on XAI methods []. XAI-related studies in this review were published since 2019, with a notable increase in the number of publications in 2025, comprising half of the papers. This exponential growth was also found in [,], where the trend in published papers occurred between 2016 and 2022. Thus, our review demonstrates this exponential distribution in the health and behavioral sciences, where XAI methods are gaining prominence.
Although tree-based algorithms, especially decision trees, are known for facilitating interpretation, SHAP values can be applied to any type of model []. The adoption of XAI in lifestyle studies remains low (33.84%). One possible explanation for this is that explainability algorithms are often not integrated into a standard ML pipeline, thereby increasing the technical complexity of the workflow. However, some efforts are being made by R and Python developers to incorporate XAI algorithms into pipelines using libraries such as H2O [].
In this review, lifestyle components (physical activity, diet, sleep, and stress) consistently appeared among the top-ranked features in models using XAI techniques, highlighting their substantive contribution relative to nonlifestyle variables. These findings align with prior research emphasizing the integration of diverse lifestyle components [,]. However, the level of interpretability achieved also depends on the quality of data used in the models, which in some studies did not meet expected standards. For instance, in a study where the focus was on obesity, the diet component was not among the top-ranked features []. In this county-level study, a food environment index was measured as a single item, potentially inadequately representing the diet component of lifestyle.
Therefore, integrating XAI methods into the ML process could enable tailored interventions based on model results, provided that measures are collected accurately. Furthermore, the adoption of XAI algorithms contributes to increased trust and verification of the fairness of the models. This approach can also facilitate the translation of findings to stakeholders and health systems, thereby enhancing transparency, promoting the adoption of models in society, and supporting informed decision-making [].
About the Software for Implementing ML Models
Competition between Python and R for ML software dominance in data science is currently intense. Both Python and R are freely distributed, object-oriented software with large and active communities. Python, as a programming language, offers specific implementations through libraries tailored for statistical analysis, including ML and DL. In contrast, R is a statistical software that integrates fundamental statistics into its base functionalities. While Python requires libraries for each stage of analysis, its well-established libraries streamline the process. On the other hand, R faces challenges due to its heterogeneous libraries, which hinder replicability and require expertise in varying syntax across packages. To address this concern, the meta-package tidymodels (Max Kuhn and Hadley Wickham) resolves these issues by integrating all necessary packages for each ML step, using a unified syntax. Additionally, tidymodels integrates user-friendly interfaces and promotes good methodological practice, thereby preventing user errors []. Conversely, Python presents a preferable environment for DL with the TensorFlow and PyTorch frameworks. In this regard, the possibility of developing ML projects on powerful computational cloud-based platforms, such as Google Colaboratory (also referred to as Google Colab) [], offers Python a remarkable advantage over R local environments.
Methodological and Reporting Guidelines and Checklist
Based on the review’s results and to enhance transparency and replicability in multidisciplinary sciences [], we provide comprehensive methodological and reporting guidelines and a checklist for ML projects. Although various studies have proposed guidelines and checklists [,], the rapid expansion of ML algorithms in health domains necessitates iterative evaluation to incorporate new steps into the ML research workflow. The guidelines and checklist () are based on the 5 stages of the ML workflow, as depicted in , with added software tools.
Data Acquisition
The integration of multidomain data enhances the comprehension of real-world problems. Using appropriate methods to collect data ensures representativeness. We recommend the use of standardized questionnaires and validated sensors. Regarding health repositories, we recommend providing information about data characteristics such as gender distribution, sample size, and variable descriptions [].
Preprocessing
Reporting the preprocessing methods used in the data analysis is particularly crucial for ensuring replicability. While preprocessing contributes to improving data quality, different preprocessing methods can lead to different results. We propose the following recommendations for each preprocessing step, although not all steps need to be performed in every ML project.
Transformation
Categorical data should be encoded using methods such as one-hot encoding and dummy variables. Continuous data should be transformed using normalization or rescaling of features with different units to ensure algorithm performance, particularly for those sensitive to the raw form of variables [].
Missing Imputation
Some algorithms cannot handle missing data and require imputation before modeling. Depending on the number of observations and the data distribution, imputation with the mean, median, or mode is typical. For time-series data, imputation with the last or next observation is preferred, though rolling statistics imputation or interpolation may offer better solutions [].
Resampling
Imbalanced datasets can bias models, resulting in poor performance on underrepresented classes []. While SMOTE is an effective technique for handling imbalanced datasets, it is not without its limitations. However, when the class imbalance ratio is extremely high, SMOTE can potentially bias the model performance by overfitting the minority class. This issue is particularly pronounced in datasets containing noise, as synthetic observations may replicate these artifacts. To mitigate these challenges, recent studies have proposed tree algorithms, which have shown effectiveness in handling class imbalance []. Furthermore, due to their robustness, the use of tree-based algorithms is increasingly recommended when working with class imbalance [].
Dimensionality Reduction
Removing noise from the dataset and retaining features directly related to the outcome can enhance both data acquisition and modeling efficiency. Removing correlated features is particularly beneficial [].
Modeling
Our guidelines focused on SL algorithms for classification and regression problems. The choice of algorithms depends on the measurement of the outcome. We recommend using multiple algorithms to compare results and select the best fit, given that there is significant variability across problems []. Additionally, comparing different families of algorithms is also advisable, as some improvements exist within the same family. When it comes to replicability and transparency, reporting algorithm hyperparameters is crucial in ML problems, as different configurations can yield varying results [].
Evaluation
Avoiding overfitting requires dividing the original dataset appropriately. This step is fundamental in ML implementation and should be considered in every study to ensure that the extracted insights are reliable and generalizable. This division depends on data typology; we identified 3 different typologies relevant at this stage, defined by whether time is an implicit factor in the data acquisition process.
Cross-Sectional
Data are collected at a single point in time, with no temporal dependencies. Leave-one-out cross-validation is effective for small datasets (n<150) typically seen in life sciences; the computational complexity increases []. K-fold cross-validation is recommended for large datasets to balance the bias-variance trade-off []. Cross-validation with bootstrap resampling can also be used to evaluate the performance of the models and estimate CIs of performance metrics [].
Longitudinal
In longitudinal studies, researchers collect repeated measures, potentially with dependencies between observations and high correlation that can bias the model []. Data should be divided by grouping individual participants’ information. The methods used are similar to those for cross-sectional data, but with consideration of partitioning.
Time Series
Time series data are a sequence of data points in chronological order. Rolling forecast origin resampling is suitable for this data []. The training set should include observations occurring before those in the test set.
Evaluation Metrics
Appropriate metrics should be selected based on the type of ML problem (regression or classification) [] and the characteristics of the study field []. It is also recommended to compute evaluation metrics repeatedly across cross-validation samples and to apply nonparametric tests, such as the Wilcoxon signed-rank test and the Friedman test, to assess model performance []. For imbalanced datasets, performance metrics such as balanced accuracy and F1-score are recommended. For example, in an imbalanced dataset, a model may achieve high accuracy by predominantly predicting the majority class. In contrast, balanced accuracy assigns equal weight to each class regardless of its frequency, providing a more informative evaluation of model performance [].
Explainability
Explainability Overview
Reporting the method of explainability used in the ML projects is essential. It is important to distinguish between interpretability and explainability. While explainability refers to understanding the effect of each feature on the original model, interpretability involves deriving actionable insights from the model’s prediction. Although tree-based methods include importance metrics, they do not indicate the direction of relationships with the outcome. Incorporating explainable resources such as SHAP values [] or LIME [] enhances the interpretability of the results, providing both actionability and transparency, and transforming black boxes into glass box models. The H2O package in R offers XAI algorithms that are compatible with the framework tidymodels, enabling a unified workflow for modeling and explainability. For a more detailed taxonomy of XAI packages in R, we refer the reader to [].
Additionally, the choice of appropriate visualization methods and the number of features displayed are crucial for ensuring comprehensive results and supporting fine-grained decision-making. Regarding the number of features, we recommend visualizing at least the top 20 features whenever possible, as this allows a broader understanding of the contribution of the most important variables. Furthermore, future studies should report all model features used in the ML analysis, not only the top-ranked ones, to ensure full transparency and allow readers to verify which HL components were excluded or had low impact in the model. Based on SHAP values, we propose the following XAI visualization techniques [,]:
Beeswarm Plot
This is the most common summary plot, where features are ranked on the y-axis from most to least contribution by their mean absolute SHAP value. The x-axis represents the SHAP values, which express the change in the log-odds (or model output), resulting in a positive or negative contribution for a specific observation. Each dot represents an observation in the dataset, and the color is indicative of the original value for that observation. Higher values are displayed as red and lower values as blue. The vertical dotted line represents the zero SHAP contribution. Contributions to the right assign a positive effect, and those to the left assign a negative effect.
Dependency Plots
Scatter plots that show the effect a single feature has on the model’s predictions. Each point represents an instance’s feature value and its corresponding SHAP value. This plot provides a detailed and clear explanation of the direction and magnitude of the relationship (whether linear or nonlinear) between the feature and the outcome.
Bar Plot
This plot provides information on the global feature importance. Features are ordered on the y-axis from the highest to the lowest average contribution to the outcome.
Waterfall Plot
This plot is a local explanation that illustrates the contribution of each feature in transforming the expected value into the final prediction for a specific instance. Each row represents a feature’s contribution; positive contributions are red, while negative contributions are blue.
Force Plot
This plot is a local explanation that shows the effect and the direction of the most impactful features for a given observation.
Software
Python and R were the most widely used software in the review. However, the variety of libraries can complicate the process. To address this, the metapackage “tidymodels” provides unified syntax, enhancing replicability []. For DL, Python offers a more powerful environment thanks to PyTorch and TensorFlow frameworks and the possibility of developing analyses on cloud services such as Google Colab. Furthermore, we encourage researchers to make their code publicly available in open repositories such as GitHub.
Limitations
At the level of the review itself, some limitations emerge. Although time-series data were frequently used in the analysis of lifestyle behaviors, the primary focus of this review was on the methodological framework of supervised ML, rather than on specific time-series modeling approaches. Future research should specifically address this gap by examining the current use of time-series models applied to wearable data to better capture human behavior. Moreover, the exclusion of UL algorithms limits the scope of ML algorithms covered in this review. For example, cluster analysis has been used to classify children according to their eating behaviors and identify features related to obesity [], with findings suggesting that interventions such as reducing eating speed may help prevent childhood obesity. This scoping review focuses on supervised ML approaches, but future research could examine the current application of UL in lifestyle data. Finally, a further limitation of this review is that the search strategy prioritized studies using the umbrella term “healthy lifestyle.” While this approach captured a representative sample of research explicitly conceptualized as HL, it may have excluded multidimensional studies examining combinations of physical activity, diet, sleep, or stress that did not use this terminology. Although an interaction block was incorporated to address the interconnectedness among these domains, we acknowledge that some multidomain research may not have been retrieved due to the specificity of the search string.
Conclusion
This review has identified several limitations within the studies reviewed that need to be addressed. First, ensuring data quality remains a significant challenge that must be addressed by carefully selecting data acquisition methods to build reliable and robust models. Second, the evaluation process is crucial for preventing overfitting, and using hold-out cross-validation can lead to high variance partitioning. Therefore, it is recommended to implement k-fold cross-validation at various stages, such as during validation; for time-series data, rolling forecast origin resampling is recommended [].
In conclusion, this scoping review provides a comprehensive analysis of lifestyle using ML models and serves as a guideline for future research. While the relationship between lifestyle and health is well-established, ongoing efforts are needed to refine how we measure lifestyle to create robust models. It is essential to focus not only on model performance but also on data representativeness, which is closely related to the granularity established during data collection. Although RF algorithms are prominent in lifestyle data analysis, it is recommended to compare their performance with other algorithms within and across families. Future research should also incorporate SHAP values to enhance interpretability within the ML workflow. Additionally, the tidymodels metapackage (R software) with H2O for XAI can assist researchers in evaluating process quality with unified syntax, thereby contributing to replicability.
Acknowledgments
This work was supported by the Catalonian Government through a FI-SDUR grant (ref 2023FISDU_00217).
Funding
This work was supported by grants PID2019-107473RB-C21 and PID2022-141403NB-I00, funded by MCIN/AEI/10.13039/501100011033/FEDER, UE, and 2021SGR-00806, funded by the Catalonian Government. Additional funding was provided by the Catalonian Government through an FI-SDUR grant (ref 2023FISDU_00217).
Data Availability
The datasets generated or analyzed during this study are available in the
CORA Repositori de Dades de Recerca repository [].
Authors' Contributions
TE contributed to conceptualization, validation, formal analysis, investigation, data curation, writing of the original draft, and visualization. LC contributed to conceptualization, writing (review and editing), supervision, project administration, and funding acquisition. CA contributed to validation and writing (review and editing). JML contributed to conceptualization, methodology, writing (review and editing), supervision, project administration, and funding acquisition.
Conflicts of Interest
None declared.
Checklist 2
Reporting checklist for machine learning analysis in healthy lifestyle data.
DOCX File, 17 KBReferences
- Furihata R, Konno C, Suzuki M, et al. Unhealthy lifestyle factors and depressive symptoms: a Japanese general adult population survey. J Affect Disord. Jul 2018;234:156-161. [CrossRef] [Medline]
- Nudelman G, Shiloh S. Connectionism and behavioral clusters: differential patterns in predicting expectations to engage in health behaviors. Ann Behav Med. Sep 13, 2018;52(10):890-901. [CrossRef] [Medline]
- Nudelman G, Shiloh S. Mapping health behaviors: constructing and validating a common-sense taxonomy of health behaviors. Soc Sci Med. Dec 2015;146:1-10. [CrossRef] [Medline]
- Kaminsky LA, German C, Imboden M, Ozemek C, Peterman JE, Brubaker PH. The importance of healthy lifestyle behaviors in the prevention of cardiovascular disease. Prog Cardiovasc Dis. Jan 2022;70:8-15. [CrossRef]
- Braun C, Foreyt JP, Johnston CA. Stress: a core lifestyle issue. Am J Lifestyle Med. 2016;10(4):235-238. [CrossRef] [Medline]
- Wong VWH, Ho FYY, Shi NK, Sarris J, Chung KF, Yeung WF. Lifestyle medicine for depression: a meta-analysis of randomized controlled trials. J Affect Disord. Apr 2021;284:203-216. [CrossRef]
- Sagner M, Katz D, Egger G, et al. Lifestyle medicine potential for reversing a world of chronic disease epidemics: from cell to community. Int J Clin Pract. Nov 2014;68(11):1289-1292. [CrossRef]
- Cerf ME. Healthy lifestyles and noncommunicable diseases: nutrition, the life‐course, and health promotion. Lifestyle Medicine. Apr 2021;2(2):e31. URL: https://onlinelibrary.wiley.com/toc/26883740/2/2 [Accessed 2026-02-20] [CrossRef]
- Gurrin C, Smeaton AF, Doherty AR. LifeLogging: personal big data. Foundations and Trends® in Information Retrieval. Jun 16, 2014;8(1):1-125. [CrossRef]
- Lim J, Jeong CY, Lim JM, et al. Assessing sleep quality using mobile EMAs: opportunities, practical consideration, and challenges. IEEE Access. 2022;10:2063-2076. [CrossRef]
- Kline A, Wang H, Li Y, et al. Multimodal machine learning in precision health: a scoping review. NPJ Digit Med. Nov 7, 2022;5(1):171. [CrossRef] [Medline]
- Beam AL, Kohane IS. Big data and machine learning in health care. JAMA. Apr 3, 2018;319(13):1317-1318. [CrossRef] [Medline]
- Maleki F, Ovens K, Najafian K, Forghani B, Reinhold C, Forghani R. Overview of machine learning part 1. Neuroimaging Clin N Am. Nov 2020;30(4):e17-e32. [CrossRef]
- Secinaro S, Calandra D, Secinaro A, Muthurangu V, Biancone P. The role of artificial intelligence in healthcare: a structured literature review. BMC Med Inform Decis Mak. Apr 10, 2021;21(1):125. [CrossRef] [Medline]
- Breiman L. Statistical modeling: the two cultures (with comments and a rejoinder by the author). Statist Sci. Aug 1, 2001;16(3). [CrossRef]
- Choi RY, Coyner AS, Kalpathy-Cramer J, Chiang MF, Campbell JP. Introduction to machine learning, neural networks, and deep learning. Transl Vis Sci Technol. Feb 27, 2020;9(2):14. [CrossRef] [Medline]
- Efron B. Prediction, estimation, and attribution. Int Statistical Rev. Dec 2020;88(S1). URL: https://onlinelibrary.wiley.com/toc/17515823/88/S1 [Accessed 2026-02-20] [CrossRef]
- Bruce PC, Bruce A, Gedeck P. Estadística Práctica Para Ciencia de Datos Con R y Python: Más de 50 Conceptos Esenciales. 2nd ed. Marcombo; 2022. ISBN: 978-84-267-3443-3
- James G, Witten D, Hastie T, Learning TRS. An Introduction to Statistical Learning. Springer; 2021:15-57. [CrossRef] ISBN: 978-1-0716-1417-4
- Saqib K, Khan AF, Butt ZA. Machine learning methods for predicting postpartum depression: scoping review. JMIR Ment Health. Nov 24, 2021;8(11):e29838. [CrossRef] [Medline]
- Shatte ABR, Hutchinson DM, Teague SJ. Machine learning in mental health: a scoping review of methods and applications. Psychol Med. Jul 2019;49(9):1426-1448. [CrossRef] [Medline]
- Ono S, Goto T. Introduction to supervised machine learning in clinical epidemiology. Ann Clin Epidemiol. 2022;4(3):63-71. [CrossRef] [Medline]
- Bi Q, Goodman KE, Kaminsky J, Lessler J. What is machine learning? A primer for the epidemiologist. Am J Epidemiol. Dec 31, 2019;188(12):2222-2239. [CrossRef] [Medline]
- Aggarwal A, Tam CC, Wu D, Li X, Qiao S. Artificial intelligence-based chatbots for promoting health behavioral changes: systematic review. J Med Internet Res. Feb 24, 2023;25:e40789. [CrossRef] [Medline]
- Goh YS, Ow Yong JQY, Chee BQH, Kuek JHL, Ho CSH. Machine learning in health promotion and behavioral change: scoping review. J Med Internet Res. Jun 2, 2022;24(6):e35831. [CrossRef] [Medline]
- Lai X, Chen J, Lai Y, et al. Using large language models to enhance exercise recommendations and physical activity in clinical and healthy populations: scoping review. JMIR Med Inform. May 27, 2025;13:e59309. [CrossRef] [Medline]
- Mersha M, Wood J, AlShami A, Kalita J, Lam K. Explainable artificial intelligence: a survey of needs, techniques, applications, and future direction. arXiv. Preprint posted online on Aug 30, 2024. [CrossRef]
- Ali S, Abuhmed T, El-Sappagh S, et al. Explainable artificial intelligence (XAI): what we know and what is left to attain trustworthy artificial intelligence. Information Fusion. Nov 2023;99:101805. [CrossRef]
- Munn Z, Peters MDJ, Stern C, Tufanaru C, McArthur A, Aromataris E. Systematic review or scoping review? Guidance for authors when choosing between a systematic or scoping review approach. BMC Med Res Methodol. Nov 19, 2018;18(1):143. [CrossRef] [Medline]
- Tricco AC, Lillie E, Zarin W, et al. PRISMA Extension for Scoping Reviews (PRISMA-ScR): checklist and explanation. Ann Intern Med. Oct 2, 2018;169(7):467-473. [CrossRef] [Medline]
- Estrella T, Alfonso C, Capdevila L, Losilla JM. Machine learning for the analysis of healthy lifestyle data: a scoping review protocol. Universitat Autònoma de Barcelona. Mar 2023. URL: https://portalrecerca.uab.cat/en/publications/machine-learning-for-the-analysis-of-healthy-lifestyle-data-a-sco/ [Accessed 2026-02-20]
- Estrella T, Capdevila L, Alfonso C, Losilla JM. Replication data for machine learning for the analysis of healthy lifestyle data: a scoping review. CORA.Repositori de Dades de Recerca; 2024. [CrossRef]
- McGowan J, Sampson M, Salzwedel DM, Cogo E, Foerster V, Lefebvre C. PRESS Peer Review of Electronic Search Strategies: 2015 guideline statement. J Clin Epidemiol. Jul 2016;75:40-46. [CrossRef] [Medline]
- Rethlefsen ML, Kirtley S, Waffenschmidt S, et al. PRISMA-S: an extension to the PRISMA Statement for reporting literature searches in systematic reviews. Syst Rev. Jan 26, 2021;10(1):39. [CrossRef] [Medline]
- Loef M, Walach H. The combined effects of healthy lifestyle behaviors on all cause mortality: a systematic review and meta-analysis. Prev Med. Sep 2012;55(3):163-170. [CrossRef] [Medline]
- Yu J, Feng Q, Kim JH, Zhu Y. Combined effect of healthy lifestyle factors and risks of colorectal adenoma, colorectal cancer, and colorectal cancer mortality: systematic review and meta-analysis. Front Oncol. Jul 22, 2022;12:827019. [CrossRef]
- Lippman D, Stump M, Veazey E, et al. Foundations of lifestyle medicine and its evolution. Mayo Clin Proc Innov Qual Outcomes. Feb 2024;8(1):97-111. [CrossRef] [Medline]
- Sayburn A. Lifestyle medicine: a new medical specialty? BMJ. Oct 25, 2018:k4442. [CrossRef]
- Fleig L, Ngo J, Roman B, et al. Beyond single behaviour theory: adding cross‐behaviour cognitions to the health action process approach. British J Health Psychol. Nov 2015;20(4):824-841. URL: https://bpspsychub.onlinelibrary.wiley.com/toc/20448287/20/4 [Accessed 2026-02-20] [CrossRef]
- Butkevičiūtė E, Bikulčienė L, Žvironienė A. Physiological state evaluation in working environment using expert system and random forest machine learning algorithm. Healthcare (Basel). Jan 11, 2023;11(2):220. [CrossRef] [Medline]
- Cai T, Long J, Kuang J, You F, Zou T, Wu L. Applying machine learning methods to develop a successful aging maintenance prediction model based on physical fitness tests. Geriatrics Gerontology Int. Jun 2020;20(6):637-642. [CrossRef]
- Lim J, Kim J, Cheon S. A deep neural network-based method for early detection of osteoarthritis using statistical data. Int J Environ Res Public Health. Apr 10, 2019;16(7):1281. [CrossRef] [Medline]
- Recenti M, Ricciardi C, Edmunds KJ, et al. Healthy aging within an image: using muscle radiodensitometry and lifestyle factors to predict diabetes and hypertension. IEEE J Biomed Health Inform. Jun 2021;25(6):2103-2112. [CrossRef] [Medline]
- Recenti M, Ricciardi C, Edmunds K, Jacob D, Gambacorta M, Gargiulo P. Testing soft tissue radiodensity parameters interplay with age and self-reported physical activity. Eur J Transl Myol. Jul 12, 2021;31(3):9929. [CrossRef] [Medline]
- Staudenmayer J, He S, Hickey A, Sasaki J, Freedson P. Methods to estimate aspects of physical activity and sedentary behavior from high-frequency wrist accelerometer measurements. J Appl Physiol. Aug 15, 2015;119(4):396-403. [CrossRef]
- Zhou M, Fukuoka Y, Goldberg K, Vittinghoff E, Aswani A. Applying machine learning to predict future adherence to physical activity programs. BMC Med Inform Decis Mak. Aug 22, 2019;19(1):169. [CrossRef] [Medline]
- Dianati-Nasab M, Salimifard K, Mohammadi R, et al. Machine learning algorithms to uncover risk factors of breast cancer: insights from a large case-control study. Front Oncol. 2023;13:1276232. [CrossRef] [Medline]
- Matta SC, Sankari Z, Rihana S. Heart rate variability analysis using neural network models for automatic detection of lifestyle activities. Biomed SIGNAL Process CONTROL. Apr 2018;42:145-157. [CrossRef]
- Park H, Jung SY, Han MK, et al. Lowering barriers to health risk assessments in promoting personalized health management. J Pers Med. Mar 18, 2024;14(3):316. [CrossRef] [Medline]
- Li Y, Song H. The association between sports social capital and cognitive health: a longitudinal study of middle-aged and elderly adults in China. SSM Popul Health. Jun 2025;30:101778. [CrossRef] [Medline]
- Lin Y, Li C, Wang X, Li H. Development of a machine learning-based risk assessment model for loneliness among elderly Chinese: a cross-sectional study based on Chinese longitudinal healthy longevity survey. BMC Geriatr. Nov 14, 2024;24(1). [CrossRef]
- Ren H, Zheng Y, Li C, et al. Using machine learning to predict cognitive decline in older adults from the Chinese longitudinal healthy longevity survey: model development and validation study. JMIR Aging. Apr 30, 2025;8:e67437. [CrossRef] [Medline]
- Liu X, Luo Z, Jing F, et al. Estimating cardiovascular mortality in patients with hypertension using machine learning: the role of depression classification based on lifestyle and physical activity. J Psychosom Res. Feb 2025;189:112030. [CrossRef] [Medline]
- Amú Ruiz FA, Gonzalez Bustamante DA, Ortiz González K, Sandoval KM. Árbol de clasificación para la identificación de síntomas asociados a la depresión en estudiantes de una universidad pública (Classification tree for the identification of symptoms associated with depression in students of a public university). Retos. 2024;52:104-114. [CrossRef]
- Wallace ML, Coleman TS, Mentch LK, et al. Physiological sleep measures predict time to 15‐year mortality in community adults: application of a novel machine learning framework. J Sleep Res. Dec 2021;30(6):e13386. [CrossRef] [Medline]
- Su W, Jia H, Chang W, et al. The secrets of medical students’ psychological resilience: a dual perspective of machine learning and path analysis. Int J Med Inform. Jan 2026;205:106111. [CrossRef]
- Luo W, Gong L, Chen X, et al. Lifestyle and chronic kidney disease: a machine learning modeling study. Front Nutr. 2022;9:918576. [CrossRef] [Medline]
- Afrash MR, Bayani A, Shanbehzadeh M, Bahadori M, Kazemi-Arpanahi H. Developing the breast cancer risk prediction system using hybrid machine learning algorithms. J Educ Health Promot. 2022;11:272. [CrossRef] [Medline]
- Alshurafa N, Sideris C, Pourhomayoun M, Kalantarian H, Sarrafzadeh M, Eastwood JA. Remote health monitoring outcome success prediction using baseline and first month intervention data. IEEE J Biomed Health Inform. Mar 2017;21(2):507-514. [CrossRef] [Medline]
- Gu X, Li Q, Wang X. Using Life’s Essential 8 and heavy metal exposure to determine infertility risk in American women: a machine learning prediction model based on the SHAP method. Front Endocrinol (Lausanne). 2025;16:1586828. [CrossRef] [Medline]
- Nichols ES, Pathak HS, Bgeginski R, et al. Machine learning-based predictive modeling of resilience to stressors in pregnant women during COVID-19: a prospective cohort study. PLoS ONE. 2022;17(8):e0272862. [CrossRef] [Medline]
- Bôto JM, Marreiros A, Diogo P, et al. Health behaviours as predictors of the Mediterranean diet adherence: a decision tree approach. Public Health Nutr. Jul 2022;25(7):1864-1876. [CrossRef] [Medline]
- Huang Y, Huang Z, Yang Q, et al. Predicting mild cognitive impairment among Chinese older adults: a longitudinal study based on long short-term memory networks and machine learning. Front AGING Neurosci. 2023;15. [CrossRef]
- Kiss O, Alzueta E, Yuksel D, et al. The pandemic’s toll on young adolescents: prevention and intervention targets to preserve their mental health. J Adolesc Health. Mar 2022;70(3):387-395. [CrossRef] [Medline]
- Majcherek D, Ciesielski A, Sobczak P. AI-driven analysis of diabetes risk determinants in U.S. adults: exploring disease prevalence and health factors. PLoS ONE. 2025;20(9):e0328655. [CrossRef] [Medline]
- Mousavi H, Karandish M, Jamshidnezhad A, Hadianfard AM. Determining the effective factors in predicting diet adherence using an intelligent model. Sci Rep. Jul 19, 2022;12(1):12340. [CrossRef] [Medline]
- Jin T, Halili A. Predicting the risk of depression in older adults with disability using machine learning: an analysis based on CHARLS data. Front Artif Intell. 2025;8:1624171. [CrossRef] [Medline]
- Luo Y, Guo M, Zhang Q. Cross-national analysis of social determinants of frailty among middle-aged and older adults: a machine learning study in the USA, England, and China. Humanit Soc Sci Commun. May 30, 2025;12(1). [CrossRef]
- Puterman E, Weiss J, Hives BA, et al. Predicting mortality from 57 economic, behavioral, social, and psychological factors. Proc Natl Acad Sci USA. Jul 14, 2020;117(28):16273-16282. [CrossRef]
- Qasrawi R, Vicuna Polo S, Abu Khader R, et al. Machine learning techniques for identifying mental health risk factor associated with schoolchildren cognitive ability living in politically violent environments. Front PSYCHIATRY. 2023;14:1071622. [CrossRef] [Medline]
- Zhou X, Wang Y, Chang Q, et al. Lifestyle-associated serum metabolites profiling in relation to risk of late-onset psoriasis. J Eur Acad Dermatol Venereol. Sep 2025. [CrossRef] [Medline]
- Abdul Rahman H, Kwicklis M, Ottom M, et al. Machine learning-based prediction of mental well-being using health behavior data from university students. Bioengineering (Basel). May 10, 2023;10(5):575. [CrossRef] [Medline]
- Cortés-Ibañez FO, Nagaraj SB, Cornelissen L, et al. Prediction of incident cancers in the lifelines population-based cohort. Cancers (Basel). Apr 28, 2021;13(9):2133. [CrossRef] [Medline]
- Cortés-Ibañez FO, Belur Nagaraj S, Cornelissen L, Sidorenkov G, de Bock GH. A classification approach for cancer survivors from those cancer-free, based on health behaviors: analysis of the lifelines cohort. CANCERS (Basel). May 12, 2021;13(10):2335. [CrossRef] [Medline]
- Faruqui SHA, Du Y, Meka R, et al. Development of a deep learning model for dynamic forecasting of blood glucose level for type 2 diabetes mellitus: secondary analysis of a randomized controlled trial. JMIR Mhealth Uhealth. Nov 1, 2019;7(11):e14452. [CrossRef] [Medline]
- Guthrie NL, Carpenter J, Edwards KL, et al. Emergence of digital biomarkers to predict and modify treatment efficacy: machine learning study. BMJ Open. Jul 23, 2019;9(7):e030710. [CrossRef] [Medline]
- Hu L, Liu B, Li Y. Ranking sociodemographic, health behavior, prevention, and environmental factors in predicting neighborhood cardiovascular health: a Bayesian machine learning approach. Prev Med. Dec 2020;141:106240. [CrossRef] [Medline]
- Hu L, Liu B, Ji J, Li Y. Tree-based machine learning to identify and understand major determinants for stroke at the neighborhood level. J Am Heart Assoc. Nov 17, 2020;9(22):e016745. [CrossRef] [Medline]
- Luo L, Yuan J, Xu C, et al. Mental health issues and 24-Hour movement guidelines–based intervention strategies for university students with high-risk social network addiction: cross-sectional study using a machine learning approach. J Med Internet Res. Jun 13, 2025;27:e72260. [CrossRef] [Medline]
- Majcherek D, Kowalski AM, Lewandowska MS. Lifestyle, demographic and socio-economic determinants of mental health disorders of employees in the European countries. Int J Environ Res Public Health. Sep 21, 2022;19(19):11913. [CrossRef] [Medline]
- Mun EY, Geng F. Predicting post-experiment fatigue among healthy young adults: random forest regression analysis. Psychol Test Assess Model. Nov 8, 2019;61(4):471-493. [Medline]
- Park J, Edington DW. Application of a prediction model for identification of individuals at diabetic risk. Methods Inf Med. 2004;43(3):273-281. [Medline]
- Park S. Association of a high healthy eating index diet with long-term visceral fat loss in a large longitudinal study. Nutrients. Feb 2024;16(4):534. [CrossRef]
- Shi H, Fang Y, Ma X. Application of machine learning algorithms in osteoporosis analysis based on cardiovascular health assessed by life’s essential 8: a cross-sectional study. J Health Popul Nutr. May 29, 2025;44(1). [CrossRef]
- Xin Y, Ren X. Predicting depression among rural and urban disabled elderly in China using a random forest classifier. BMC Psychiatry. Feb 15, 2022;22(1):118. [CrossRef] [Medline]
- Zhang L, Zhao S, Yang Z, Zheng H, Lei M. An artificial intelligence platform to stratify the risk of experiencing sleep disturbance in university students after analyzing psychological health, lifestyle, and sports: a multicenter externally validated study. Psychol Res Behav Manag. 2024;17:1057-1071. [CrossRef] [Medline]
- Allen B. An interpretable machine learning model of cross-sectional U.S. county-level obesity prevalence using explainable artificial intelligence. PLOS ONE. 2023;18(10):e0292341. [CrossRef] [Medline]
- Morris MC, Moradi H, Aslani M, et al. Predicting incident cardiovascular disease among African-American adults: a deep learning approach to evaluate social determinants of health in the Jackson heart study. PLOS ONE. 2023;18(11):e0294050. [CrossRef] [Medline]
- Moon Y, Woo H. Key risk factors of generalized anxiety disorder in adolescents: machine learning study. Front Public Health. 2024;12:1504739. [CrossRef] [Medline]
- Cheung YK, Hsueh PYS, Qian M, et al. Are nomothetic or ideographic approaches superior in predicting daily exercise behaviors? METHODS Inf Med. 2017;56(6):452-460. [CrossRef] [Medline]
- Kim J, Jeong K, Lee S, Baek Y. Machine-learning model predicting quality of life using multifaceted lifestyles in middle-aged South Korean adults: a cross-sectional study. BMC Public Health. Jan 11, 2024;24(1):159. [CrossRef] [Medline]
- Pereira MG, Santos M, Magalhães R, Rodrigues C, Araújo O, Durães D. Burnout risk profiles in psychology students: an exploratory study with machine learning. Behav Sci (Basel). Apr 9, 2025;15(4):505. [CrossRef] [Medline]
- Ai M, Morris TP, Zhang J, et al. Resting-state MRI functional connectivity as a neural correlate of multidomain lifestyle adherence in older adults at risk for Alzheimer’s disease. Sci Rep. May 9, 2023;13(1):7487. [CrossRef] [Medline]
- Sandri E, Cerdá Olmedo G, Piredda M, Werner LU, Dentamaro V. Explanatory AI predicts the diet adopted based on nutritional and lifestyle habits in the Spanish population. Eur J Investig Health Psychol Educ. Jan 24, 2025;15(2):11. [CrossRef] [Medline]
- Birk N, Matsuzaki M, Fung TT, et al. Exploration of machine learning and statistical techniques in development of a low-cost screening method featuring the global diet quality score for detecting prediabetes in rural India. J Nutr. Oct 23, 2021;151(12 Suppl 2):110S-118S. [CrossRef] [Medline]
- Wallace ML, Buysse DJ, Redline S, et al. Multidimensional sleep and mortality in older adults: a machine-learning comparison with other risk factors. J Gerontol A Biol Sci Med Sci. Nov 13, 2019;74(12):1903-1909. [CrossRef] [Medline]
- Wang X, Zhang D, Lu L, et al. Development and validation of an explainable machine learning model for predicting the risk of sleep disorders in older adults with multimorbidity: a cross-sectional study. Front Public Health. 2025;13:1619406. [CrossRef] [Medline]
- Zhou RZ, Hu Y, Tirabassi JN, Ma Y, Xu Z. Deriving neighborhood-level diet and physical activity measurements from anonymized mobile phone location data for enhancing obesity estimation. Int J Health Geogr. Dec 30, 2022;21(1):22. [CrossRef] [Medline]
- Oladeji O, Zhang C, Moradi T, et al. Monitoring information-seeking patterns and obesity prevalence in Africa with internet search data: observational study. JMIR Public Health Surveill. Apr 29, 2021;7(4):e24348. [CrossRef] [Medline]
- Stemmer M, Parmet Y, Ravid G. Identifying patients with inflammatory bowel disease on Twitter and learning from their personal experience: retrospective cohort study. J Med Internet Res. Aug 2, 2022;24(8):e29186. [CrossRef] [Medline]
- Sathyanarayana A, Joty S, Fernandez-Luque L, et al. Sleep quality prediction from wearable data using deep learning. JMIR Mhealth Uhealth. Nov 4, 2016;4(4):e125. [CrossRef] [Medline]
- Chiang PH, Dey S. Offline and online learning techniques for personalized blood pressure prediction and health behavior recommendations. IEEE Access. 2019;7:130854-130864. [CrossRef]
- Zhong J, Liu W, Niu B, Lin X, Deng Y. Role of built environments on physical activity and health promotion: a review and policy insights. Front Public Health. 2022;10:950348. [CrossRef] [Medline]
- Kimura N, Aota T, Aso Y, et al. Predicting positron emission tomography brain amyloid positivity using interpretable machine learning models with wearable sensor data and lifestyle factors. Alz Res THERAPY. 2023;15(1). [CrossRef]
- The R project for statistical computing. R Foundation for Statistical Computing; 2022. URL: https://www.R-project.org/ [Accessed 2026-02-20]
- Wickham H, Averick M, Bryan J, et al. Welcome to the Tidyverse. JOSS. Nov 21, 2019;4(43):1686. [CrossRef]
- Groothuis-Oudshoorn K. mice: multivariate imputation by chained equations in R. J Stat Soft. 2011;45(3). [CrossRef]
- Kursa MB, Rudnicki WR. Feature selection with the Boruta package. J Stat Soft. 2010;36(11). [CrossRef]
- Kuhn M. Building predictive models in R using the caret package. J Stat Soft. 2008;28(5). [CrossRef]
- van der Laan MJ, Polley EC, Hubbard AE. Super learner. Stat Appl Genet Mol Biol. 2007;6(1). [CrossRef] [Medline]
- Lin S(. The “loneliness epidemic”, intersecting risk factors and relations to mental health help-seeking: a population-based study during COVID-19 lockdown in Canada. J Affect Disord. Jan 2023;320:7-17. [CrossRef]
- Berthold MR, Cebron N, Dill F, et al. KNIME - the Konstanz information miner: version 2.0 and beyond. SIGKDD Explor Newsl. Nov 16, 2009;11(1):26-31. [CrossRef]
- Castro R, Ribeiro-Alves M, Oliveira C, et al. What are we measuring when we evaluate digital interventions for improving lifestyle? A scoping meta-review. Front Public Health. 2021;9:735624. [CrossRef] [Medline]
- Pasquetto IV, Borgman CL, Wofford MF. Uses and reuses of scientific data: the data creators’ advantage. HDSR. 2019;1(2). URL: https://hdsr.mitpress.mit.edu/collection/af83430a [Accessed 2026-02-20] [CrossRef]
- Aldoseri A, Al-Khalifa KN, Hamouda AM. Re-thinking data strategy and integration for artificial intelligence: concepts, opportunities, and challenges. Appl Sci (Basel). Jun 13, 2023;13(12):7082. [CrossRef]
- Kim KH, Oh SW, Ko SJ, Lee KH, Choi W, Choi IY. Healthcare data quality assessment for improving the quality of the Korea Biobank Network. In: Meri A, editor. PLoS ONE. 2023;18(11):e0294554. [CrossRef] [Medline]
- Iqbal SMA, Mahgoub I, Du E, Leavitt MA, Asghar W. Advances in healthcare wearable devices. npj Flex Electron. 2021;5(1):9. [CrossRef]
- Mohanta B, Das P, Patnaik S. Healthcare 5.0: a paradigm shift in digital healthcare system using artificial intelligence, IOT and 5G communication. 2019. Presented at: 2019 International Conference on Applied Machine Learning (ICAML); May 25-26, 2019:191-196; Bhubaneswar, India. [CrossRef]
- Kamel Boulos MN, Koh K. Smart city lifestyle sensing, big data, geo-analytics and intelligence for smarter public health decision-making in overweight, obesity and type 2 diabetes prevention: the research we should be doing. Int J Health Geogr. Mar 3, 2021;20(1):12. [CrossRef] [Medline]
- Perez-Pozuelo I, Zhai B, Palotti J, et al. The future of sleep health: a data-driven revolution in sleep science and medicine. NPJ Digit Med. 2020;3(1):42. [CrossRef] [Medline]
- Reinertsen E, Clifford GD. A review of physiological and behavioral monitoring with digital sensors for neuropsychiatric illnesses. Physiol Meas. May 15, 2018;39(5):05TR01. [CrossRef] [Medline]
- Breiman L. Random forests. Mach Learn. Oct 2001;45(1):5-32. [CrossRef]
- Kuhn M, Johnson K. Feature Engineering and Selection: A Practical Approach for Predictive Models. CRC Press, Taylor & Francis Group; 2020. ISBN: 978-1-138-07922-9
- Thompson NC, Ge S, Manso GF. The importance of (exponentially more) computing power. arXiv. Preprint posted online on Jun 28, 2022. [CrossRef]
- Arora A, Alderman JE, Palmer J, et al. The value of standards for health datasets in artificial intelligence-based applications. Nat Med. Nov 2023;29(11):2929-2938. [CrossRef]
- Chawla NV, Bowyer KW, Hall LO, Kegelmeyer WP. SMOTE: Synthetic Minority Over-sampling Technique. jair. Jun 1, 2002;16:321-357. [CrossRef]
- Aliferis C, Overfitting SG. Underfitting and general model overconfidence and under-performance pitfalls and best practices in machine learning and AI. In: Simon GJ, Aliferis C, editors. Artificial Intelligence and Machine Learning in Health Care and Medical Sciences. Springer International Publishing; 2024:477-524. [CrossRef] ISBN: 978-3-031-39354-9
- James G, Witten D, Hastie T, Methods TRR. Resampling Methods An Introduction to Statistical Learning. Springer; 2021:197-223. [CrossRef] ISBN: 978-1-0716-1417-4
- Hyndman RJ, Athanasopoulos G. Forecasting: Principles and Practice. 2nd ed. OTexts; 2018. URL: https://otexts.org/fpp2/ [Accessed 2026-02-20]
- Saarela M, Podgorelec V. Recent applications of explainable AI (XAI): a systematic literature review. Appl Sci (Basel). Oct 2, 2024;14(19):8884. [CrossRef]
- Rodríguez-Pérez R, Bajorath J. Interpretation of compound activity predictions from complex machine learning models using local approximations and Shapley values. J Med Chem. Aug 27, 2020;63(16):8761-8777. [CrossRef] [Medline]
- Overview. H2O.ai. 2024. URL: https://docs.h2o.ai/h2o/latest-stable/h2o-docs/index.html [Accessed 2026-02-20]
- Hosker DK, Elkins RM, Potter MP. Promoting mental health and wellness in youth through physical activity, nutrition, and sleep. Child Adolesc Psychiatr Clin N Am. Apr 2019;28(2):171-193. [CrossRef] [Medline]
- Kris-Etherton PM, Sapp PA, Riley TM, Davis KM, Hart T, Lawler O. The dynamic interplay of healthy lifestyle behaviors for cardiovascular health. Curr Atheroscler Rep. Dec 2022;24(12):969-980. [CrossRef] [Medline]
- Kuhn M, Silge J. Tidy Modeling with R: A Framework for Modeling in the Tidyverse. 1st ed. O’Reilly; 2022. ISBN: 978-1-4920-9648-1
- Colaboratory BEG. Building Machine Learning and Deep Learning Models on Google Cloud Platform. Apress; 2019:59-64. [CrossRef] ISBN: 978-1-4842-4469-2
- Han S, Olonisakin TF, Pribis JP, et al. A checklist is associated with increased quality of reporting preclinical biomedical research: a systematic review. In: Boltze J, editor. PLoS ONE. 2017;12(9):e0183591. [CrossRef]
- Luo W, Phung D, Tran T, et al. Guidelines for developing and reporting machine learning predictive models in biomedical research: a multidisciplinary view. J Med Internet Res. Dec 16, 2016;18(12):e323. [CrossRef] [Medline]
- Al-Zaiti SS, Alghwiri AA, Hu X, et al. A clinician’s guide to understanding and critically appraising machine learning studies: a checklist for Ruling Out Bias Using Standard Tools in Machine Learning (ROBUST-ML). Eur Heart J Digit Health. Jun 2022;3(2):125-140. [CrossRef] [Medline]
- Stevens GA, Alkema L, Black RE, et al. Guidelines for accurate and transparent health estimates reporting: the GATHER statement. The Lancet. Dec 2016;388(10062):e19-e23. [CrossRef]
- Moritz S, Sardá A, Bartz-Beielstein T, Zaefferer M, Stork J. Comparison of different methods for univariate time series imputation in r. arXiv. Preprint posted online on Oct 13, 2015. [CrossRef]
- Mohsen F, Al-Absi HRH, Yousri NA, El Hajj N, Shah Z. A scoping review of artificial intelligence-based methods for diabetes risk prediction. NPJ Digit Med. Oct 25, 2023;6(1):197. [CrossRef] [Medline]
- Safi SK, Gul S. An enhanced tree ensemble for classification in the presence of extreme class imbalance. Mathematics. Oct 16, 2024;12(20):3243. [CrossRef]
- Velarde G, Weichert M, Deshmunkh A, et al. Tree boosting methods for balanced and imbalanced classification and their robustness over time in risk assessment. Intelligent Systems with Applications. Jun 2024;22:200354. [CrossRef]
- Caruana R, Niculescu-Mizil A. An empirical comparison of supervised learning algorithms. 2006. Presented at: Proceedings of the 23rd international conference on Machine learning - ICML ’06; Jul 24-28, 2006:161-168; Pittsburgh, PA. URL: http://portal.acm.org/citation.cfm?doid=1143844 [Accessed 2026-02-20] [CrossRef]
- Arnold C, Biedebach L, Küpfer A, Neunhoeffer M. The role of hyperparameters in machine learning models and how to tune them. PSRM. Oct 2024;12(4):841-848. [CrossRef]
- Arlot S, Celisse A. A survey of cross-validation procedures for model selection. Statist Surv. Preprint posted online on 2009. [CrossRef]
- Tsamardinos I, Greasidou E, Borboudakis G. Bootstrapping the out-of-sample predictions for efficient and accurate cross-validation. Mach Learn. Dec 2018;107(12):1895-1922. [CrossRef]
- Rainio O, Teuho J, Klén R. Evaluation metrics and statistical tests for machine learning. Sci Rep. Mar 13, 2024;14(1):6086. [CrossRef]
- Grandini M, Bagli E, Visani G. Metrics for multi-class classification: an overview. arXiv. Preprint posted online on Aug 13, 2020. [CrossRef]
- Lundberg S, Lee SI. A unified approach to interpreting model predictions. arXiv. Preprint posted online on May 22, 2017. URL: http://arxiv.org/abs/1705.07874 [Accessed 2024-04-22]
- Ribeiro MT, Singh S, Guestrin C. Model-agnostic interpretability of machine learning. arXiv. Preprint posted online on Jun 16, 2016. [CrossRef]
- Maksymiuk S, Gosiewska A, Biecek P. Landscape of R packages for explainable artificial intelligence. arXiv. Preprint posted online on Sep 24, 2020. [CrossRef]
- Ponce‐Bobadilla AV, Schmitt V, Maier CS, Mensing S, Stodtmann S. Practical guide to SHAP analysis: explaining supervised machine learning model predictions in drug development. Clinical Translational Sci. Nov 2024;17(11). URL: https://ascpt.onlinelibrary.wiley.com/toc/17528062/17/11 [Accessed 2026-02-20] [CrossRef]
- Lim H, Lee H. Eating habits and lifestyle factors related to childhood obesity among children aged 5-6 years: cluster analysis of panel survey data in Korea. JMIR Public Health Surveill. Apr 5, 2024;10:e51581. [CrossRef] [Medline]
Abbrevations
| AI: artificial intelligence |
| DASS: Depression Anxiety Stress Scale |
| FFQ: Food Frequency Questionnaire |
| FNN: fuzzy neural network |
| GPAQ: Global Physical Activity Questionnaire |
| HL: healthy lifestyle |
| IMS-PAQ: Indian Migration Study Physical Activity Questionnaire |
| INPLASY: International Platform of Registered Systematic Review and Meta-Analysis Protocols |
| IPAQ: International Physical Activity Questionnaire |
| LIME: local interpretable model-agnostic explanations |
| ML: machine learning |
| NN: neural network |
| NutSo-HH: Nutritional and Social Healthy Habits Scale |
| PRESS: Peer Review of Electronic Search Strategies |
| PRISMA-S: Preferred Reporting Items for Systematic Reviews and Meta-Analyses–Search extension |
| PRISMA-ScR: Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews |
| RF: random forest |
| SHAP: Shapley Additive Explanation |
| SL: supervised learning |
| SMOTE: synthetic minority oversampling technique |
| SQUASH: Short Questionnaire to Assess Health-Enhancing Physical Activity |
| SVM: support vector machine |
| UL: unsupervised learning |
| XAI: explainable artificial intelligence |
Edited by Andre Kushniruk; submitted 06.Jun.2025; peer-reviewed by Adewumi Adepoju, Vedamurthy Gejjegondanahalli Yogeshappa; final revised version received 06.Dec.2025; accepted 14.Dec.2025; published 27.Feb.2026.
Copyright© Tony Estrella, Lluis Capdevila, Carla Alfonso, Josep-Maria Losilla. Originally published in JMIR Human Factors (https://humanfactors.jmir.org), 27.Feb.2026.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Human Factors, is properly cited. The complete bibliographic information, a link to the original publication on https://humanfactors.jmir.org, as well as this copyright and license information must be included.

