- Research
- Open access
- Published:
Exploring the potential and limitations of deep learning and explainable AI for longitudinal life course analysis
BMC Public Health volume 25, Article number: 1520 (2025)
Abstract
Background
Understanding the complex interplay between life course exposures, such as adverse childhood experiences and environmental factors, and disease risk is essential for developing effective public health interventions. Traditional epidemiological methods, such as regression models and risk scoring, are limited in their ability to capture the non-linear and temporally dynamic nature of these relationships. Deep learning (DL) and explainable artificial intelligence (XAI) are increasingly applied within healthcare settings to identify influential risk factors and enable personalised interventions. However, significant gaps remain in understanding their utility and limitations, especially for sparse longitudinal life course data and how the influential patterns identified using explainability are linked to underlying causal mechanisms.
Methods
We conducted a controlled simulation study to assess the performance of various state-of-the-art DL architectures including CNNs and (attention-based) RNNs against XGBoost and logistic regression. Input data was simulated to reflect a generic and generalisable scenario with different rules used to generate multiple realistic outcomes based upon epidemiological concepts. Multiple metrics were used to assess model performance in the presence of class imbalance and SHAP values were calculated.
Results
We find that DL methods can accurately detect dynamic relationships that baseline linear models and tree-based methods cannot. However, there is no one model that consistently outperforms the others across all scenarios. We further identify the superior performance of DL models in handling sparse feature availability over time compared to traditional machine learning approaches. Additionally, we examine the interpretability provided by SHAP values, demonstrating that these explanations often misalign with causal relationships, despite excellent predictive and calibrative performance.
Conclusions
These insights provide a foundation for future research applying DL and XAI to life course data, highlighting the challenges associated with sparse healthcare data, and the critical need for advancing interpretability frameworks in personalised public health.
Introduction
The complexity of human health trajectories is shaped by the intricate interplay of biological, environmental, and social factors across the individuals’ lifespans [1]. Life course epidemiology seeks to understand these dynamics, revealing how early-life exposures, cumulative stressors, and critical developmental periods influence disease risk and health outcomes [2,3,4,5,6]. For instance, early-life adversities, such as childhood trauma or poverty, have been linked to adult-onset diseases, with the risk depending heavily on the timing, intensity and interplay with other exposures [7, 8]. With the increasing availability of large-scale longitudinal datasets, particularly those derived from national healthcare registers, researchers now have an unprecedented opportunity to examine health trajectories over the life course. These datasets are comprised of individual-level exposures recorded over a long time period, ranging from adverse childhood experiences (ACEs) to hospital records and education attainment, enabling the investigation of how risks accumulate and interact over time [1, 3]. However, traditional analytical methods, such as regression models and risk scoring, often fail to capture the full extent of the non-linear, multi-factorial relationships inherent in these processes [9, 10]. These methods rely on assumptions of linearity, independence, and population-level homogeneity, often leading to oversimplified interpretations that mask individual-level variations and temporal nuances [11]. As a result, the full richness of life course data remains underutilised, limiting the ability to inform targeted public health interventions.
Recent advancements in deep learning (DL) have allowed these models to overcome the limitations of traditional models in many fields. Within various healthcare applications, these models have proven to accurately predict disease risk for multiple outcomes using electronic health records (EHRs), including cardiovascular disease, cancer and diabetes [12,13,14]. They are particularly suited to these applications because of their ability to autonomously learn complex patterns and relationships directly from data, without requiring the pre-specification of their form or depending on strong assumptions [15]. Additionally, DL architectures, such as recurrent neural networks (RNNs) and convolutional neural networks (CNNs), are uniquely equipped to handle multivariate time-series (MTS) data, preserving temporal information that is often lost in traditional methods [16]. These qualities mean that DL models have the potential to transform life course epidemiology and the insights that can be gained from these vast datasets [17]. With the appropriate application of these methods, it may be possible to improve the accuracy of individual risk predictions, identify new risk factors, and find sensitive periods during which interventions are more effective. However, there is still a substantial amount of development required in order for them to be reliably and ethically implemented within life course epidemiology.
Whilst DL models have performed well in many other applications, there have been no studies examining the effectiveness of these models in life course epidemiology tasks. Moreover, although DL architectures designed for MTS data, such as RNNs and Transformers, have outperformed traditional methods in many contexts, some studies have found evidence of epidemiological settings where DL models perform worse [18, 19]. The application of DL to life course data remains limited, with significant gaps in understanding how these models perform under the unique constraints of longitudinal datasets, including sparsity and confounding. Many DL models have become state-of-the-art in specific applications, however, there is limited knowledge on which architectures are best suited to life course analysis. For instance, InceptionTime has performed well in physiological signal classification, such as ECG and EEG analysis [20, 21], and Transformers have revolutionised natural language processing (NLP) by achieving superior performance in tasks such as machine translation and text generation [22, 23]. While some models excel in capturing certain data characteristics, such as sparsity or temporal complexity, their relative performance can vary widely across different datasets and tasks. Understanding which DL models are most appropriate for life course analysis remains a critical gap in the field.
Another critical challenge is the interpretability of DL models, which often function as “black boxes”, with limited insights into the specific temporal or inter-variable relationships that they identify [24]. Recent advances in explainable artificial intelligence (XAI) aim to address this issue, with methods like Shapley additive explanations (SHAP) emerging as popular tool [25,26,27,28,29]. Methods such as SHAP provide post hoc explanations by quantifying the contribution of each input feature to a model’s predictions [25]. In life course epidemiology, this capability is particularly valuable for identifying critical risk factors and key exposure periods that influence health outcomes, providing actionable insights for public health [27]. SHAP is now one of the most widely used XAI methods in healthcare with a recent surge in studies using it to identify potential risk factors and disease mechanisms [30,31,32]. However, XAI methods remain limited in their ability to establish causal relationships, an issue that is especially pronounced in time-series data where confounding and sparsity complicate interpretability [33]. Within epidemiology, establishing causal relationships is considered essential and is one of the significant knowledge gaps restricting the use of DL within this field [34]. Without established causal approaches for DL models, many researchers use XAI methods like SHAP to identify the potential causal mechanisms, even though the SHAP values only relate to the predictive influence of exposures. Misinterpreting predictive markers as causal mechanisms poses significant risks, underscoring the need for rigorous evaluation of XAI and how it is related to the underlying causal structures present in life course applications.
There are several key ways in which longitudinal life course data sets differ from MTS data. One particular challenge is the sparsity which is a defining characteristic of many life course datasets. Sparse data-characterised by irregular sampling, rare events, and high proportions of missing or zero entries-pose significant challenges for traditional methods. For example, ACEs or exposures to rare diseases often occur infrequently, creating datasets with high sparsity. While many DL models were initially developed for tasks like NLP or computer vision, many models have been adapted for application to MTS data which is distinct from life course data [35]. Although DL models are frequently claimed to excel in handling such sparsity, there is little systematic evidence supporting this advantage. In fact, sparsity is considered to be one of the key limitations preventing the wider use of DL methods in healthcare, alongside model opacity and data heterogeneity [36]. Instead, most studies on DL performance in sparse datasets are case-specific and lack generalisability, leaving a critical gap in the literature.
This study addresses these gaps by systematically evaluating the utility of DL models and XAI methods for life course analysis for the first time. Through a series of simulation experiments, we compare the ability of four state-of-the-art DL models, including MLSTM-FCN, LSTMAttention, ResNet and InceptionTime, to handle varying levels of sparsity, exposure complexity, and confounding. These models are benchmarked against traditional methods, including logistic regression and XGBoost, to provide a comprehensive understanding of their strengths and limitations [37]. Additionally, we use SHAP to examine the consistency and validity of model explanations, comparing SHAP-derived insights to the ground truth causal mechanisms embedded in the simulations.
Our findings represent three key contributions to the advancement of life course epidemiology. First, we evaluate whether there is a universal best-performing DL model for analysing different possible relationships in life course data or if model performance is inherently scenario-dependent. Second, this study evaluates the reliability of SHAP as an interpretability tool, highlighting its limitations in aligning with causal relationships and cautioning against its misuse for identifying disease mechanisms or risk factors. Third, it investigates the capability of DL models to handle sparse healthcare data, providing evidence for their effectiveness in contexts where traditional methods often fail. The findings not only offer practical guidance for researchers but also contribute to the broader discussion on integrating causality, interpretability, and predictive accuracy in AI-driven public health solutions. This work aims to advance the responsible and effective use of AI in healthcare, paving the way for a new DL framework for life course analysis that has the power to improve our understanding of disease risk trajectories.
Methods
This section outlines the methodological approach to simulating data as well as the parameter tuning and model fitting procedures. Further technical details are provided in the supplementary materials for reproducibility.
Data simulation
We used the simcausal [38] R package to generate a longitudinal dataset with a Directed Acyclic Graph (DAG) designed to reflect multiple exposure occurring over time as well as realistic temporal and inter-variable dependencies. The use of the simcausal package ensures interpretability by maintaining an explicit causal structure. While alternative methods, such as simulated data generation using Generative Adversarial Networks, are capable of replicating real-world datasets with high fidelity, they often lack the transparency and generalisability required for the analysis of causal mechanisms, which is central to this study. Parameters for the DAG were derived from a subset of the DANLIFE cohort as established in Davies et al. to ensure that the simulated data captures the important relationships present in life course data [39]. The simulated dataset includes 100,000 individuals, followed from birth until their 16 th birthday, incorporating both time-dependent (dynamic) and time-independent (stationary) exposures. Specifically, three time-independent covariates (maternal age at birth, parental diabetes status, and parental ethnicity) and three time-dependent exposures (counts of ACEs as defined by Rod et al. [8]) were included.
The simulated cohort differs from the original DANLIFE cohort in the way ACEs are represented, the original cohort includes a single ACE variable that represents fitted trajectory groups which is replaced by the three time series variables (Loss, SES and Dynamic) that were used by Davies et al. to create the trajectory groups in the simulated cohort, thereby allowing a study of temporal patterns. The stationary variables were simulated using categorical distributions with probabilities to match the DANLIFE cohort, given in Table S1. The time-dependent variables were modelled as zero-inflated Poisson distributions that depended on the values of the time-independent variables and their value at the previous time point. This methodology generated individual-level trajectories over time, enabling the exploration of complex interactions among variables throughout the lifespan. To evaluate how well the synthetic data reflects the real-world DANLIFE cohort, we have included detailed statistical comparisons in the supplementary materials in Figure S5.
Multiple binary outputs \(y\) were generated from the input data \(X\) using custom functions, designed to replicate four Life Course Patterns (LCPs) common within the literature: period, repeats, order, and timing (Table 1). The World Health Organization originally conceptualised LCPs to categorise how different exposures at various life stages may affect health [40]. The period conceptual model is based on research showing that there are certain time periods during which exposure events have greater impact on future health [3]. For instance, childhood exposure to passive smoking is more detrimental to lung function than the same exposure occurring in adulthood [41]. Critical periods represent periods during which exposures must occur to give rise to certain health consequences, such as the irreversible effects of alcohol exposure during the first trimester of pregnancy on the child’s organ development, leading to lifelong impacts on learning, behaviour, and predisposition to various chronic diseases [42]. In contrast, sensitive periods increase the likelihood of poor outcomes but do not guarantee them and other factor outside this period may also impact outcome likelihood. Furthermore, the accumulation-of-risk model illustrates how repeated exposures can compound risk over time, such as chronic poverty exacerbating respiratory issues like asthma [3, 43]. Another key life course paradigm is the timing of exposures. The relative timing of events can magnify their health impacts, for example, experiencing multiple adversities, such as unemployment and illness within a short period, may compound negative outcomes [44]. Finally, the temporal ordering of exposures has a substantial impact on the health trajectory [45]. For instance, an individual’s likelihood of a heart attack is greater if they have experienced depression previously [46].
For each LCP, three separate outputs were generated that embodied the pattern, varying in complexity to evaluate model performance across increasingly challenging scenarios with the group name reflecting this. To ensure that the simulated outcomes embodied the challenges found in healthcare studies, we maintained a class imbalance similar to that of the DANLIFE cohort, with approximately 2.5% of outcomes classified as positive [39]. The rules used to generate the outcomes are deterministic and therefore, noise was added to the train set to better imitate realistic scenarios. Specifically, ten percent of positive outcomes were randomly switched to negative outcomes, while an equal number of individuals originally classified with negative outcomes were reclassified as positive. The test set remained unaltered, meaning that models that have successfully learnt the underlying pattern can theoretically correctly diagnose all individuals and achieve perfect performance.
Our approach adopts a supervised learning perspective, focusing on modelling the functional relationship between the input data \(X\) and the target variable \(y\). The objective is to learn a mapping function \(f\) such that \(y = f(X) + \epsilon\), where \(\epsilon\) denotes the noise or error term. It is important to note that, by design, the precise nature of the input data \(X\) has limited impact on the model performance and it is instead solely the relationship between the input data and disease outcome that drives the results. In this way, our results are applicable to many diverse healthcare datasets and contexts.
Model architectures
We compared the predictive performance of six models on each of the multiple simulated outcomes: a baseline logistic regression (LR) model, a tree-based ML model (Extreme Gradient Boosting [XGBoost] [37]), and four DL architectures tailored for time series analysis. The DL models included: the Multivariate Long Short-Term Memory Fully Convolutional Networks (MLSTM-FCN [47]), which integrates LSTM networks for sequential learning with fully convolutional networks for capturing spatial patterns; ResNet [47], a deep residual network designed to address vanishing gradient issues while learning hierarchical representations; InceptionTime [48], a convolutional neural network optimised for multivariate time-series data utilising inception modules to capture features at various scales; and LSTMAttention [49], which employs an attention mechanism to emphasise long-range dependencies in sequential data. Each model architecture is explained in detail in the supplementary material.
To handle class imbalances effectively, various strategies were implemented to enhance model performance while minimising the risk of the model prioritising the larger negative class. The LR model employed class balancing, whereas the XGBoost model adjusted class weights in accordance with their distribution in the loss function. For the DL models, we adopted the focal loss function to prioritise the learning of harder-to-classify individuals, assigning greater weight to the minority class. The model weights were trained from scratch. Further details regarding the model architectures and fitting procedures can be found in the supplementary materials, with Tables S2 and S3 giving the hyperparameters optimised for the XGBoost and DL models, respectively.
Model fitting & hyperparameter optimisation
Hyperparameter optimisation was conducted using the optuna package, to allow for fine-tuning of parameters such as learning rate, batch size, weight decay, and model-specific parameters (Tables S2 and S3 in the supplementary materials contain the parameters that were optimised as well as their ranges for XGBoost and the DL models, respectively). To ensure robust evaluation, we employ stratified \(k\)-fold cross-validation (with \(k = 2\)), maintaining class balance within each fold. For the ML models, the Tree-structured Parzen Estimator (TPE) sampler was employed to efficiently explore the hyperparameter space. Unpromising trials were pruned at each epoch, allowing computational resources to be focused on the most promising configurations. An early stopping mechanism was implemented to mitigate overfitting, with training halted based on validation loss improvement over a specified number of epochs, preserving the best-performing model weights. The dataset was divided into training (80%) and test (20%) sets, with each model trained for a maximum of 50 epochs.
Once hyperparameters were selected to maximise the average precision score (AP - an approximation of the area under the precision-recall curve [AUPRC]), final training was conducted using 90% of the train set, with the remaining data reserved for early stopping. Model performance was assessed using various metrics, including F1-score, area under the receiver operating characteristic curve (AUROC), and AUPRC [50, 51]. The AUPRC is particularly relevant for imbalanced datasets, as it captures the model’s ability to distinguish between the positive and negative classes [52]. Additionally, the Brier score was employed to evaluate calibration quality, providing insight into the degree of over- and under-prediction, with lower scores being preferable. A well-calibrated model should have predicted probabilities that are a true reflection of the likelihood of an event occurring, and a poorly calibrated model can be harmful to decision making [53].
Experimental setup
The performance of LR, XGBoost, and multiple DL models (InceptionTime, LSTMAttention, MLSTM-FCN, and ResNet) was evaluated across the four LCPs: Period, Repeats, Order, and Timing (Table 1). Each LCP represents a different type of temporal pattern, testing the models’ abilities to predict binary outcomes under challenging conditions, particularly given the high class imbalance (only 2.5% positive outcomes) and the presence of 10% noise. Each model’s performance is described using F1 scores, AUPRC, AUROC, and Brier scores to assess model discrimination, precision-recall trade-offs, and calibration.
The second part of the analysis examines explainability using the SHAP methodology which quantifies the relative influence of each exposure on the model’s predictions, they are individual-level and have temporal specificity, which means it is possible to not only examine which exposures are important but also at which points in time they exert the strongest influence. The individual-level SHAP values are first studied, followed by population-level marginal SHAP beeswarm plots to investigate how closely the SHAP values align with the simulated causal trends.
Results
Comparison of model performance for different LCPs
Figure 1 shows the model performance of the six models within each LCP measured by the AUPRC score, with higher values indicating better predictive performance. Firstly, the figure shows that LR demonstrates a sharp decline in AUPRC as the LCP pattern complexity increases and incorporates relative temporal and inter-variable dependencies, reflecting its limited capacity to handle non-linear and temporal patterns. Similarly, XGB also exhibits a drop in performance with increasing complexity, albeit less steep than LR. Both of these models perform comparably to the DL models for the Period LCP but fall behind once the LCPs become dynamic. Secondly, across the four LCPs, there is no single DL model that consistently outperforms the rest with DL performance exhibiting variability depending on the underlying causal structure. Another point of note is that the LSTMAttention model performs poorly for the Timing LCP overall, demonstrating vulnerability despite its excellent performance on the other LCPs. The model performance measured by the AUC score is shown in Figure S6.
Predictive performance of models for each of the four LCPs (Period, Repeats, Order and Timing). Model performance, measured by AUPRC scores, shown for each of the three patterns that fall within each LCP. All the AUPRC scores that resulted in each of the three different seeds used to control model training and hyperparameter optimisation are shown. Boxes show the interquartile range, midpoints depict the median values and whiskers show the minimum and maximum values excluding outliers. LR = logistic regression, XGB = XGBoost, MLF = MLSTM-FCN, IT = InceptionTime, LSTMA = LSTMAttention
The different evaluation metric values averaged across the three random seeds are displayed in Table 2. Further threshold-dependent evaluation metric values are provided in the supplementary materials (Table S4). Higher scores reflect greater discriminative performance for all metrics except for the Brier score where lower values indicate better model calibration. There are several instances where there is a big discrepancy between model performance when measured by the AUROC score compared to the AUPRC and F1 scores. For example, LR frequently achieved high AUROC values when the associated AUPRC scores were very low, for example, the AUPRC scores for Order3 and Timing1 were 17.12% and 26.59%, respectively, whilst the AUROC exceeded 90%.
For the Period LCPs, all models have exceptionally good performance for the simpler patterns, with AUROC and AUPRC scores approaching 100%. Notably, LSTMAttention showed perfect performance for rules Period1 and Period2 which typify critical and sensitive periods, with all metrics, including Brier scores, indicating a robust fit to the data. All other models, including LR and XGBoost, also excelled in discriminative performance and had Brier scores indicating good calibration, although MLSTM-FCN showed a slightly elevated Brier score of 4.10 for rule Period1. As the complexity of Period patterns increased from Period1 to Period3, differences between models became evident. LR, although it performed comparably for the other rules, struggled particularly with the Period3 rule which incorporates different weights for the impact of an exposure according to the time when it occurred, achieving an AUPRC of only 89.26%. Moreover, the AUPRC values for the MLSTM-FCN (95.47%) and InceptionTime (97.65%) were lower than XGBoost (99.22%) as well as all other DL models (99.73–99.89%).
The Repeats patterns show substantial disparities in model performance. LR consistently underperformed compared to the other models, with AUPRC values consistently below 52% and as low as 30.78% for rule Repeats2. In contrast, LSTMAttention, MLSTM-FCN and ResNet performed exceptionally well. LSTMAttention reached an AUPRC of 100% for rule Repeats1, reflecting perfect precision-recall balance, and maintained an AUROC of 100% across all rules. MLSTM-FCN and ResNet also demonstrated high predictive performance, with AUPRC scores above 98% and Brier scores indicating strong calibration. InceptionTime followed closely behind in terms of predictive performance but the associated Brier scores were markedly higher than the other models with values as high as 3.93% for Repeats3. In contrast, the performance of XGBoost varied, achieving strong metrics for simpler rules (rule Repeats1, AUPRC 99.98%) but experiencing greater declines in performance for complex patterns, giving F1 scores as low as 90.07% as seen in rule Repeats3.
Order patterns, which involve the sequential ordering of events, further highlighted differences in model performance. InceptionTime stood out with near-perfect scores, particularly for rule Order3, where it had an AUPRC of 99.98% and a very low Brier score of 0.02. LSTMAttention similarly performed well, with an AUROC and AUPRC of 100% for rule Order1. On the other hand, LR exhibited significant weaknesses, with AUPRC values falling to 24.72% and 17.12% for rules Order2 and Order3, respectively, despite maintaining AUROC scores above 90%. Higher Brier scores accompanied the poor predictive performance with values as high as 4.2% for Order1. ResNet and MLSTM-FCN scored consistently high metrics with AUPRC scores above 97%, while XGBoost displayed reduced performance for more complex order rules, such as rule Order2 (AUPRC 90.63% and F1 score 86.39%), particularly in precision-recall metrics.
The Timing patterns were also challenging for LR, with rule Timing1 showing particularly poor outcomes, as evidenced by an AUPRC of only 26.59% and a Brier score of 2.67, indicating poor calibration and difficulty in modelling the timing of events. Generally, DL models outperformed LR significantly. However, LSTMAttention, while having the highest performance for the other LCPs, struggled with rules Timing2 and Timing3, achieving AUPRC scores of 75.01% and 59.82%, respectively, despite AUROC scores greater than 97%. ResNet maintained high performance with all AUPRC values greater than 99.67% and low Brier scores, suggesting excellent calibration. InceptionTime and MLSTM-FCN both performed very well, followed by XGBoost which had slightly lower AUPRC scores but markedly lower F1 scores that dropped to 93.81% for the rule Timing3.
Explainability
Figure 2 shows population-level beeswarm plots for the LCP Order3 marginalised over time and the features, separately. This pattern generates a positive outcome if at least one Dynamic and Loss exposure event occur with the first Dynamic exposure event preceding the first Loss exposure event. The beeswarm plot marginalised over time shows that the majority of the models identify Dynamic and Loss variables as being the most influential on model predictions. All models show similar patterns with lower values of these features, i.e. non-occurrences, being more likely to have negative SHAP values meaning that positive predictions depend on the occurrence of these exposures. The pattern is most pronounced for XGBoost, followed by ResNet and MLSTM-FCN. Notably, InceptionTime shows very limited variation in SHAP values, unlike the other models, and the pattern of lower SHAP values for lower feature values is very slight. These discrepancies show that the models do not agree on the most influential factors. Additionally, the SHAP values for individuals whose parents have diabetes (high feature value) appear to negatively influence model predictions for LSTMAttention alone. The second beeswarm plot, showing SHAP values marginalised over the features, has no clear discernable trend. It is notable that, again, XGBoost has the greatest spread in the magnitude of SHAP values, however it is not possible to discern any temporal dependencies. Therefore, whilst it is possible to use these plots to gain insight into the influence of the occurrence of multiple exposure events, these population-level plots alone do not reflect dynamic causal relationships.
SHAP Beeswarm Plots for LCP Order1. The top panel illustrates SHAP values marginalised over time for each feature, grouped by model, with features ordered by mean absolute SHAP values. The bottom panel visualises SHAP values marginalised over features for each age, also grouped by model. Feature values are colour-coded, ranging from low (“Low”) to high (“High”). LR = logistic regression, XGB = XGBoost, MLF = MLSTM-FCN, IT = InceptionTime, LSTMA = LSTMAttention
In some very simplified scenarios, it is possible to correctly guess the causal mechanism from the population-level beeswarm plots alone. Figure S7 in the supplementary materials, shows the marginalised beeswarm SHAP plots for the Period3 pattern; a simple linear pattern that assigns greater weights to earlier Loss exposures in calculating the likelihood of a positive outcome. The beeswarm plot marginalised over time clearly picks out the Loss feature as the only influential feature and the lower SHAP values associated with lower feature values suggests that the greater number of Loss events increases the likelihood of a positive outcome. The beeswarm SHAP plot marginalised over time links greater feature values at younger ages with higher SHAP values, and the magnitude of the SHAP values gets lower with age. Due to the linear nature of this LCP, it is possible to piece these two pieces of information together to link the occurrence of more Loss exposures at younger ages with the outcome which is consistent with the causal pathways in the generated data. However, the marginalised SHAP beeswarm plots for the non-linear patterns are not as easily interpretable.
Figure 3 shows the individual-level data and related SHAP values for two individuals (one positive and one negative). The SHAP values for the positive individual have large positive values associated with when the Dynamic and Loss exposure events occur. This shows that these two events have the greatest influence on the positive model prediction. It is also possible to see that the non-occurrence of Dynamic and Loss events, particularly early Dynamic events and late Loss events, have small negative impact on the model prediction. This is most noticeable with XGBoost, which also has the SHAP values of the highest magnitude, followed by ResNet. Similarly, the negative individual also has the loss exposure event increasing the likelihood of a positive outcome. However, each model has multiple negative SHAP values related to the non-occurrence of any Dynamic events prior to the Loss exposure event, resulting in the correct negative model prediction. These patterns directly align with the causal pathways in the simulated data. Finally, notice that the marginal plots obscure the temporal relationships as the SHAP values all have small magnitudes.
SHAP Comparison for Individuals for LCP Order1. This figure compares SHAP values for two individuals, one positive for the outcome of interest and one negative. The top row includes individual feature heatmaps and SHAP values marginalised over features. The bottom two rows depict the individual-level SHAP values marginalised over time for each individual split by model. SHAP values are shown on a consistent scale across all plots, with colour gradients indicating magnitude and direction (blue to red). LR = logistic regression, XGB = XGBoost, MLF = MLSTM-FCN, IT = InceptionTime, LSTMA = LSTMAttention
Using individual-level SHAP values to identify the temporal relationships is also possible for other LCPs. For instance, the Timing1 LCP also has causal patterns that are easy to discern from individuals. Figure S8 in the supplementary material shows that the negative individual has positive SHAP values associated with the occurrence of the single Loss exposure event indicating that this event increases the likelihood of positive classification. However, this is offset by an equal or greater negative SHAP value related to the non-occurrence of a SES event in the same year. The magnitude of the SHAP values vary greatly between the different models. The positive individual has large positive SHAP values when the Loss and SES events co-occur. As in the negative case, some models, such as XGBoost, assign positive SHAP values to the additional occurrences of SES and Loss exposure events and corresponding negative values to the lack of the other. Moreover, several of the DL models assign small SHAP values to the other exposure events in ways that do not relate to the underlying causal relationships.
There are other LCPs and individuals for which is is significantly harder to discern the causal relationships from the individual-level SHAP values, where the SHAP values bare little resemblance to the underlying simulated mechanisms. Figure 3 in the supplementary material shows the SHAP values and individual-level data for the Order3 LCP, where a positive outcome is generated if there is no SES exposure event and a Dynamic exposure occurs prior to a Loss exposure. Some models, primarily XGBoost and ResNet, have negative SHAP values aligning with the SES exposure event, correctly identifying its existence as acting against a positive outcome. However, this SHAP value for many models is approaching zero and incorrectly indicating that this exposure has little influence on model predictions. For the positive individual, the SHAP values related to the SES feature are all slightly positive at the non-occurrence of any exposure events, showing that this is important in generating a positive outcome. Moreover, the positive SHAP values associated with the single Loss exposure event suggest that this event is increasing the likelihood of the positive outcome, although the magnitude exhibits substantial variation between models. However, the Dynamic exposure events are assigned a variety of different SHAP values by each of the models. InceptionTime associates all Dynamic events with negative SHAP values, indicating that this event occurring negatively influencing model predictions, when the opposite is true. XGBoost has both positive and negative SHAP values for the different Dynamic events which is also contrary to the simulated pattern, as only the first Dynamic exposure event is in the LCP mechanism. None of the models correctly identify the significance of this feature in the underlying causal relationship.
Discussion
This study provides the first systematic evaluation of DL models and XAI methods for life course epidemiology. While DL is widely used in healthcare research, its application to life course data remains underexplored. Our findings demonstrate that DL models outperform traditional epidemiological approaches, such as LR and XGBoost, in identifying LCPs, particularly those involving complex temporal and intervariable dependencies between the exposures and the outcome. XGBoost performed well for simpler LCPs, however it was outperformed by DL models as pattern complexity increased, supporting evidence that DL models are better suited for dynamic, high-dimensional data [54,55,56], despite XGBoost’s advantages in tabular datasets [57]. One key finding of this study is the absence of a universal best-performing DL model, reinforcing the importance of selecting models according to the specific datasets [16]. These findings underscore the importance of context-aware model selection to maximise predictive performance in life course research.
The findings of this study build on existing research demonstrating the strengths of DL for longitudinal data analysis in healthcare [36, 58, 59]. However, most prior research has focused on EHRs rather than life course data, which presents additional challenges including sparse exposure histories and long-term dependencies. While DL models appear to be well-suited to handling sparsity [35, 36], empirical validation has been limited. Our results confirm that DL models can handle these complexities better than traditional methods.
Explainability & causality
Despite their predictive advantages, DL models present interpretability challenges, particularly in epidemiological contexts where causal inference is a key objective. This study evaluated SHAP, a widely used post-hoc XAI method, to assess how well SHAP explanations align with causal patterns in life course exposures. While SHAP performed well for simple LCPs, it became increasingly misaligned as the simulated exposure-disease relationships became more complex, often identifying variables inconsistent with the true simulated causal pathways. These findings support broader concerns that post-hoc XAI methods struggle with high-dimensional, temporally structured data [24, 25, 30, 31]. Recent research has raised concerns about the sensitivity of SHAP to correlated variables and confounding [33, 34], with studies observing discrepancies between SHAP explanations and true causal mechanisms in biomedical datasets [28]. Our results reinforce these concerns, highlighting the need for alternative XAI techniques that better align with epidemiological principles.
Several alternative XAI methods could have been considered for this study [29, 60]; however, each has limitations when applied to longitudinal health data. Gradient-based methods such as Integrated Gradients (IG), Deep Learning Important Features (DeepLIFT), and Layer-Wise Relevance Propagation (LRP) provide feature importance scores via backpropagation [61, 62] but struggle with capturing feature interactions and generalising across architectures [63, 64]. Additionally, these methods lack temporal modelling, making them less suitable for life course analysis. Model-agnostic approaches, such as Local Interpretable Model-Agnostic Explanations (LIME) and Anchors, approximate model behaviour with simpler interpretable models [26, 65, 66], but their sensitivity to parameter choices and instability in high-dimensional datasets limit their reliability. Attention-based techniques can highlight influential time points in sequential data, but may introduce biases from patterns in the training data [22].
Shapley Interaction Quantification (SHAPIQ) is a recently developed approach that offers a promising alternative by refining feature importance estimates in the presence of correlated exposures and evolving dependencies over time, making it particularly valuable for longitudinal epidemiology [67]. Emerging causal-aware methods, such as CausalSHAP and counterfactual explanations, aim to provide causal insights but often require predefined causal structures, which may not be feasible in observational health studies [68, 69]. In contrast, SHAPIQ offers a flexible alternative that does not require an explicit causal graph. Our findings suggest that SHAPIQ could improve interpretability in complex epidemiological models, particularly in situations involving correlated exposures, temporal dependencies, and confounding.
As DL and XAI methods become increasingly integrated into healthcare decision-making and policy, careful consideration of their limitations is essential. Misinterpreting model outputs-such as equating predictive markers with causal significance-can lead to poorly designed interventions that risk exacerbating health inequalities or introducing unintended harms. Clinicians and policymakers must recognise that XAI tools like SHAP do not inherently provide causal explanations. Our findings underscore the need for rigorous validation, transparency, and appropriate methodological safeguards when applying these models in public health, ensuring that policy decisions are not based on incorrect assumptions about causality
Study limitations and future research
Several limitations of this study should be acknowledged. First, while our use of simulated datasets enabled controlled evaluations of model performance and interpretability, it also has inherent limitations. Simulated data may not fully capture the complexities of real-world life course data, which are often subject to measurement error, unmeasured confounding, and selection bias. As a result, the relative simplicity of the simulated data may have contributed to the somewhat elevated model performance observed in this study. Future research should extend our findings by evaluating DL models on empirical life course datasets to assess their performance under real-world conditions. Second, although we focused on SHAP due to its widespread adoption, other XAI methods warrant further investigation. Epidemiological applications frequently require causal insights as well as insights into variables influential on predictive performance. Existing XAI tools primarily identify predictive markers rather than causal relationships, limiting their applicability for causal inference in life course research. Therefore, an approach such as SHAPIQ or CausalSHAP may be suitable for this area, with the potential to improve interpretability by aligning feature attributions with underlying causal structures [70]. Future research should focus on developing or adapting XAI methods for longitudinal data and life course research, by integrating causal inference principles and explicitly accounting for both temporal dependencies and complex feature interactions. Third, the high computational demands of DL models, particularly for hyperparameter tuning and training, present practical constraints for large-scale epidemiological studies. These computational requirements may limit the feasibility of implementing DL methods in resource-constrained settings and any changes made to simplify the model training pipeline may lead to reduced performance and utility.
Conclusions
This study has several important implications for both life course epidemiology and healthcare systems as a whole. First, with healthcare systems increasingly digitising and register-based datasets expanding, there is an unprecedented opportunity to leverage DL models to uncover key risk factors for a wide range of health conditions. These models can facilitate a deeper understanding of how the timing, intensity, and cumulative impact of exposures experienced throughout the life course influence long-term health outcomes. By identifying and analysing these complex patterns, DL methods have the potential to uncover novel relationships that might otherwise remain hidden, enabling more precise insights into disease mechanisms. This capability is particularly important for advancing precision public health, where tailored interventions can be designed based on a more nuanced understanding of how different risk factors interact over time. In this regard, our findings suggest that when life course patterns are present in the data, DL models show the most promise in unravelling these intricate relationships. By integrating time-sensitive and cumulative risk factors, DL could open new pathways for designing interventions and clinical prediction models that are more responsive to individual health trajectories and more sensitive to the timing of exposures. This offers the potential to significantly improve public health interventions, enabling them to be more context-specific and effective.
This study demonstrates the potential of DL and XAI to advance life course epidemiology. Our findings suggest that while DL models can capture complex exposure-disease relationships, model selection should be context-dependent. Moreover, we caution against over-reliance on post-hoc XAI methods for deriving causal insights and emphasise the need for continued innovation in XAI tools that consider causal principles. By addressing these challenges, DL and XAI methods have the potential to transform life course epidemiology, offering novel insights into disease risk trajectories and informing more effective public health interventions.
Data availability
Data is provided within the manuscript or supplementary information files.
References
Maret-Ouda J, Tao W, Wahlin K, Lagergren J. Nordic registry-based cohort studies: Possibilities and pitfalls when combining Nordic registry data. Scand J Public Health. 2017;45:14–9.
Jones NL, Gilman SE, Cheng TL, Drury SS, Hill CV, Geronimus AT. Life Course Approaches to the Causes of Health Disparities. Am J Public Health. 2019;109(Suppl 1):S48.
Kuh D, Ben-Shlomo Y, Lynch J, Hallqvist J. Life course epidemiology. J Epidemiol Community Health. 2003;57:778–83.
Elder GH, Johnson MK, Crosnoe R. The Emergence and Development of Life Course Theory. Boston: Springer US; 2003. pp. 3–19.
Ben-Shlomo Y, Kuh D. A life course approach to chronic disease epidemiology: conceptual models, empirical challenges and interdisciplinary perspectives. Int J Epidemiol. 2002;31(2):285–93.
Kalmakis KA, Chandler GE. Adverse childhood experiences: towards a clear conceptual meaning. J Adv Nurs. 2014;70(7):1489–501.
Davey Smith G, Hart C, Blane D, Hole D. Adverse socioeconomic conditions in childhood and cause specific adult mortality: prospective observational study. BMJ. 1998;316:1631–1263.
Rod NH, Bengtsson J, Elsenburg LK, Taylor-Robinson D, Rieckmann A. Hospitalisation patterns among children exposed to childhood adversity: a population-based cohort study of half a million children. Lancet Public Health. 2021;6(11):e826–35.
Hutchison ED. An update on the relevance of the life course perspective for social work. Fam Soc. 2019;100(4):351–66.
Pollet TV, Stulp G, Henzi SP, Barrett L. Taking the aggravation out of data aggregation: A conceptual guide to dealing with statistical issues related to the pooling of individual-level observational data. Am J Primatol. 2015;77(7):727–40.
Smith BJ, Smith ADAC, Dunn EC. Statistical Modeling of Sensitive Period Effects Using the Structured Life Course Modeling Approach (SLCMA). Cham: Springer; 2021.
Weng SF, Reps J, Kai J, Garibaldi JM, Qureshi N. Can machine-learning improve cardiovascular risk prediction using routine clinical data? PLoS ONE. 2017;12(4):1–14.
Swanson K, Wu E, Zhang A, Alizadeh AA, Zou J. From patterns to patients: Advances in clinical machine learning for cancer diagnosis, prognosis, and treatment. Cell. 2023;186:1772–91.
Iparraguirre-Villanueva O, Espinola-Linares K, Castañeda ROF, Cabanillas-Carbonell M. Application of Machine Learning Models for Early Detection and Accurate Classification of Type 2 Diabetes. Diagnostics. 2023;13:2383.
Aisenbrey S, Fasang AE. New life for old ideas: The “second wave’’ of sequence analysis bringing the “course’’ back into the life course. Sociol Methods Res. 2010;38(3):420–62.
Ismail Fawaz H, Forestier G, Weber J, Idoumghar L, Muller PA. Deep learning for time series classification: a review. Data Min Knowl Discov. 2019;33(4):917–63.
Chen S, Yu J, Chamouni S, Wang Y, Li Y. Integrating machine learning and artificial intelligence in life-course epidemiology: pathways to innovative public health solutions. BMC Med. 2024;22:354.
Salganik MJ, Lundberg I, Kindel AT, Ahearn CE, Al-Ghoneim K, Almaatouq A, et al. Measuring the predictability of life outcomes with a scientific mass collaboration. Proc Natl Acad Sci. 2020;117(15):8398–403.
Bjerre L, Peixoto C, Alkurd R, Talarico R, Abielmona R. Comparing AI/ML approaches and classical regression for predictive modeling using large population health databases: Applications to COVID-19 case prediction. Glob Epidemiol. 2024;8:100168.
Borra D, Andalo A, Severi S, Corsi C. On the Application of Convolutional Neural Networks for 12-lead ECG Multi-label Classification Using Datasets from Multiple Centers. Comput Cardiol. 2020. https://doiorg.publicaciones.saludcastillayleon.es/10.22489/CinC.2020.349.
Kastrati A, Plomecka MB, Wattenhofer R, Langer N. Using deep learning to classify saccade direction from brain activity. In ACM Symposium on Eye Tracking Research and Applications, ETRA. New York: Association for Computing Machinery; 2021. p. 1–6.
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, et al. Attention is All you Need. In: Guyon I, Luxburg UV, Bengio S, Wallach H, Fergus R, Vishwanathan S, et al., editors. Advances in Neural Information Processing Systems. vol. 30. Newry: Curran Associates, Inc.; 2017.
Devlin J, Chang MW, Lee K, Google KT, Language AI. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Minneapolis: Association for Computational Linguistics; 2019. p. 4171–86.
Doshi-Velez F, Kim B. Towards A Rigorous Science of Interpretable Machine Learning. arXiv preprint arXiv:1702.08608, 2017.
Lundberg SM, Lee SI. A Unified Approach to Interpreting Model Predictions. Adv Neural Inf Process Syst. 2017;2017-Decem:4766–75.
Lundberg SM, Erion G, Chen H, DeGrave A, Prutkin JM, Nair B, et al. From local explanations to global understanding with explainable AI for trees. Nat Mach Intell. 2020;2:56–67.
Fauvel K, Lin T, Masson V, Élisa Fromont, Termier A. XCM: An Explainable Convolutional Neural Network for Multivariate Time Series Classification. Mathematics. 2020;9.
Jin D, Sergeeva E, Weng WH, Chauhan G, Szolovits P. Explainable deep learning in healthcare: A methodological survey from an attribution view. WIREs Mech Dis. 2022;14:e1548.
Saluja R, Malhi A, Knapič S, Främling K, Cavdar C. Towards a Rigorous Evaluation of Explainability for Multivariate Time Series. arXiv preprint arXiv:2104.04075. 2021.
Ibrahim L, Mesinovic M, Yang KW, Eid MA. Explainable Prediction of Acute Myocardial Infarction Using Machine Learning and Shapley Values. IEEE Access. 2020;8:210410–7.
Chia AHT, Khoo MS, Lim AZ, Ong KE, Sun Y, Nguyen BP, et al. Explainable machine learning prediction of ICU mortality. Inform Med Unlocked. 2021;25:100674.
Westerlund AM, Hawe JS, Heinig M, Schunkert H. Risk Prediction of Cardiovascular Events by Exploration of Molecular Data with Explainable Artificial Intelligence. Int J Mol Sci. 2021;22:10291.
Amann J, Vetter D, Blomberg SN, Christensen HC, Coffee M, Gerke S, et al. To explain or not to explain?-Artificial intelligence explainability in clinical decision support systems. PLoS Digit Health. 2022;1(2):1–18.
Ghassemi M, Naumann T, Schulam P, Beam AL, Chen IY, Ranganath R. A Review of Challenges and Opportunities in Machine Learning for Health. AMIA Summits Transl Sci Proc. 2020;2020:191.
Ruiz AP, Flynn M, Large J, Middlehurst M, Bagnall A. The great multivariate time series classification bake off: a review and experimental evaluation of recent algorithmic advances. Data Min Knowl Discov. 2021;35:401–49.
Xie F, Yuan H, Ning Y, Ong MEH, Feng M, Hsu W, et al. Deep learning for temporal data representation in electronic health records: A systematic review of challenges and methodologies. J Biomed Inform. 2022;126:103980.
Chen T, Guestrin C. XGBoost: A Scalable Tree Boosting System. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. KDD ’16. New York: Association for Computing Machinery; 2016. pp. 785–94.
Sofrygin O, van der Laan MJ, Neugebauer R. simcausal R Package: Conducting Transparent and Reproducible Simulation Studies of Causal Effect Estimation with Complex Longitudinal Data. J Stat Softw. 2017;81(2):1–47.
Davies M, van Houten CS, Bengtsson J, Elsenburg LK, Kragelund Nielsen K, Andersen GS, et al. Childhood adversity and the risk of gestational diabetes: A population-based cohort study of nulliparous pregnant women. Diabet Med. 2024;41(1):e15242. E15242 DME-2023-00361.R1.
World Health Organization, International Longevity Centre-UK. The Implications for Training of Embracing a Life Course Approach to Health. Geneva: World Health Organization; 2000.
Grant T, Brigham EP, McCormack MC. Childhood Origins of Adult Lung Disease as Opportunities for Prevention. J Allergy Clin Immunol Pract. 2020;8(3):849.
Chung DD, Pinson MR, Bhenderu LS, Lai MS, Patel RA, Miranda RC. Toxic and Teratogenic Effects of Prenatal Alcohol Exposure on Fetal Development, Adolescence, and Adulthood. Int J Mol Sci. 2021;22(16):8785.
Nikiéma B, Gauvin L, Zunzunegui VM, Séguin L. Longitudinal patterns of poverty and health in early childhood: Exploring the influence of concurrent, previous, and cumulative poverty on child health outcomes. BMC Pediatr. 2012;12(1):1–13.
Dupre ME, George LK, Liu G, Peterson ED. The Cumulative Effect of Unemployment on Risks for Acute Myocardial Infarction. Arch Intern Med. 2012;172(22):1731–7.
Mishra GD, Cooper R, Kuh D. A life course approach to reproductive health: theory and methods. Maturitas. 2010;65(2):92–7.
May HT, Horne BD, Knight S, Knowlton KU, Bair TL, Lappé DL, et al. The association of depression at any time to the risk of death following coronary artery disease diagnosis. Eur Heart J Qual Care Clin Outcome. 2017;3(4):296–302.
Wang Z, Yan W, Oates T. Time Series Classification from Scratch with Deep Neural Networks: A Strong Baseline. Proc IJCNN. 2016;2017-May:1578–85.
Ismail Fawaz H, Lucas B, Forestier G, Pelletier C, Schmidt DF, Weber J, et al. InceptionTime: Finding AlexNnet for time series classification. Data Min Knowl Discov. 2020;34(6):1936–62.
Zerveas G, Jayaraman S, Patel D, Bhamidipaty A, Eickhoff C. A Transformer-based Framework for Multivariate Time Series Representation Learning. Proc ACM SIGKDD Int Conf Knowl Discov Data Min. 2021;11(21):2114–24.
Qi Q, Luo Y, Xu Z, Ji S, Yang T. Stochastic Optimization of Areas Under Precision-Recall Curves with Provable Convergence. Adv Neural Inf Process Syst. 2021;3:1752–65.
Huang J, Ling CX. Using AUC and accuracy in evaluating learning algorithms. IEEE Trans Knowl Data Eng. 2005;17(3):299–310.
Davis J, Goadrich M. The relationship between Precision-Recall and ROC curves. In: Proc. ICML. 2006. pp. 233–40.
Van Calster B, McLernon DJ, van Smeden M, Wynants L, Steyerberg EW, Topic Group ‘Evaluating diagnostic tests and prediction models’ of the STRATOS initiative. Calibration: the Achilles heel of predictive analytics. BMC Med. 2019;17(1):230.
Shmueli G. To Explain or to Predict? Stat Sci. 2010;25(3):289–310.
Miotto R, Wang F, Wang S, Jiang X, Dudley JT. Deep learning for healthcare: review, opportunities and challenges. Brief Bioinforma. 2017;19(6):1236–46.
Alex Quistberg D, Mooney SJ, Tasdizen T, Arbelaez P, Nguyen QC. Invited Commentary: Deep Learning - Methods to Amplify Epidemiological Data Collection and Analyses. Am J Epidemiol. 2024; 194(2):322–6
Shwartz-Ziv R, Armon A. Tabular Data: Deep Learning is Not All You Need. Inf Fusion. 2022;81:84–90
Rajkomar A, Oren E, Chen K, Dai AM, Hajaj N, Hardt M, et al. Scalable and accurate deep learning with electronic health records. NPJ Digit Med. 2018;1(1):1–10.
Ismail AA, Gunady MK, Bravo HC, Feizi S. Benchmarking Deep Learning Interpretability in Time Series Predictions. Adv Neural Inf Process Sys. 2020;33:6441–52.
Simonyan K, Vedaldi A, Zisserman A. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps. arXiv preprint arXiv:1312.6034. 2013.
Shrikumar A, Greenside P, Kundaje A. Learning Important Features Through Propagating Activation Differences. In: International conference on machine learning, PMIR. 2017;3145–53
Montavon G, Binder A, Lapuschkin S, Samek W, Müller KR. Layer-Wise Relevance Propagation: An Overview. Samek W, Montavon G, Vedaldi A, Hansen LK, Müller KR, editors. Explainable AI: Interpreting, Explaining and Visualizing Deep Learning. vol. 11700. Cham: Springer International Publishing; 2019. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/978-3-030-28954-6_10.
Szczepankiewicz K, Popowicz A, Charkiewicz K, Nałęcz-Charkiewicz K, Szczepankiewicz M, Lasota S, et al. Ground truth based comparison of saliency maps algorithms. Sci Rep. 2023;13:16887.
Adebayo J, Gilmer J, Muelly M, Goodfellow IJ, Hardt M, Kim B. Sanity Checks for Saliency Maps. Adv Neural Inf Process Systems. 2018;31.
Ribeiro MT, Singh S, Guestrin C. “Why Should I Trust You?”: Explaining the Predictions of Any Classifier. In: NAACL-HLT 2016 - 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Demonstrations Session. San Diego; 2016. p. 97–101.
Ribeiro MT, Singh S, Guestrin C. Anchors: high-precision model-agnostic explanations. In: Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence and Thirtieth Innovative Applications of Artificial Intelligence Conference and Eighth AAAI Symposium on Educational Advances in Artificial Intelligence. AAAI’18/IAAI’18/EAAI’18. New Orleans: AAAI Press; 2018.
Muschalik M, Baniecki H, Fumagalli F, Kolpaczki P, Hammer B, Hüllermeier E. shapiq: Shapley Interactions for Machine Learning. In: The Thirty-eight Conference on Neural Information Processing Systems Datasets and Benchmarks Track. Vancouver; 2024.
Heskes T, Sijben E, Bucur IG, Claassen T. Causal Shapley Values: Exploiting Causal Knowledge to Explain Individual Predictions of Complex Models. Adv Neural Inf Process Sys. 2020;33:4778–89.
Wachter S, Mittelstadt BD, Russell C. Counterfactual Explanations without Opening the Black Box: Automated Decisions and the GDPR. Harv JL Tech. 2017;31:841.
Tronchin L, Cordelli E, Celsi LR, Maccagnola D, Natale M, Soda P, et al. Translating Image XAI to Multivariate Time Series. IEEE Access. 2024;12:27484–500.
Acknowledgements
Not applicable.
Funding
S.B., H.J.T.U. and H.C. acknowledge funding from the MRC Centre for Global Infectious Disease Analysis (reference MR/X020258/1), funded by the UK Medical Research Council (MRC). This UK funded award is carried out in the frame of the Global Health EDCTP3 Joint Undertaking. S.B. is funded by the National Institute for Health and Care Research (NIHR) Health Protection Research Unit in Modelling and Health Economics, a partnership between UK Health Security Agency, Imperial College London and LSHTM (grant code NIHR200908). Disclaimer: “The views expressed are those of the author(s) and not necessarily those of the NIHR, UK Health Security Agency or the Department of Health and Social Care.”. S.B. acknowledges support from the Novo Nordisk Foundation via The Novo Nordisk Young Investigator Award (NNF20OC0059309) which previously supported S.M and supports H.C. SB acknowledges the Danish National Research Foundation (DNRF160) through the chair grant which also supports N.S.. S.B. acknowledges support from The Eric and Wendy Schmidt Fund For Strategic Innovation via the Schmidt Polymath Award (G- 22 - 63345). M.D. acknowledges funding from the Independent Research Fund Project (1030 - 00171B). S.M. was funded via S.B. via The Novo Nordisk Young Investigator Award. S.M. acknowledges support from the National Research Foundation, Singapore, under its NRF FELLOWSHIP (NRF-NRFF15 - 2023 - 0010). N.H.R. acknowledges funding from the Danish Research Council (reference1030 - 00171B). N.H.R. acknowledges funding from the European Research Council (ERC) consolidator grant (no. 101124807). The Copenhagen Health Complexity Center is funded by TrygFonden. S.F. acknowledges EPSRC funding (EP/V002910/2). A.K. acknowledges support from the Novo Nordisk Foundation via The Novo Nordisk Young Investigator Award (NNF20OC0059309). For the purpose of open access, the author has applied a ‘Creative Commons Attribution’ (CC BY) licence to any Author Accepted Manuscript version arising from this submission.
Author information
Authors and Affiliations
Contributions
S.B., S.M. and H.C. conceived and designed the study. H.J.T.U. and H.C. conceived the experiments. H.C., S.M., H.J.T.U., S.B., and N.S. conducted the experiments. H.C., S.B. and H.J.T.U. wrote the original draft. H.C., N.S., A.K., M.D., N.H.R., S.F., S.M., S.B. and H.J.T.U. all reviewed the manuscript and contributed to its scientific interpretation.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Coupland, H., Scheidwasser, N., Katsiferis, A. et al. Exploring the potential and limitations of deep learning and explainable AI for longitudinal life course analysis. BMC Public Health 25, 1520 (2025). https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12889-025-22705-4
Received:
Accepted:
Published:
DOI: https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12889-025-22705-4