Personalized Disease Risk Prediction from Multimodal Health Data Using Large Language Models
Hanieh
Arjmand
1✉
Emailhanieh.arjmand@gmail.com
Alexandre
Tomberg
1
1
Lydia AI (Knowtions Research Inc.)
161 Bay Street, Suite 2700
M5J 2S1
Toronto
ON
Canada
Hanieh Arjmand (corresponding author: hanieh.arjmand@gmail.com), Alexandre Tomberg
Lydia AI (Knowtions Research Inc.), 161 Bay Street, Suite 2700 Toronto, ON M5J 2S1, Canada
Abstract
A
This study presents a system designed to enhance disease prediction accuracy by integrating electronic health records (EHR) and wearable device data. EHR provides structured health information, such as hospital visits and diagnostic codes, while wearable devices capture real-time physiological data, like step count, offering insights into behavior patterns and activity levels. Traditionally, these data sources have been studied independently, resulting in limited utilization of their combined potential and inconsistencies in predictions. Our system uses a multimodal approach in which specialized encoders process each data type separately, and integrate extracted features into a shared embedding space. This common embedding is leveraged by large language models (LLMs) to capture meaningful relationships between health records and activity patterns, allowing the system to predict disease risks with increased accuracy and relevance to individual patient characteristics. The system’s risk scores, generated end-to-end, can be further utilized by LLMs to provide personalized recommendations based on individual risk profiles. This multimodal approach not only addresses the challenges of integrating disparate data formats but also provides a holistic view of patient health, identifying subtle trends that may be missed when using EHR or wearable data in isolation. The result is a robust, comprehensive framework for proactive and personalized healthcare.
Keywords:
Large Language Models (LLMs)
Electronic Health Records (EHR)
Wearable data
Multimodal learning
Disease prediction
Health data integration
Patient risk assessment
Dynamic health tracking
A
Introduction:
Predictive analytics with the integration of Electronic Health Records (EHR) and wearable data stands at the forefront of a transformative shift in healthcare [1]. This approach not only enhances disease prediction for individual patients but also scales to provide insights for public health monitoring and epidemiology. At the heart of this transformation is the potential use of advanced models, like foundational Large Language Models (LLMs), that can analyze complex, multimodal data to reveal underlying health risks [2–4].
A
The use of EHR and wearable data integration for predictive analytics relies on the ability to ingest and interpret complex data from multiple sources. Wearable devices offer continuous, real-time monitoring of physiological parameters, such as heart rate, activity, and sleep quality, while EHRs provide a longitudinal view of a patient’s health, including lab results, diagnoses, and clinical notes [
1]. Integrating these data streams offers a unique opportunity to build predictive models that can identify at-risk patients early, enabling timely and personalized interventions. By combining the continuous flow of real-time wearable data with the depth of historical EHR information, predictive analytics provides a comprehensive health profile for each patient, enhancing proactive healthcare models that emphasize prevention.
The development of foundational LLMs has shown promise for various healthcare applications, including disease prediction. These models have the capability to process and synthesize multimodal, individual-specific data, ranging from textual clinical notes to time-series data from wearables, giving them a nuanced understanding of a patient's potential health risks [2]. LLMs can be fine-tuned to adapt their responses based on a patient’s risk profile, integrating contextual data such as lab values, genetic predispositions, and other indicators [2–7]. This approach goes beyond traditional models by leveraging the extensive prior medical knowledge encoded within LLMs, which can identify connections between conditions, such as recognizing that "hypertension" and "high blood pressure" are synonymous and understanding the clinical significance of various lab ranges.
Unlike classical predictive models, which often require explicit coding and complex data handling to incorporate prior knowledge, LLMs offer a more flexible and adaptive framework. Through prompt engineering, they can dynamically incorporate new information, making them highly versatile across a wide range of tasks and domains. Additionally, LLM-based architectures are well-suited for handling diverse and incomplete healthcare datasets. Traditional models often struggle with missing data, requiring imputation or exclusion. In contrast, LLM-based systems can work effectively with partial inputs by focusing on available information, without needing architectural modifications. This makes them ideal for modeling heterogeneous data sources such as real-time wearable signals, structured EHRs, and even unstructured clinical notes.
Predictive analytics using LLMs still faces the complex challenge of integrating diverse health data types. Health data is often stored in varied formats and modalities, from numerical lab results to visual data such as medical imaging. Traditional methods have long sought to address this through multimodal fusion techniques, such as logistic regression models that integrate several data inputs or fusion architectures combining features from disparate sources. However, these approaches often lack the flexibility and scalability of LLMs. For instance, traditional models require cumbersome methods to incorporate prior knowledge or to deal with fuzzy, non-standardized labels [2]. The challenge lies in how to efficiently integrate these data into a format that LLMs can process while retaining its richness and clinical relevance. In healthcare, crucial indicators of disease often span multiple data modalities, making it necessary to combine wearable sensor readings with diagnostic information for accurate prediction.
Recent advances in multimodal LLMs have explored integrating complex data types such as images and time-series signals, though applications in healthcare are still emerging. This strategy enriches the model's contextual capacity, as shown in recent studies where models like HealthAlpaca improve predictions by combining contextual and physiological data [3, 8]. Another example is HeLM, a framework that uses modality-specific encoders to combine clinical data such as demographic and laboratory data, with spirogram curves, which are used to evaluate lung function, for disease risk prediction [2]. Through modality specific encoders, HeLM creates unified representations suitable for LLM processing, demonstrating the potential of large models to handle diverse biomedical inputs.
Building on similar principles, we use modality-specific encoders tailored for more complex inputs, longitudinal EHR and wearable data, to generate unified representations for LLM-based prediction. This enables integration of detailed medical history with real-time physiological patterns to support dynamic health modeling. Figure 1 shows a high-level overview of our proposed system, which integrates structured EHR data, wearable sensor data, and optionally, other available information through modality-specific encoders and an LLM for personalized disease prediction. Our approach is akin to enhancing a language model's vocabulary with new, specialized words and retraining it to understand them in context. Here, the EHR and wearable embeddings act as newly introduced "words" that enrich the model's understanding of health-related information. By embedding these novel data types into the shared embedding space, we effectively retrain the model to interpret these embeddings with the same familiarity as its existing knowledge base. This allows the system to capture complex interactions and nuanced meanings between clinical history and real-time physiological signals, thereby deepening the model's predictive and contextual capabilities. This process empowers the system to "learn" from both structured EHR data and dynamic activity patterns, much like adding new vocabulary terms expands a language model's understanding and contextual relevance.
Methods:
For this study, we utilized data from the UK Biobank, a large-scale biomedical database that contains in-depth genetic and health information on approximately 500,000 participants from across the United Kingdom [9]. This extensive resource provides a rich dataset for health research, including longitudinal EHR and accelerometer data, enabling comprehensive analysis of both historical health events and real-time movement patterns. For each individual, the dataset included a sequence of diagnosis codes across health events, capturing a detailed timeline of clinical diagnoses, treatments, and health outcomes. In addition, accelerometer data was collected for a subset of participants over a continuous one-week period. The accelerometer data was preprocessed to extract minute-by-minute step counts, forming a structured time series representation of each participant’s activity. In total, we identified 69,326 individuals with both activity and EHR data available meeting the criteria for our model training and evaluation. Among these, approximately 3% (or 2,193 participants) had recorded mortality events.
Our diagnostic system utilizes a multimodal, end-to-end training approach designed to integrate EHR and wearable time-series data to improve disease prediction. By combining historical health records with real-time physiological data, the system leverages both long-term and immediate health indicators, capturing a comprehensive view of patient health. The method employs distinct encoding models: one tailored to the structured nature of EHR data, capturing longitudinal health patterns, and another optimized for wearable data, identifying activity-based trends relevant to disease risk. The latent representations from each encoder are then integrated into a shared embedding space, along with contextual information, to enhance prediction accuracy. This section details each component of the system, from input data processing to disease probability prediction, demonstrating how each element contributes to a robust, adaptive model for personalized disease assessment, as illustrated in Fig. 2.
1.
Input Data Processing: The system takes both structured EHR data and time-series wearable activity data as inputs. These are processed by separate encoders tailored for each data type.
2.
EHR Data Encoding: Our EHR encoder builds on the Med-BERT architecture, capturing patterns in diagnosis codes within structured health records [10]. While independently pre-trained, we used the same pretraining tasks as Med-BERT, such as masked code prediction and prolonged length of stay prediction, which help the model capture essential temporal and contextual patterns within diagnosis sequences. This approach enables our encoder to effectively model longitudinal EHR data for predictive insights. Although we utilized this specific encoder, the same approach can be achieved with alternative architectures, such as TAPER, CEHR-BERT, BEHRT, as long as the encoder is capable of extracting relevant clinical information and incorporating temporal details of patient visits into its representations [11–13].
3.
Wearable Data Encoding: The activity encoder is a 1D CNN network with fully connected layers, originally trained to predict mortality using step sequences per minute. Following training, the layer preceding the classification layer was adjusted in dimensions to function as the encoder, enabling it to capture temporal patterns in activity data. This configuration allows the model to recognize trends in activity levels (e.g., periods of high or low activity) that may relate to health outcomes, creating a robust time-series encoding suited to wearable data.
4.
Extension to Additional Modalities: While our system currently integrates EHR and wearable activity data as its core components, the architecture is designed with inherent flexibility to incorporate additional data modalities for a more comprehensive understanding of patient health and behavior. For example, nutritional data can be integrated by introducing a dedicated encoder, such as an MLP, specifically designed to process structured dietary information. This encoder would capture patterns in nutritional intake that may influence disease risk, allowing the model to learn from dietary behaviors alongside clinical and activity data. By employing this modular approach, additional encoders for other relevant data types can be seamlessly added to the system, enhancing its ability to capture complex health patterns and improving the accuracy and personalization of disease predictions.
5.
Encoding and Latent Representations: Each encoder produces latent representations of the input data, which condense the raw information into lower-dimensional vectors focused on key health indicators. The EHR encoder’s latent representation reflects complex relationships in clinical data, while the activity encoder’s representation captures wearable-derived health patterns. These representations are designed to highlight features that are most predictive of disease outcomes and provide an efficient encoding of the raw data’s essential characteristics.
6.
Integration of Additional Information: Additional contextual information, such as demographic data, text prompts, or specific instructions, can be incorporated directly to further personalize disease prediction. This auxiliary information is encoded using embeddings from a pre-trained LLM, generating embedding vectors that capture nuances in the extra data. For example, instructions like “provide yes/no as the model output” or demographic details are embedded as vectors using an LLM. These embeddings are then combined with the EHR and activity latent representations by replacing the embeddings of designated placeholder tokens, creating a comprehensive embedding that integrates all relevant data sources.
7.
Disease Prediction: The integrated embedding is passed through an LLM, which generates a textual output representing the prediction, such as “yes” or “no” for binary classification tasks. In this study, we used OutreAI/Lite-Mistral-150M-v2-Instruct as an example LLM; however, the framework is model-agnostic and can be adapted to use other language models. The model was trained to predict mortality risk, but it can be easily extended to other disease prediction tasks by replacing the mortality labels with labels for the disease of interest. The logits for each class are computed from the LLM’s output and are used to compute the loss for model optimization.
The model is trained end-to-end using a weighted binary cross-entropy loss function, designed to handle class imbalances often present in disease prediction tasks. This loss function adjusts the penalty for misclassification by applying higher weights to the minority class (e.g., true disease cases), encouraging the model to learn from both positive and negative cases effectively. Training proceeds until the loss function converges. During training, we monitor AUROC (Area Under the Receiver Operating Characteristic) in addition to minimizing loss, as it is a critical metric for evaluating model performance in clinical prediction. This approach helps ensure the model performs well across key metrics as training progresses.
The model is trained jointly on both EHR and wearable data. To evaluate the contribution of each modality, we assess performance under three test-time configurations: (1) using both modalities, (2) using only EHR data, and (3) using only wearable data. We also compare the model’s performance to baseline models, as detailed in the Results section.
Results:
This section presents the model's performance under three test configurations: (1) using both EHR and wearable activity data, (2) using only EHR data (with activity data replaced by zeros), and (3) using only activity data (with EHR data replaced by zeros). Additionally, these results are compared to single-modality encoder models (Med-BERT and Activity Encoder) each paired with a simple classifier and trained specifically on EHR or activity data alone to assess the standalone predictive power of each data modality.
Configuration 1: EHR and Activity Data (Full Model)
When evaluated with both EHR and activity data, the model achieved the results shown in Table 1.
Table 1
Performance of the full multimodal model using both EHR and wearable activity data.
|
Dataset
|
Accuracy
|
AUROC
|
True Positives
|
True Negatives
|
False Positives
|
False Negatives
|
|
Training
|
0.8590
|
0.8945
|
1231
|
40454
|
6531
|
312
|
|
Validation
|
0.8442
|
0.8334
|
219
|
8560
|
1518
|
102
|
|
Test
|
0.8476
|
0.8536
|
234
|
8580
|
1490
|
95
|
These high accuracy and AUROC values demonstrate strong predictive performance, indicating the benefit of combining both EHR and wearable data to capture comprehensive health insights.
Configuration 2: EHR Data Only (Activity Data Replaced by Zeros)
In this setup, the activity data was replaced with zeros, effectively isolating the EHR data's contribution. The results are shown in Table 2.
Table 2
Performance of the model using EHR data only (activity data zeroed out).
|
Dataset
|
Accuracy
|
AUROC
|
True Positives
|
True Negatives
|
False Positives
|
False Negatives
|
|
Training
|
0.8431
|
0.8895
|
1244
|
39669
|
7316
|
299
|
|
Validation
|
0.8285
|
0.8265
|
226
|
8390
|
1688
|
95
|
|
Test
|
0.8291
|
0.8483
|
236
|
8386
|
1684
|
93
|
These results indicate that EHR data alone provides substantial predictive power, achieving accuracy and AUROC values close to the full model’s performance. This outcome underscores EHR data's importance as a primary source of health insights.
Configuration 3: Activity Data Only (EHR Data Replaced by Zeros)
In this configuration, EHR data was replaced with zeros, leaving only wearable activity data for predictions. The model performance is shown in Table 3.
Table 3
Performance of the model using activity data only (EHR data zeroed out).
|
Dataset
|
Accuracy
|
AUROC
|
True Positives
|
True Negatives
|
False Positives
|
False Negatives
|
|
Training
|
0.5857
|
0.6633
|
994
|
27431
|
19554
|
549
|
|
Validation
|
0.5876
|
0.6725
|
206
|
5904
|
4174
|
115
|
|
Test
|
0.5854
|
0.6455
|
205
|
5883
|
4187
|
124
|
Baseline Encoder-Only Performance (Single-Modality Models)
To contextualize the above results, individual encoder models (Med-BERT and Activity Encoder) were evaluated with simple MLP layers on their latent spaces. This approach provides baseline performance metrics for each modality used independently, offering a point of comparison for the multimodal configurations.
Table 4
Performance of the Med-BERT encoder using only EHR data.
|
Dataset
|
Accuracy
|
AUROC
|
True Positives
|
True Negatives
|
False Positives
|
False Negatives
|
|
Training
|
0.8420
|
0.8915
|
1238
|
39622
|
7363
|
305
|
|
Validation
|
0.8292
|
0.8240
|
214
|
8409
|
1669
|
107
|
|
Test
|
0.8340
|
0.8440
|
237
|
8436
|
1634
|
92
|
The Med-BERT encoder shows relatively high accuracy and AUROC, as shown in Table 4, which are comparable to the results of the model evaluated with only EHR data in the multimodal setup (Table 2). This result highlights the strength of EHR data as a robust predictor on its own.
The activity encoder alone performs similarly to the activity-only configuration of the multimodal model, as shown in Table 5 and Table 3, respectively. Despite lower accuracy and AUROC than the EHR modality, it demonstrates the utility of activity data in providing predictive insights, particularly for dynamic health changes.
Table 5
Performance of the activity encoder using only wearable activity data.
|
Dataset
|
Accuracy
|
AUROC
|
True Positives
|
True Negatives
|
False Positives
|
False Negatives
|
|
Training
|
0.6366
|
0.6700
|
732
|
23970
|
13604
|
494
|
|
Validation
|
0.6390
|
0.6711
|
308
|
10299
|
5776
|
217
|
|
Test
|
0.6312
|
0.6675
|
260
|
8450
|
4913
|
177
|
While the accuracy and AUROC values for activity data alone are lower than those for EHR data, they still demonstrate the unique predictive contributions of wearable data, especially for capturing dynamic physiological patterns.
Discussion:
In this work, we developed a multimodal LLM-based framework that integrates EHR and wearable data for disease prediction. The model was trained jointly on both data types and evaluated under different input configurations to assess the contribution of each modality. We also compared the model's performance against single-modality baselines to better understand the predictive strength of each input source. The results highlight key findings regarding model performance across these configurations and the independent predictive power of each modality.
The model achieved the highest accuracy and AUROC when combining both EHR and activity data, underscoring the advantage of integrating historical health records with real-time physiological signals. This multimodal integration provides a more comprehensive and nuanced assessment of disease risk than either data source alone. By capturing both long-term health patterns and real-time activity, the model offers a fuller view of patient health. For example, a patient with cardiovascular indicators in their EHR may also exhibit low physical activity and irregular sleep in their wearable data. The encoders process these modalities separately and integrate them to produce a cardiovascular risk prediction, which can be used by an LLM to generate tailored recommendations such as increased activity or dietary adjustments.
The model’s architecture demonstrates resilience by adapting to different data availability scenarios. A notable finding is that when either EHR or activity data is zeroed out, the model’s performance aligns closely with that of single-modality encoder models. This suggests that each data type retains its predictive power independently, even within a multimodal framework. The model’s ability to preserve performance when only partial data is available highlights its adaptability, a practical advantage in healthcare where complete data may not always be available.
The EHR-only configuration yielded relatively high accuracy and AUROC values, comparable to the full model’s performance. This result suggests that EHR data provides foundational predictive insights due to its comprehensive historical information. The model’s strong performance with EHR alone underscores the critical role of structured health records in disease prediction.
Although activity data alone produced lower accuracy and AUROC compared to EHR data, it still provided valuable predictive insights, capturing real-time physiological and behavioral patterns. The performance of the model evaluated with only activity data closely matches that of the activity-specific encoder model, confirming the wearable data's distinct contributions. To further quantify its impact, we analyzed the differences in predictions between the full multimodal system (Configuration 1: EHR and Activity Data) and the EHR-only system (Configuration 2). Specifically, we investigated whether the changes in the full model's predictions relative to the EHR-only model aligned with the predictions from the activity-only system (Configuration 3). Our analysis revealed that 79.8% of cases exhibited this alignment, indicating that the changes in predictions from the full model (both increases and decreases) corresponded with the activity-only predictions. This suggests that the activity data meaningfully contributes to the adjustments in risk predictions in the full system, either enhancing or attenuating the predicted risk based on the activity patterns. Additionally, we observed a moderately strong positive correlation (r = 0.771) between the prediction differences (Configuration 1 vs. Configuration 2) and the predictions from the activity-only system (Configuration 3). This correlation underscores the significant role that activity data plays in enriching the model's predictive capabilities when combined with EHR data.
Through the end-to-end training approach, the EHR encoder and the activity encoder learn collaboratively, with each encoder adapting to shared features to provide complementary health insights. By leveraging these advanced encoders, the model effectively combines clinical and activity data with additional contextual inputs, resulting in a robust, accurate, and personalized system for disease prediction.
Our system can also incorporate other available information and patient‑specific factors, such as genetic history and lifestyle choices, into prompts, making disease prediction and health management more precise. Building on these predictions, the same multimodal backbone or a separate conversational LLM operating on its outputs, can be used to generate actionable insights and personalized recommendations. These models can analyze complex data patterns across multiple sources to support clinical decision‑making and enhance patient care. For example, LLM‑based systems like PhysioLLM engage users in personalized health conversations, utilizing physiological and contextual data to deepen individuals’ understanding of their health [14]. Another example, the Personal Health Insights Agent (PHIA), leverages LLMs, code generation, and web search to interpret wearable health data, providing accurate responses to health‑related questions and empowering users to develop data‑driven wellness strategies [15]. Our proposed multimodal framework can adopt these capabilities: by examining the aggregated multimodal data, comprehensive risk scores, and the model’s attention weights, it can reveal the relative importance of each data source, surface key cross‑modal interactions, and convert those findings into tailored, patient‑specific recommendations. Operating across scales, it can flag high‑risk groups for targeted public‑health measures while simultaneously delivering individualized guidance, thereby advancing proactive, data‑driven care for clinicians and patients.
In summary, the analysis illustrates the robustness and flexibility of the multimodal approach, which captures the complementary strengths of EHR and wearable data. While each modality independently contributes predictive value, their integration provides a fuller picture of patient health, enhancing the potential for proactive and precise disease risk assessments. The model’s ability to perform well with partial data further highlights its resilience, making it suitable for diverse healthcare applications focused on personalized, preventive care. By bridging the gap between clinical data and real-world lifestyle behaviors, the system represents an advancement in preventive healthcare, offering healthcare providers and patients deeper, more actionable insights for disease management and health risk mitigation.
Data Availability Statement:
The data used in this study were obtained from the UK Biobank under approved application number 94364. Access to the UK Biobank dataset is restricted to approved researchers who meet the access criteria set by the UK Biobank Access Management System. Due to legal and ethical considerations, individual-level data are not publicly available. Interested parties may apply for access directly through the UK Biobank portal (Apply for access - UK Biobank, https://www.ukbiobank.ac.uk/enable-your-research/apply-for-access).
Ethics Statement:
This research was conducted using the UK Biobank Resource under Application Number 94364. UK Biobank received ethical approval from the North West Multi-centre Research Ethics Committee (MREC; approval number 11/NW/0382). Under UK Biobank's Research Tissue Bank (RTB) approval, researchers using approved applications do not require separate institutional ethical approval. All participants provided written informed consent at recruitment. All methods were carried out in accordance with relevant guidelines and regulations, including the Declaration of Helsinki. All experimental protocols were approved by UK Biobank's Access Management System under the aforementioned application.
Code availability Statement:
The custom code used in this study for data processing and model development is not publicly available due to proprietary restrictions. However, it may be made available to qualified researchers upon reasonable request to the corresponding author and subject to institutional or licensing agreements.
A
Author Contribution
H.A. contributed to the conception and design of the study, developed and trained the model, performed the analysis, interpreted the results, and wrote the first draft of the manuscript. A.T. supervised the project, contributed to the study design and interpretation of results, and critically reviewed and revised the manuscript. Both authors approved the final version of the manuscript and agreed to be accountable for the accuracy and integrity of the work.
A
Acknowledgement
This research was conducted using the UK Biobank Resource under Application Number 94364. The authors thank the UK Biobank participants and coordinators for their invaluable contributions.
A
A
A
Data Availability
The data used in this study were obtained from the UK Biobank under approved application number 94364. Access to the UK Biobank dataset is restricted to approved researchers who meet the access criteria set by the UK Biobank Access Management System. Due to legal and ethical considerations, individual-level data are not publicly available. Interested parties may apply for access directly through the UK Biobank portal ( [Apply for access - UK Biobank](https:/www.ukbiobank.ac.uk/enable-your-research/apply-for-access) , https://www.ukbiobank.ac.uk/enable-your-research/apply-for-access).
References
1.- Singh, N., Khare, S., Thakur, P. & Sarawagi, K. Leveraging advanced AI algorithms to revolutionize health monitoring for seniors: a comprehensive analysis of data from wearables. EHRs beyond EasyChair Preprint 14103 (2024).
2.- Belyaeva, A. et al. Multimodal LLMs for health grounded in individual-specific data. In Machine Learning For Multimodal Healthcare Data, 86–102 (Springer Nature Switzerland, 2023). https://doi.org/10.1007/978-3-031-47679-2_7
3.- Kim, Y., Xu, X., McDuff, D., Breazeal, C. & Park, H. W. Health–LLM: large language models for health prediction via wearable sensor data. Proc. Health, Inference, and Learning, PMLR 248, 522–539 (2024). https://arxiv.org/abs/2401.06866
4.- Kim, Y. et al. Ben Shoham, O. & Rappoport, N. CPLLM: clinical prediction with large language models. PLOS Digit. Health 3, e0000680 (2024). https://doi.org/10.1371/journal.pdig.0000680
5.- Ding, J. E. et al. Large language multimodal models for new-onset type 2 diabetes prediction using five-year cohort electronic health records. Sci. Rep. 14, 20774. https://doi.org/10.1038/s41598-024-71020-2 (2024).
6.- Ding, J. E. et al. Cui, H.et al LLMs–based few–shot disease predictions using EHR: a novel approach combining predictive agent reasoning and critical agent instruction. arXiv (2024). abs/2403.15464.
7.-Ding, S., Ye, J., Hu, X. & Zou, N. Distilling the knowledge from large-language model for health event prediction. Sci. Rep. 14, 30675 (2024).
8.- Cosentino, J. et al. Towards a personal health large language model. arXiv abs (2024). /2406.06474.
9.- Sudlow, C. et al. UK Biobank: An open access resource for identifying the causes of a wide range of complex diseases of middle and old age. PLoS Med. 12, e1001779. https://doi.org/10.1371/journal.pmed.1001779 (2015).
10.- Rasmy, L., Xiang, Y., Xie, Z., Tao, C. & Zhi, D. Med-BERT: pretrained contextualized embeddings on large-scale structured electronic health records for disease prediction. npj Digit. Med. 4, 1 (2021). (2021).
11.- Pang, C. et al. CEHR-BERT: incorporating temporal information from structured EHR data to improve prediction tasks. In Proc. Mach. Learn. Health, PMLR 158, 239–260 (2021).
12.- Li, Y. et al. BEHRT: transformer for electronic health records. Sci. Rep. 10, 7155. https://doi.org/10.1038/s41598-020-62922-y (2020).
13.- Darabi, S., Kachuee, M., Fazeli, S. & Sarrafzadeh, M. TAPER: time-aware patient EHR representation. IEEE J. Biomed. Health Inform. 24, 3268–3275 (2020). (2020). https://doi.org/10.1109/JBHI.2020.2984931
14.- Fang, C. M. et al. IEEE,. PhysioLLM: supporting personalized health insights with wearables and large language models. In Proc. IEEE Conf. Biomed. Health Inform. (2024). Available at: https://arxiv.org/abs/2406.19283
15.- Merrill, M. A. et al. Transforming wearable data into health insights using large language model agents. arXiv abs. https://doi.org/10.48550/arXiv.2406.06464 (2024). /2406.06464.