Marrying Generative Model of Healthcare Events with Digital Twin of Human-Environment Interaction for Disease Reasoning
Ziquan Wei ⋅ Tingting Dan ⋅ Guorong Wu
Abstract
Despite the central role of sensor-derived measurements such as imaging traits and plasma biomarkers in biomedical research and clinical practice, existing generative models for disease prediction largely depend on event-level representations from hospital and registry data. Given the multi-factorial nature of human disease, the absence of human-environment interaction modeling limits the capacity for personalized disease modeling and clinical decision support. To address this limitation, we propose a generative model with human-environment interaction for \textit{in silico} modeling of disease reasoning, a conditioned latent diffusion framework that establishes the connection between multi-organ sensor data with tokenized healthcare events. Specifically, we introduce a novel geometric diffusion model to characterize the temporal evolution of complex data representation such as brain networks (region-to-region connectivity encoded in a graph), in parallel with diffusion models for tabular data from other organ systems. Together, we integrate the generative model with digitalized human-environment interaction (coined DiffDT) for simulated intervention and reasoning of future disease trajectories. We conduct extensive experiments on the UK Biobank (UKB) dataset, which contains organ-specific imaging traits, including brain (44,834), heart (23,987), liver (28,722), and kidney (32,155), along with nearly 500k medical history sequences (age range: 25$\sim$89 years). Our DiffDT achieves significant improvements over state-of-the-art human disease autoregressive models and imaging trait generative baselines.
Successful Page Load