Representation learning for health with multi-modal clinical data

Seminar: 
Applied Mathematics
Event time: 
Monday, April 17, 2017 - 12:15pm to 1:15pm
Location: 
LOM 201
Speaker: 
Marzyeh Ghassemi
Speaker affiliation: 
MIT
Event description: 

The explosion of clinical data provides an exciting new opportunity to use machine learning to discover new and impactful clinical information. Among the questions that can be addressed are establishing the value of treatments and interventions in heterogeneous patient populations, creating risk stratification for clinical endpoints, and investigating the benefit of specific practices or behaviors. However, there are many challenges to overcome. First, clinical data are noisy, sparse, and irregularly sampled. Second, many clinical endpoints (e.g., the time of disease onset) are ambiguous, resulting in ill-defined prediction targets.

I tackle these problems by learning abstractions that generalize across applications despite missing and noisy data. My work spans coded records from administrative staff, vital signs recorded by monitors, lab results from ordered tests, notes taken by clinical staff, and accelerometer signals from wearable monitors. The learned representations capture higher-level structure and dependencies between multi-modal time series data and multiple time-varying targets. I focus on learning techniques that transform diverse data modalities into a consistent intermediate that improves prediction in clinical investigation.

In this talk, I will present work that addresses the problem of learning good representations using clinical data. I will discuss the need for practical, evidence-based medicine, and the challenges of creating multi-modal representations for prediction targets varying both spatially and temporally. I will present work using electronic medical records for over 30,000 intensive care patients from the MIMIC-III dataset to predict both mortality and clinical interventions. To our knowledge, classification results on these task are better than those of previous work. Moreover, the learned representations hold intuitive meaning - as topics inferred from narrative notes, and as latent autoregressive states over vital signs. I will also present work from a non-clinical setting that uses non-invasive wearable data to detect harmful vocal patterns and their pathological physiology. I present two sets of results in this area: 1) it is possible to detect pathological anatomy from the ambulatory signal, and 2) it is possible to detect the impact of therapy on vocal behaviors.