Skip to content

Commit

Permalink
Added sections on embeddings for medical ontologies and causal infere…
Browse files Browse the repository at this point in the history
…nce (#339)

This build is based on
22c54f0.

This commit was created by the following Travis CI build and job:
https://travis-ci.org/greenelab/deep-review/builds/225555590
https://travis-ci.org/greenelab/deep-review/jobs/225555591

[ci skip]

The full commit message that triggered this build is copied below:

Added sections on embeddings for medical ontologies and causal inference (#339)

* Add reference to Multi-task Prediction of Disease Onsets from Longitudinal Lab Tests in Electronic Helath Records data set.

* Added in the anchor and learn framework.  This isn't strictly deep learning, so I don't know if it should be included, but it is relevant ot the section.

* Added in references to neural embeddings in medical coding

* Added causal inference references

* Changed pmid to doi

* Changed PMC id to regular PubMed id.
  • Loading branch information
DaveDeCaprio committed Apr 25, 2017
1 parent 15c79b2 commit 3c41159
Show file tree
Hide file tree
Showing 5 changed files with 1,671 additions and 4 deletions.
37 changes: 33 additions & 4 deletions all-sections.md
Original file line number Diff line number Diff line change
Expand Up @@ -316,6 +316,22 @@ This indicates a potential strength of deep methods. It may be possible to
repurpose features from task to task, improving overall predictions as the field
tackles new challenges.

Several authors have created reusable feature sets for medical terminologies using
neural embeddings, as popularized by word2Vec [@1GhHIDxuW]. This approach
was first used on free text medical notes by De Vine et al.
[@XQtuRkTU] with results at or better than traditional methods.
Y. Choi et al.[@1qa47hoP] built embeddings of standardized
terminologies, such as ICD and NDC, used in widely available administrative
claims data. By learning terminologies for different entities in the same
vector space, they can potentially find relationships between different
domains (e.g. drugs and the diseases they treat). Medical claims data does not
have the natural document structure of clinical notes, and this issue was
addressed by E. Choi et al. [@TwvauiTv], who built
embeddings using a multi-layer network architecture which mimics the structure
of claims data. While promising, difficulties in evaluating the quality of
these kinds of features and variations in clinical coding practices remain as
challenges to using them in practice.

Identifying consistent subgroups of individuals and individual health
trajectories from clinical tests is also an active area of research. Approaches
inspired by deep learning have been used for both unsupervised feature
Expand All @@ -331,9 +347,10 @@ scale analysis of an electronic health records system found that a deep
denoising autoencoder architecture applied to the number and co-occurrence of
clinical test events, though not the results of those tests, constructed
features that were more useful for disease prediction than other existing
feature construction methods [@WrNCJ9sO]. Taken together, these
results support the potential of unsupervised feature construction in this
domain. However, numerous challenges including data integration (patient
feature construction methods [@WrNCJ9sO]. Razavian et al.
[@c6MfDdWP] used a set of 18 common lab tests to predict disease onset
using both CNN and LSTM architectures and demonstrated and improvement over baseline
regression models. However, numerous challenges including data integration (patient
demographics, family history, laboratory tests, text-based patient records,
image analysis, genomic data) and better handling of streaming temporal data
with many features, will need to be overcome before we can fully assess the
Expand Down Expand Up @@ -410,7 +427,9 @@ making methodological choices that either reduce the need for labeled examples
or that use transformations to training data to increase the number of times it
can be used before overfitting occurs. For example, the unsupervised and
semi-supervised methods that we've discussed reduce the need for labeled
examples [@5x3uMSKi]. The adversarial training example
examples [@5x3uMSKi]. The anchor and learn framework
[@A9JeoGV8] uses expert knowledge to identify high confidence
observations from which labels can be inferred. The adversarial training example
strategies that we've mentioned can reduce overfitting, if transformations are
available that preserve the meaningful content of the data while transforming
irrelevant features [@Xxb4t3zO]. While adversarial training examples
Expand Down Expand Up @@ -1262,6 +1281,16 @@ interpretability of deep learning models, fitting deep models to limited and
heterogeneous data, and integrating complex predictive models into a dynamic
clinical environment.

A critical challenge in moving from prediction to treatment recommendations
is the necessity to establish a causal relationship for a recommendation.
Causal inference is often framed in terms of counterfactual question
[@cpNVdlL7]. Johansson et al [@173ftiSzF] use deep neural
networks to create representation models for covariates that capture nonlinear
effects and show significant performance improvements over existing models. In
a less formal approach, Kale et al [@FUIfIdE] first create a deep neural
network to model clinical time series and then analyze the relationship of the
hidden features to the output using a causal approach.

#### Applications

##### Trajectory Prediction for Treatment
Expand Down
48 changes: 48 additions & 0 deletions bibliography.bib
Original file line number Diff line number Diff line change
Expand Up @@ -1690,3 +1690,51 @@ @article{xl1ijigK
year = {2017}
}


@article{173ftiSzF,
abstract = {Observational studies are rising in importance due to the widespread
accumulation of data in fields such as healthcare, education, employment and
ecology. We consider the task of answering counterfactual questions such as, "Would this patient have lower blood sugar had she received a different
medication?". We propose a new algorithmic framework for counterfactual
inference which brings together ideas from domain adaptation and representation
learning. In addition to a theoretical justification, we perform an empirical
comparison with previous approaches to causal inference from observational
data. Our deep learning algorithm significantly outperforms the previous
state-of-the-art.},
archiveprefix = {arXiv},
author = {Fredrik D. Johansson and Uri Shalit and David Sontag},
eprint = {1605.03661v2},
file = {1605.03661v2.pdf},
link = {http://arxiv.org/abs/1605.03661v2},
month = {May},
primaryclass = {stat.ML},
title = {Learning Representations for Counterfactual Inference},
year = {2016}
}


@article{c6MfDdWP,
abstract = {Disparate areas of machine learning have benefited from models that can take
raw data with little preprocessing as input and learn rich representations of
that raw data in order to perform well on a given prediction task. We evaluate
this approach in healthcare by using longitudinal measurements of lab tests, one of the more raw signals of a patient's health state widely available in
clinical data, to predict disease onsets. In particular, we train a Long
Short-Term Memory (LSTM) recurrent neural network and two novel convolutional
neural networks for multi-task prediction of disease onset for 133 conditions
based on 18 common lab tests measured over time in a cohort of 298K patients
derived from 8 years of administrative claims data. We compare the neural
networks to a logistic regression with several hand-engineered, clinically
relevant features. We find that the representation-based learning approaches
significantly outperform this baseline. We believe that our work suggests a new
avenue for patient risk stratification based solely on lab results.},
archiveprefix = {arXiv},
author = {Narges Razavian and Jake Marcus and David Sontag},
eprint = {1608.00647v3},
file = {1608.00647v3.pdf},
link = {http://arxiv.org/abs/1608.00647v3},
month = {Aug},
primaryclass = {cs.LG},
title = {Multi-task Prediction of Disease Onsets from Longitudinal Lab Tests},
year = {2016}
}

Loading

0 comments on commit 3c41159

Please sign in to comment.