180 likes | 326 Views
Predicting Genetic Regulatory Response Using Classification. Us v. Them (“Them” being Manuel Middendorf, Anshul Kundaje, Chris Wiggins, Yoav Freund, and Christina Leslie, in “Predicting Genetic Regulatory Response Using Classification” (2004). The Problem.
E N D
Predicting Genetic Regulatory Response Using Classification Us v. Them (“Them” being Manuel Middendorf, Anshul Kundaje, Chris Wiggins, Yoav Freund, and Christina Leslie, in “Predicting Genetic Regulatory Response Using Classification” (2004)
The Problem • Current studies of gene transcription tend to be descriptive • Need for a predictive system – the ability to predict gene regulation for new experiments • Rather than determining patterns in sets of genes and conditions, look at underyling causes of those patterns
The Important Parts of Genes & Experiments • Regulation is determined by binding sites (motifs) and regulators (parents) • The significance of experiments, then, is how they affect regulators • The significance of genes is what motifs they contain
Binding Sites and Regulation • Discretize gene response into only up-regulated (1), down-regulated(-1), or unchanged (0) • A motif is either present (1) or absent (0) • A parent is either up-regulated (1), down-regulated (-1), or unchanged (0) • Assume (and we need to check with someone who actually knows something about biology on this) that things only happen if motif is present and parent is either up- or down-regulated
What our matrix really looks like • g = # of genes • e = # of experiments • p = # of parents • m = # of motifs • Then we have g*e response values • For each response, we have p*m parent/motif combinations • For each parent/motif combination, there are three possibilities – present and up-regulated, present and down-regulated, or all those other possibilities where nothing happens • Represent these possibilities as a pair of binary variables, one for up and one for down
Some Numbers • In the paper, the initial dataset had 6110 genes and 173 experiments • 354 motifs are considered • 475 regulators are considered • Set of genes to consider is reduced to only 1411 genes of interest
Some More Numbers • Only train on genes that are up- or down-regulated • Approx. 8% of gene/experiment pairs from the overall sample appear to be, so, assuming this holds true in the reduced sample, we have 19,632 gene / experiment pairs to train on • For each of these values we have 2*354*475 = 336,300 predictor variables
Some Problems • 19,632 by 336,300 is an awfully large matrix to want to do any calculations with • We have far more variables than observations
Possible Solutions • Random Projection: • Pro: we can reduce our dimensionality • Con: it seems like a somewhat silly approach • Con: there’ll still be a lot of calculations just to make the projection
Possible Solutions • Variable Selection or PCA • Pro: we could reduce our dimensionality in a more informed way • Con: computationally painful
Possible Solutions • Random Forest • From Breiman and Cutler: • Grow a number of classification trees, and take the vote of the classifications of all trees • For each tree, if we have n cases, sample, with replacement, n cases, using some number of randomly chosen variables much smaller than the full number of variables • Pro: allows for flexibility in reduction of dimensions being considered • Pro: dimension is reduced without computation problems • Con: ?
Their Solution • Alternating decision tree: • A tree with alternating levels of prediction nodes and splitter nodes • Combines a set of weak prediction rules (the splitter nodes) to make one strong rule
Alternating Decision Treesv.Random Forests • Both the same rough concept – take a vote from many weak rules to get a strong rule • In ADTs, rules are based on single variables, and may be conditioned on values of other variables • In random forests, rules can be based on multiple variables, and are only marginal over all values of other variables • ADTs are fairly straightforward to read and interpret
Alternating Decision Trees • Pro: computationally kind • Pro: works better than Random Forests (Creamer & Freund, 2004) • Con: we didn’t come up with it