|
| 1 | +## P1B2: Sparse Classifier Disease Type Prediction from Somatic SNPs |
1 | 2 |
|
| 3 | +**Overview**: Given patient somatic SNPs, build a deep learning network that can classify the cancer type. |
| 4 | + |
| 5 | +**Relationship to core problem**: Exercise two core capabilities we need to build: (1) classification based on very sparse input data; (2) evaluation of the information content and predictive value in a molecular assay with auxiliary learning tasks. |
| 6 | + |
| 7 | +**Expected outcome**: Build a DNN that can classify sparse data. |
| 8 | + |
| 9 | +### Benchmark Specs Requirements |
| 10 | + |
| 11 | +#### Description of the Data |
| 12 | +* Data source: SNP data from GDC MAF files |
| 13 | +* Input dimensions: 28,205 (aggregated variation impact by gene from 2.7 million unique SNPs) |
| 14 | +* Output dimensions: 10 class probabilities (9 most abundant cancer types in GDC + 1 “others”) |
| 15 | +* Sample size: 4,000 (3000 training + 1000 test) |
| 16 | +* Notes on data balance and other issues: data balance achieved via undersampling; “others” category drawn from all remaining lower-abundance cancer types in GDC |
| 17 | + |
| 18 | +#### Expected Outcomes |
| 19 | +* Classification |
| 20 | +* Output range or number of classes: 10 |
| 21 | + |
| 22 | +#### Evaluation Metrics |
| 23 | +* Accuracy or loss function: Standard approaches such as F1-score, accuracy, ROC-AUC, cross entropy, etc. |
| 24 | +* Expected performance of a naïve method: linear regression or ensemble methods without feature selection |
| 25 | + |
| 26 | +#### Description of the Network |
| 27 | +* Proposed network architecture: MLP with regularization |
| 28 | +* Number of layers: ~5 layers |
| 29 | + |
| 30 | +### Running the baseline implementation |
| 31 | + |
| 32 | +``` |
| 33 | +cd P1B2 |
| 34 | +python p1b2_baseline.py |
| 35 | +``` |
| 36 | +The training and test data files will be downloaded the first time this is run and will be cached for future runs. |
| 37 | + |
| 38 | +#### Example output |
2 | 39 |
|
3 | 40 | ```
|
4 | 41 | Using Theano backend.
|
@@ -64,5 +101,13 @@ best_val_loss=1.31111 best_val_acc=0.59500
|
64 | 101 |
|
65 | 102 | Best model saved to: model.A=sigmoid.B=64.D=None.E=20.L1=1024.L2=512.L3=256.P=1e-05.h5
|
66 | 103 |
|
67 |
| -Evaluation on test data: {'accuracy': 0.55000000000000004} |
| 104 | +Evaluation on test data: {'accuracy': 0.5500} |
| 105 | +``` |
| 106 | + |
| 107 | +### Running the XGBoost classifier |
| 108 | + |
| 109 | +``` |
| 110 | +cd P1B2 |
| 111 | +python p1b2_xgboost.py |
| 112 | +
|
68 | 113 | ```
|
0 commit comments