1 / 49

Biomedical Applications of Prototype-Based Classifiers and Relevance Learning

This is a study on the application of prototype-based classifiers and relevance learning in the biomedical field. It explores the use of supervised learning techniques for classification, regression, and prediction based on labeled example data. The study also discusses the significance of interpretable "white-box" systems and compares different performance measures for evaluating classifier models.

billyjames
Download Presentation

Biomedical Applications of Prototype-Based Classifiers and Relevance Learning

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Biomedical applicationsof prototype-based classifiersandrelevancelearning Michael BiehlIntelligent Systems Johann Bernoulli Institute for Mathematics and Computing Science University of Groningen / NL www.cs.rug.nl/~biehl Introduction: prototype-based classification, relevance learning Generalized Matrix Relevance LVQ Illustration: three bio-medical applications

  2. supervised learning • classification/ regression / prediction • based on labeledexample data generic workflow: example data model apply to novel data working training validation estimate working performance set parameters of model / training compare different models obvious performance measures: overall / class-wise accuracy ROC, Precision Recall ... accuracy is not enough - interpretable “white-box” systems example: prototype-based models, distance-based classifiers

  3. distance-based classifiers a simple distance-based system: (K) NN classifier • store a set of labeled examples • classify a query according to the • label of the Nearest Neighbor • (or the majority of K NN) • piece-wise linear decision • boundaries according • to (e.g.) Euclidean distancefrom all examples ? N-dim. feature space • expensive (storage, computation) • sensitive to mislabeled data • overly complex decision boundaries +conceptually simple, + no training phase + only one parameter (K)

  4. prototype-based classification Learning Vector Quantization [Kohonen, 1990] • represent the data by one or • several prototypes per class • classify a query according to the • label of the nearest prototype • (or alternative schemes) ? • local decision boundaries acc. • to (e.g.) Euclidean distances +robust, low storage needs, little computational effort N-dim. feature space + parameterization in feature space, interpretability - model selection: number of prototypes per class, etc. requires training: placement of prototypes in feature space

  5. Learning Vector Quantization N-dimensional data, feature vectors ∙identification of prototype vectors from labeled example data ∙distance based classification (e.g. Euclidean) competitive learning: LVQ1 [Kohonen, 1990] •initialize prototype vectors for different classes • present a single example • identify the winner (closest prototype) •move the winner -closertowards the data (same class) -away from the data (different class)

  6. Learning Vector Quantization N-dimensional data, feature vectors ∙identification of prototype vectors from labeled example data ∙distance based classification (e.g. Euclidean) ∙distance-based classification [here: Euclidean distances] ∙tesselation of feature space [piece-wise linear] ∙aim: discrimination of classes ( ≠ vector quantization or density estimation ) ∙generalization ability correct classification of new data

  7. cost function based LVQ one example: Generalized LVQ (GLVQ) cost function [Sato&Yamada, 1995] minimize two winning prototypes: Efavors - small number of misclassifications, e.g. with - large margins between classes - small , large - class-typical prototypes

  8. LVQ distance measures ? key question: appropriate distance / (dis-) similarity measure fixed, pre-defined distance measures: (G)LVQ can formulated for general (differentiable) distances examples: Minkowski distances (p≠2), correlation based, statistical divergences, ... not necessarily metrics! standard work-flow - consider several distance measures according to prior knowledge - compare performances in, e.g., cross-validation elegant approach: Relevance Learning / adaptive distances - employ parameterized distance measure - optimize in the data-driven training process (cost function!)

  9. GMLVQ Generalized Matrix Relevance LVQ: [Schneider, Biehl, Hammer, 2009] generalized quadratic distancein LVQ:

  10. GMLVQ Generalized Matrix Relevance LVQ: [Schneider, Biehl, Hammer, 2009] generalized quadratic distancein LVQ: training: adaptation of prototypes and distance measure guided by GLVQ cost function variants: one global, several local, class-wise relevance matrices diagonal matrices: single feature weights [Hammer et al., 2002] rectangular low-dim. representation / visualization [Bunte et al., 2012]

  11. interpretation after training: prototypes represent typical class properties or subtypes • summarizes • the contribution of a single dimension • the relevance of original features in the classifier Relevance Matrix quantifies the contribution of the pair of features (i,j) to the distance Note: interpretation assumes implicitlythat features have equal order of magnitude e.g. after z-score-transformation → (averages over data set)

  12. three application examples I) steroid metabolomics: - detection of malignancy in adrenocortical tumors based on urinary steroid metabolite excretion GMLVQ: ~ 150 samples, 32-dim. feature vectors II) cytokine expression data: - diagnosis of (early) rheumatoid arthritis based on synovial tissue samples ~ 50 samples represented by 117 cytokine expressions in synovial tissue, PCA+GMLVQ combined III) gene expression data: - recurrence risk prediction from tumor samples ~ 400 samples, ~20000 dim. feature space outlier analysis + GMLVQ on (80) pre-selected genes

  13. Steroid metabolomics: detectingmalignancyin adrenocorticaltumors www.ensat.org W. Arlt, M. Biehl, A. Taylor, S. Hahner, R. Libé, B. Hughes, P. Schneider, D. Smith, H. Stiekema, N. Krone, E. Porfiri, G. Opocher, J. Bertherat, F. Mantero, B. Allolio, M. Terzolo, P. Nightingale, C. Shackleton, X. Bertagna, M.Fassnacht, P. Stewart Urine Steroid Metabolomics as a Biomarker Tool for Detecting Malignancy in Patients with Adrenal Tumors J Clinical Endocrinology & Metabolism 96:3775-3784 (2011)

  14. steroid metabolomics • classification of adrenocorticaltumors (adenoma vs. carcinoma) • based on steroid hormone excretion profiles benign ACA malignant ACC features: 32 steroidmetaboliteexcretionvalues non-invasive measurement (24 hrs. urinesamples) www.ensat.org aim:develop a novel biomarker tool for differential diagnosis idea: identify characteristic steroid profiles (prototypes)

  15. steroid metabolomics [Arlt et al., 2011] [Biehl et al., 2012] Generalized Matrix LVQ , ACCvs. ACAclassification ∙data divided in 90% training, 10% test set, (z-score transformed) ∙determine prototypes typical profiles (1 per class) ∙adaptive generalized quadratic distance measure parameterized by ∙apply classifier to test data evaluate performance (error rates, ROC) ∙repeat and average over many random splits

  16. steroid metabolomics prototypes: steroid excretion in ACA/ACC ACA ACC

  17. steroid metabolomics Relevance matrix … of pairs of markers relevance of single markers frequency of markers to be among top 9 subset of selected steroids ↔ technical realization (patented, UoB) using9 markers only, similar ROC

  18. steroid metabolomics ROC characteristics clear improvement due to adaptive distances 90% / 10% randomized splits of the data in training and test set averages over 1000 runs (sensitivity) AUC 0.87 0.93 0.97 Euclidean diagonal rel. full matrix (1-specificity)

  19. steroid metabolomics Relevance matrix diagonal elements off-diagonal 19 ACA ACC discriminative e.g. steroid 19 (THS)

  20. steroid metabolomics TH-Doc (12) highly discriminative combination of markers! weakly discriminative markers 5a-THA (8)

  21. adrenocortical tumors GRLVQ (sensitivity) GMLVQ AUC 0.87 0.93 0.97 Euclidean 8 diagonal rel. full matrix (1-specificity)

  22. visualization of the data set generic property: relevance matrix becomes highly singular ACA ACC

  23. work in progress • high-throughput LC/MS assay to replace GC/MS • on-going prospective study w.r.t. ~ 2000 patients • monitoring of patients after surgeryand/or under medication aim: recurrence detection / prediction • identification of tumor subtypes ? • other disorders affecting / related to steroid metabolism

  24. Early diagnosisof Rheumatoid Arthritis L. Yeo, N. Adlard, M. Biehl, M. Juarez, M. Snow C.D. Buckley, A. Filer, K. Raza, D. Scheel-Toellner Expression of chemokines CXCL4 and CXCL7 by synovial macrophages defines an early stage of rheumatoid arthritis Annals of the Rheumatic Diseases75:763-771 (2016)

  25. rheumatoid arthritis (RA) uninflamed control early inflammation established RA resolving early RA ultimate goals: understand pathogenesis and mechanism of progression ? cytokine based diagnosis of RA at earliest possible stage ?

  26. synovial tissue cytokine expression synovium tissue section mRNA extraction real-time PCR • panel of 117 cytokines • cell signaling proteins • regulate immune response • produced by, e.g. • T-cells, macrophages, • lymphocytes, fibroblasts, etc.

  27. GMLVQ analysis • pre-processing: • log-transformed expression values • 21 leading principal components explain 95% of the variation • Twotwo-class problems: (A) established RA vs. uninflamed controls • (B) early RA vs. resolving inflammation • 1 prototype per class, global relevance matrix, distance measure: • leave-two-out validation (one from each class) • evaluation in terms of Receiver Operating Characteristics

  28. Matrix Relevance LVQ diagonal relevances leave-one-out (A) established RA vs. uninflamed control true positive rate false positive rate diagonal Λiivs. cytokine index i (B) early RA vs. resolving inflammation true positive rate

  29. protein level studies CXCL4 chemokine (C-X-C motif) ligand 4 CXCL7 chemokine (C-X-C motif) ligand 7 direct study on protein level, staining / imaging of sinovial tissue: macrophages : predominant source of CXCL4/7 expression • high levels of CXCL4 and • CXLC7 in early RA • expression on macrophages • outside of blood vessels • discriminates • early RA / resolving cases

  30. relevant cytokines leave-one-out diagonal relevances (A) established RA vs. uninflamed control true positive rate false positive rate diagonal Λiivs. cytokine index i macrophage stimulating 1 (B) early RA vs. resolving inflammation true positive rate

  31. work in progress • more samples (difficult...) needed in order • to obtain a reliable early diagnosis • integrated analysis of gene expression and other data • from the same / an analogous patient cohort

  32. Predicting Recurrence in Clear Cell Renal Cell Carcinoma Analysis of TCGA data using Outlier Analysis and GMLVQ Gargi Mukherjee … Rutgers University, New Jersey Kevin Raines … Stanford University, California SrikanthSastry… JNC, Bengaluru, India Sebastian Doniach… Stanford University, California Gyan Bhanot … Rutgers University, New Jersey Michael Biehl … University of Groningen, The Netherlands In: Proc. IEEE Congress on Evolutionary Computation CEC 2016

  33. data clear cell Renal Cell Carcinoma (ccRCC) publicly available datasets: The Cancer Genome Atlas (TCGA) cancergenome.nih.gov also hosted at Broad Institutegdac.broadinstitute.org

  34. data clear cell renal cell carcinoma TCGA data @ BroadInstitute mRNA-Seqexpressiondata X normalized, log-transformed: Y=log(1+X) 65 normal samples 65 matched tumor samples 469 tumor samples in total 469 tumor samples 65 normal samples 20532 genes recurrence data: daysafter diagnosis numberof recurrences 65 + 65 matched

  35. outlier analysis 380 training samples 89 test samples fast forwardto machinelearning analysis randomizedsplit

  36. outlier analysis per gene: determine mean μ, standard deviation σ of Y for each gene: identify outlier samples 380 training samples Y > μ + σ“high outlier“ Y < μ - σ“lowoutlier“ restrictthefollowinganalysisto genes with ≥ 20 high outlier samples or ≥ 20 low outlier samples

  37. outlier analysis Kaplan-Meier (KM) analysisper gene: test for significant association ofoutlierstatusofsampleswith recurrence 1546 „high-outlier genes“ with KM log rank p < 0.001 1628 „low-outlier genes“ with KM log rank p < 0.0005 1546 genes construct two binary outlier matrices „1“ for high-outlier samples „0“ else „1“ for low-outlier samples „0“ else 380 samples  PCA 1628 genes 380 samples

  38. outlier analysis high outlier genes PCA reveals four clusters of genes genes in small clusters (B,D): outlierstatus associated with late recurrence 1475 71 A B lowoutlier genes genes in large clusters(A,C): outlierstatus associated with early recurrence 1402 226 C D

  39. recurrence risk score top 20 genes (by KM p-value) fromeachclusterA,B,C,D referencesetof 80 genes foreach sample: - determineoutlierstatusw.r.t. the 80 genes (Y>?<μ ± σ ) - addupcontributions per gene - 1 if sample isoutlierw.r.t. to a gene in A or C (earlyrec.) 0 if sample is not an outlierw.r.t. thegene + 1 if sample isoutlierw.r.t. to a gene in B or D (laterec.) recurrence risk score - 40 ≤ R ≤ + 40 observe: median = 2 over the 380 training samples crisp classification w.r.t. recurrence risk: high risk (early recurrence) if R < 2 low risk (late recurrence) if R ≥ 2

  40. recurrence risk prediction KM plots with respect to high / low risk groups: training set (380 samples) test set (89 samples) log rank p < 1.e-16 log rank p < 1.e-4 • risk score R is predictive of the actual recurrence risk • the 80 selected genes can serve as a prognostic panel

  41. extreme case analysis outlieranalysisyields 4 groups(A,B,C,D) of 20 pre-selected genes associatedwithlate/earlyrecurrence • 80-dim. feature vectors ≤ 2 years (early) numberof recurrences: > 5 years (undefined) (late or no recurrence) 2 classes: 109 samples class 2, high risk 107 samples class 1, low risk

  42. GMLVQ classifier • one prototype vector per class: • adaptive distanceforcomparisonofsamplesandprototypes: diagonal elements of Λ components of C A B D B A C D low expression | high expression

  43. GMLVQ classifier ROC of GMLVQ classifier (Leave-One-Out of the 216 extreme samples) log rank p < 1.e-7 KM plot w.r.t. all 469 samples ( L-1-O for 216 samples, plus 253 undefined )

  44. diagnostics? • the set of 80 genes is also diagnostic: • GMLVQ separates normal from tumor cells (close to) perfectly • PCA of corresponding gene expressions: gradient from normal to high risk: 65 normal samples 105 low risk samples (late rec.) 109 high risk samples (early rec.)

  45. most relevant genes (GMLVQ) 12 most relevant genes from GMLVQ classifier

  46. remarks and open questions • prospective studies • 80 genes do not necessarily reflect biological mechanismscompare, e.g., with known pathways / modules of genes • GMLVQ suggests an even smaller panel of genes (12?) • identify a minimum panel for diagnostics and prognostics • more direct, multivariate identification of relevant genes by • dimension reduction + GMLVQ with back-transform

  47. conclusion prototype- and distance based systems: - intuitive, transparent, interpretable - classification, regression, unsupervised learning, visualization ... - relevance learning: further insight into data and problem - suitable for a variety of bio-medical problems a recent review: M. Biehl, B. Hammer, T. Villmann Prototype-based models in Machine Learning Advanced Review in: WIRES Cognitive Science 7(2): 92-111 (2016)

  48. links Matlab code: Relevance and Matrix adaptation in Learning Vector Quantization (GRLVQ, GMLVQ and LiRaM LVQ): http://matlabserver.cs.rug.nl/gmlvqweb/web/ A no-nonsense beginners’ tool for GMLVQ: http://www.cs.rug.nl/~biehl/gmlvq (see also: Tutorial, Thursday 9:30) Pre- and re-prints etc.: http://www.cs.rug.nl/~biehl/

  49. thanks Barbara Hammer Thomas Villmann WiebkeArlt Dagmar Scheel-Toellner GyanBhanot Petra Schneider Kerstin Bunte

More Related