1 / 20

Data mining Satellite images indexation

Data mining Satellite images indexation. Feature Selection Marine Campedel 5 février 2004. Data Mining (1). “Fouille de données”, ECD, KDD,… Automatic process giving access to raw data in the context of a given application ;

zinnia
Download Presentation

Data mining Satellite images indexation

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Data mining Satellite images indexation Feature Selection Marine Campedel 5 février 2004

  2. Data Mining (1) • “Fouille de données”, ECD, KDD,… • Automatic process giving access to raw data in the context of a given application ; • Necessity because of databases increasing sizes -> find the “relevant” information ; • Indexation : automatic process that associates a set of labels to a raw data.

  3. Data acquisition Data Mining (2) Off-line process Raw Data (images) Features Information Extraction Semantic Models Supervised learning User Interface Information Retrieval User query On-line process

  4. Information Extraction From raw data and a priori knowledge (unsupervised) Between raw data and application-based knowledge (supervised) Information Retrieval Goal : get relevant examples (raw images) corresponding to any user query (‘find frozen woods area’) in a specified application (‘satellite image retrieval’) Data Mining (3)

  5. Any a priori knowledge from data type or final application ? Data acquisition Features Selection (1) Raw Data (images) Features Information Extraction • Computation cost and storage capacity  reduce the number of features (dimension) ; • Reduce redundancy while maintaining noise robustness and discriminative power ; •  Feature selection algorithm is needed

  6. Features Selection (2) Domain a priori knowledge ? Raw Data (images) Predefined properties ? Relevance definition ? Compute all a priori features (colour, texture, shape features,…) Construct new features (PCA, ICA,…) Feature Selection Relevant features

  7. Unsupervised Quantization Define selection criterion from a priori knowledge (‘filter’ approach) Typical use of correlation coeffs, mutual information,… and thresholding Traditional drawback : cannot evaluate set of features Supervised Define selection criterion according to the final application (‘wrapper’ or ‘embedded’ approach) Typical use of labelled databases and classifiers Traditional drawback : computation cost Features Selection (3)

  8. Supervised Features Selection • Inputs : labelled database + classification task + exhaustive feature library • Goal : select the features set that achieves the best classification score • Pbs : selection of the inputs (database, classifier type, feature library are chosen from a priori knowledge)

  9. Constraints • The (hand-)labelled database size is limited by the acquisition process (hundreds to thousands ?) • The features library size can be huge (hundreds ?) The classifier must be able to train from a limited number of data in high dimensional space, ensuring strong generalization property

  10. SVM choice • Support Vector Machine • Parametric classifier; • Support vectors : examples that define the limits of each class; • Designed to be robust to outliers; • Tractable with high dimensional data; • Lots of recent literature and tools on the web (matlab: SPIDER, C-C++: svmlib, svmlight, Java: WEKA).

  11. SVM principle (1/4) • 2 classes linear SVM without error Labelled training patterns Linearly separable if there exists w (weights) and b (bias) such that • The optimal hyperplane separates the data with the maximal margin (determines the direction w/|w| where the distance between the projections of two different classes data is maximal)

  12. SVM principle (2/4) • Support vectors • SVM problem :

  13. SVM principle (3/4) • Dual problem • Kernel

  14. SVM principle (4/4) • Soft margin • Multi-classes : 1-vs-all and MC-SVM

  15. Selection algorithms using SVM • RFE (Recursive Feature Elimination) [Guyon, 2002] • Iteratively eliminates features corresponding to small weights until the desired number of features is reached. • Minimize L0 norm of feature weights (minimize the number of non-zero weights) [Weston, 2003] • Iterative process using linear SVM ; • Update data by multiplying by the estimated weights.

  16. Proposed experiment • Database : synthetic or Brodatz (texture images) or satellite image database • Feature library : using Gabor, orthogonal wavelets, co-occurrence matrices, basic local stats,…with several neighbourhoods sizes (scales) • Classifier : SVM • Goal : compare performance of different selection algorithms (supervised + unsupervised ones) • Robustness to database modification ? to classifier parameter modification ?

  17. Spider example : 2-classes and 2-relevant dimensions synthetic linear problem The 2 first dimensions are relevant (uniform distribution) The next 6 features are noisy versions of the two first dimensions The 42 other one are independent uniformly distributed variables (noise) 400 examples, 50 dimensions Evaluation using cross-validation (train on 80% of the data, test on 20%, 5 attempts) Score = classification error rate

  18. Spider example • Results confirm the selection process gain • Correlation-based selection algorithm performs poorly compared to the proposed ‘wrapper’ methods

  19. Conclusion and what next ? • Subject : feature selection algorithms • Determine an automatic procedure for selecting relevant features in the context of satellite image indexing • Applicable to any data indexing ? (Is the data-type a-priori knowledge concentrated in the feature library design ?) • Experiment in progress…

  20. Bibliography • [Elisseeff, 2003] “Technical documentation of the multi-class SVM”,2003. • [Guyon,2002] “Gene selection for cancer classification using support vector machines”, I.Guyon, J.Weston, S.Barnhill and V.Vapnik - Machine Learning 46(1-3) 389-422, 2002. • [Guyon,2003] “An introduction to Variable and Feature selection”, I.Guyon, A.Elisseeff, JMLR 3, 1157-1182, 2003. • [Schoelkopf and Smola,2002] “Learning with Kernels-Support Vector Machines, Regularization, Optimization and Beyond”, B.Schoelkopf and A.J.Smola, MIT press, 2002. • [Weston,2003] “Use of the Zero-Norm with Linear Models and Kernel Methods”, Weston, Elisseff, Schoelkopf and Tipping, JMLR 3, 1439-1461,2003.

More Related