740 likes | 884 Views
This document explores the intricacies of similarity-based classification, focusing on art, particularly differentiating between masterpieces by Van Gogh and Monet. It delves into various similarity functions, algorithms from computational biology and computer vision, and techniques in information retrieval. Approaches like Support Vector Machines (SVM), Weighted Nearest-Neighbors, and Linear Interpolation Weights are discussed, along with their design goals. By employing kernels and similarity metrics, the paper aims to enhance classification performance across diverse domains, illustrating methods through examples, including recommendations in e-commerce contexts.
E N D
Classifying based on similarities: Van Gogh Monet Van Gogh Or Monet ?
the Similarity-based Classification Problem (paintings) (painter)
Examples of Similarity Functions Computational Biology • Smith-Waterman algorithm (Smith & Waterman, 1981) • FASTA algorithm (Lipman & Pearson, 1985) • BLAST algorithm (Altschul et al., 1990) Computer Vision • Tangent distance (Duda et al., 2001) • Earth mover’s distance (Rubner et al., 2000) • Shape matching distance (Belongie et al., 2002) • Pyramid match kernel (Grauman & Darrell, 2007) Information Retrieval • Levenshtein distance (Levenshtein, 1966) • Cosine similarity between tf-idf vectors (Manning & Schütze, 1999)
Example: Amazon similarity 96 books 96 books
Example: Amazon similarity 96 books 96 books
Example: Amazon similarity 96 books Eigenvalues Rank 96 books
Well, let’s just make S be a kernel matrix Flip, Clip or Shift? Best bet is Clip. 0 0
Well, let’s just make S be a kernel matrix Learn the best kernel matrix for the SVM: (Luss NIPS 2007, Chen et al. ICML 2009)
Let the similarities to the training samples be features • SVM (Graepel et al., 1998; Liao & Noble, 2003) • Linear programming (LP) machine (Graepel et al., 1999) • Linear discriminant analysis (LDA) (Pekalska et al., 2001) • Quadratic discriminant analysis (QDA) (Pekalska & Duin, 2002) • Potential support vector machine (P-SVM) (Hochreiter & Obermayer, 2006; Knebel et al., 2008)
Weighted Nearest-Neighbors Take a weighted vote of the k-nearest-neighbors: Algorithmic parallel of the exemplar model of human learning. ?
Weighted Nearest-Neighbors Take a weighted vote of the k-nearest-neighbors: Algorithmic parallel of the exemplar model of human learning.
Design Goals for the Weights ? Design Goal 1 (Affinity):wi should be an increasing function of ψ(x, xi).
Design Goals for the Weights (Chen et al. JMLR 2009) ? Design Goal 2 (Diversity):wi should be a decreasing function of ψ(xi, xj).
Linear Interpolation Weights Linear interpolation weights will meet these goals:
Linear Interpolation Weights Linear interpolation weights will meet these goals:
LIME weights Linear interpolation weights will meet these goals: Linear interpolation with maximum entropy (LIME) weights (Gupta et al., IEEE PAMI 2006):
LIME weights Linear interpolation weights will meet these goals: Linear interpolation with maximum entropy (LIME) weights (Gupta et al., IEEE PAMI 2006):
LIME weights Linear interpolation weights will meet these goals: Linear interpolation with maximum entropy (LIME) weights (Gupta et al., IEEE PAMI 2006):
LIME weights Linear interpolation weights will meet these goals: Linear interpolation with maximum entropy (LIME) weights (Gupta et al., IEEE PAMI 2006):
Kernelize Linear Interpolation regularizes the variance of the weights
Kernelize Linear Interpolation only need inner products – can replace with kernel or similarities!
KRI Weights Satisfy Design Goals Kernel ridge interpolation (KRI) weights:
KRI Weights Satisfy Design Goals Kernel ridge interpolation (KRI) weights: affinity:
KRI Weights Satisfy Design Goals Kernel ridge interpolation (KRI) weights: diversity:
KRI Weights Satisfy Design Goals Kernel ridge interpolation (KRI) weights:
KRI Weights Satisfy Design Goals Kernel ridge interpolation (KRI) weights: Remove the constraints on the weights: Can show equivalent to local ridge regression: KRR weights.
Weighted k-NN: Example 1 KRI weights KRR weights
Weighted k-NN: Example 2 KRI weights KRR weights
Weighted k-NN: Example 3 KRI weights KRR weights