1 / 26

Property Testing of Data Dimensionality

Property Testing of Data Dimensionality. Robert Krauthgamer ICSI and UC Berkeley Joint work with Ori Sasson (Hebrew U.). Data dimensionality. The analysis of large volumes of complex data is required in many disciplines.

duy
Download Presentation

Property Testing of Data Dimensionality

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Property Testing of Data Dimensionality Robert Krauthgamer ICSI and UC Berkeley Joint work with Ori Sasson (Hebrew U.)

  2. Data dimensionality • The analysis of large volumes of complex data is required in many disciplines. • Such data is frequently represented by vectors in a high-dimensional vector space. • E.g., sequential biological data (genome, proteins) • A common method of representing data is feature extraction (vector representation in feature space). • Images databases • Text corpora (via latent semantic indexing) Testing Data Dimensionality

  3. The issue of dimension • High-dimensional data is difficult to work with. • Complexity of many operations is heavily dependent (e.g. exponentially) on the dimension. • Real-life data often adheres to a low-dimensional structure • Which allows to effectively reduce the dimension. • E.g. in R2: • Dimensionality Reduction: Mapping into low-dimensional space (while preserving most of the data “structure”) • Trade-off accuracy for computational efficiency Testing Data Dimensionality

  4. Linear Structure Metric Structure Dimensionality reduction methods • Singular Value Decomposition (SVD) • I.e., low-rank matrix approximation. • Practical variants: Multidimensional Scaling (MDS), Principal Component Analysis (PCA) • Low-distortion embedding in low-dimensional lp • Of any Euclidean metric [Johnson-Lindenstrauss’86] • Of any metric [Bourgain’86, Linial-London-Rabinovich’93]. • Other methods, e.g. combinatorial feature selection [Charikar-Guruswami-Kumar-Rajagopalan-Sahai’00] Testing Data Dimensionality

  5. Property testing framework Relaxed decision problems: Determine whether • The input has a property P, or • The input is far from having the property P, i.e. it needs to be modified significantly in order to have the property. Goal: Obtain • Randomized algorithms (correct with probability  2/3), • Whose complexity is low (does not depend on input size). Trivial example: Testing if an input list contains only 0’s or e-fraction of the entries are not 0 – with O(1/e) queries. Testing Data Dimensionality

  6. Testing data dimensionality Given a data set S, determine whether • S has at most a (fixed) dimension d, or • S is e-far from having this property, • i.e. at least an e-fraction of the entries of (a representation of S) needs to be modified for S to have the property. Technicalities: • Interpretation of dimension (i.e. type of structure) • Representation of S • Assume it affects both query mechanism and farness measure Testing Data Dimensionality

  7. Our results – Testing for linear structure • Algorithm for testing whether vectors v1,…,vn lie in linear (or affine) subspace of dimension  d. • Algorithm queries O(d/e) vectors. • Holds for every vector space V. • Algorithm for testing whether a matrix Amn has rank  d. • Algorithm queries the entries of an O(d/e)  O(d/e) submatrix. • Holds for matrices over any field F. (Both algorithms have one-sided error.) Testing Data Dimensionality

  8. Our results – Testing for metric structure • Testing whether v1,…,vn l2mcan be embedded intol2d • Isometrically - achieved by querying O(d/e) vectors (corollary). • With distortionD<1/e - requires querying W((n/D)1/2) vectors. • With perturbationd>0 - requires W(min{n1/2 , m/log m}) queries. • Testing whether vectors v1,…,vnl1mcan be embedded isometrically into l1d requires querying W(n1/4) vectors. (Lower bounds are for algorithms with two-sided error.) Testing Data Dimensionality

  9. Our results – Testing metrics and norm • Algorithm for testing whether a matrix Mnn is the distances matrix of a d-dimensional Euclidean metric. • Algorithm queries the entries of an O(d/e) O(d/e) submatrix. • Slight improvement over O((dlog d)/e) O((dlog d)/e) of [Parnas-Ron’01]. • Algorithm for testing whether a vector has lp-norm  . • Algorithm queries O(e-3 log 1/e) entries (with two-sided error). • Holds for any p and. • Allows to test the Frobenius norm of a matrix (such as the difference between a matrix and its low-rank approximation). Testing Data Dimensionality

  10. Property testing origins • Introduced by [Rubinfeld-Sudan’96] • Testing algebraic properties of functions • Many PCPs involve testing of encodings • E.g. low-degree polynomials, Hadamard code, long code • Testing of combinatorial properties initiated by [Goldreich-Goldwasser-Ron’98] • They focused on graph properties (e.g. coloring). • Later works considered testing monotonicity of functions, satisfiability of formulas, regularity of languages, equality of distributions, clustering of Euclidean vectors, metric spaces etc. Testing Data Dimensionality

  11. Related work • Property testing • Testing whether a distances matrix represents a tree metric, ultra-metric, or a low-dimensional Euclidean metric [Parnas-Ron’01]. • Testing properties of Euclidean vectors, e.g. clustering [Alon-Dar-Parnas-Ron’00] and convexity [Czumaj-Sohler-Ziegler’00]. • Testing various matrix properties, e.g. monotonicity [Newman-Fischer’01]. • Fast low-rank approximation (by sampling) • [Frieze-Kannan-Vempala’98, Achlioptas-McSherry’01] • Farness measure considers the magnitude of the changes. • Sampling depends on input size (unless input is “uniform”). Testing Data Dimensionality

  12. Other related work • Finite point criterion for lpd– embeddability. • Namely, the minimum fp(d) such that (any) metric space embeds in lpd iff every fp(d) of its points do. • For p = 2, [Menger’28] showed fp(d) = d+3 . • For p = 1 and any d > 2, [Bandelt-Chepoi-Laurent’98] showed f1(d) d2-1, but it is not known whether f1(d) is finite. • Our results for l1and l2spaces establish somewhat similar bounds for a relaxed version of this question. Testing Data Dimensionality

  13. Algorithm for testing linear structure Thm 1. Testing whether a set of vectors Slies in a subspace of dimension dcan be achieved with O(d/e)queries. The algorithm. • Query O(d/e) vectors of S uniformly at random. • Accept if (and only if) the queried vectors lie in a linear (or affine) subspace of dimension  d. Testing Data Dimensionality

  14. Proof of testing linear structure Proof (correctness). Algorithm always accepts a data set S of dimension d. Let S be e-far from having dimension d. • Consider sampling the O(d/e) vectors one by one. • Let Xtbethe dimension of the subspace spanned by the first t sampled vectors. • Lemma 1.Pr[Xt+1 = Xt + 1 | Xt d] e . • Proof. Since S is e-far from having dimension d, the subspace spanned by the first t sampled vectors contains less than (1-e)-fraction of the vectors of S. Testing Data Dimensionality

  15. A technical lemma • Lemma 2. Let 0 X0 X1 X2 ... be random variables. If Pr[Xt+1= Xt + 1 | Xtd] e for all t  0, then for t*= 8d/ewe have Pr[Xt*  d] < 1/3. • Proof sketch.Xthas binomial distribution as long as Xtd. Then E[Xt*]8d and using Chernoff Pr[Xt* d] < 1/3. So with probability  2/3 we have Xt*> d and the algorithm rejects (for S that is e-far from dimension d). This completes the proof of Thm 1. • Similar approach allows to test if a matrix is low-rank and for distances matrix (slight improvement over [Parnas-Ron’01]). Testing Data Dimensionality

  16. Lower bound for l1 Thm 2. Testing whether nvectors in l1mcan be embedded isometrically into l1drequires querying W(n1/4)vectors. • Consider first algorithms with one-sided error. • Suppose d=1, m=2. • Consider the following point set S: • S is 1/24-far from l1d-embeddability • because every “” cannot • be embedded in the line. Testing Data Dimensionality

  17. Lower bound for l1 with one-sided error • Assume there is an algorithm that queries t << n1/2 points. • WLOG it sees a “random” sample of S. • With high probability 1 – O(t2/n) = 1 – o(1) • The sample contains no two points at distance O(1) from each other. • Then sample is l1d–embeddable (since there is a geodesic line going through all its points). • And so algorithm must accept S. • Contradiction (since S is 1/24-far). Testing Data Dimensionality

  18. Lower bound for l1 with two-sided error • We (randomly) create from S another data set S’ such that • S’ embeds in the line (WHP 1-o(1)). • The algorithm’s view of S differs from its view of S’ with probability o(1), • So probabilities of accepting S vs. that of S’ differ by o(1)<<1/3. Contradiction. • Here (to prove Thm 2): • Create S’ by choosing r <<n1/2 random points from S and duplicating each one n/r times. • Then a sample of << r1/2 points from S,S’ is almost the same. These inputs look the same Testing Data Dimensionality

  19. Lower bound for l2with perturbation Thm 3. Testing whether nvectors in l2mcan be perturbed by dto be l2d– embeddable requires W(min{n1/2 , m/log m})queries. • Let d=0 (I.e. testing if the vectors are in a ball of radius d). • Consider a sphere of radius d’ = d(1+1/2n) in l2m. • Let S’ consist of n random vectors from this sphere. • Let S consist of n/2 random vectors from the sphere and their n/2 antipodal vectors (-v). Testing Data Dimensionality

  20. Lower bound for l2with perturbation • WHP, the vectors of S’ are in a ball of radius d • By concentration of measure, WHP they are nearly orthogonal, e.g. the distance between every two is roughly d2. • In fact, WHP they are all at distance <d from their “center of mass”, as claimed. YES S’ Concentration of measure Testing Data Dimensionality

  21. Antipodals Lower bound for l2with perturbation • S is 1/2-far from being in a ball of radius d • Because the distance between antipodal vectors in S is 2d’ > 2d. • Assume algorithm queries << n1/2 • WHP view of S, S’ is the same. • So, probability of accepting S andS’ should differ by o(1). • Contradiction. This proves Thm 3. S NO Testing Data Dimensionality

  22. Lower bound for l2with distortion Thm 4: Testing whether nvectors in l2mcan be embedded in l2dwith distortion D<1/e requires W((n/D)1/2)queries. • Let d=1 (embedding into a line with distortion D). • Consider a unit circle with equally spaced 10D points. • Let S consist of points from n/10D (far apart) parallel copies of this circle in R3. Testing Data Dimensionality

  23. Lower bound for l2with distortion • S is 1/10D-far from having an embedding with distortion D • Since embedding each cycle into the line requires distortion > D. NO 10D points Testing Data Dimensionality

  24. Lower bound for one-sided error • Assume algorithm queries << (n/D)1/2 points of S • WLOG it sees a “random” sample of S. • WHP, this sample contains at most one point from each circle, • And then it can be embedded with distortion < D into the line (by mapping each point to its circle’s center). • So WHP algorithm must accept S. Contradiction. YES 10D points Testing Data Dimensionality

  25. Lower bound for two-sided error • We create S’ by choosing one point from each circle of S and duplicating it 10D times. • Then S’ can be embedded with distortion < D into the line. • WHP view of << (n/D)1/2 points from S is the same as from S’. • So, probability of accepting S andS’ should differ by o(1). • This proves Thm 4. Testing Data Dimensionality

  26. Future research • Testing whether • A matrix spectral norm ||A||2 is small. • A distances matrix represents metric (triangle inequality). • A distances matrix represents an l1d– metric. • A distances matrix represents an approximate l2d– metric. • Testing with farness measure that depends on magnitude • a la [Frieze-Kannan-Vempala’98, Achlioptas-McSherry’01] Testing Data Dimensionality

More Related