doubling dimension
Recently Published Documents


TOTAL DOCUMENTS

25
(FIVE YEARS 10)

H-INDEX

6
(FIVE YEARS 0)

Author(s):  
Johannes Blum

AbstractWe study the k-Center problem, where the input is a graph $$G=(V,E)$$ G = ( V , E ) with positive edge weights and an integer k, and the goal is to select k center vertices $$C \subseteq V$$ C ⊆ V such that the maximum distance from any vertex to the closest center vertex is minimized. In general, this problem is $$\mathsf {NP}$$ NP -hard and cannot be approximated within a factor less than 2. Typical applications of the k-Center problem can be found in logistics or urban planning and hence, it is natural to study the problem on transportation networks. Common characterizations of such networks are graphs that are (almost) planar or have low doubling dimension, highway dimension or skeleton dimension. It was shown by Feldmann and Marx that k-Center is $$\mathsf {W[1]}$$ W [ 1 ] -hard on planar graphs of constant doubling dimension when parameterized by the number of centers k, the highway dimension $$hd$$ hd and the pathwidth $$pw$$ pw (Feldmann and Marx 2020). We extend their result and show that even if we additionally parameterize by the skeleton dimension $$\kappa $$ κ , the k-Center problem remains $$\mathsf {W[1]}$$ W [ 1 ] -hard. Moreover, we prove that under the Exponential Time Hypothesis there is no exact algorithm for k-Center that has runtime $$f(k,hd,pw,\kappa ) \cdot \vert V \vert ^{o(pw+ \kappa + \sqrt{k+hd})}$$ f ( k , h d , p w , κ ) · | V | o ( p w + κ + k + h d ) for any computable function f.


Author(s):  
Hu Ding ◽  
Fan Yang ◽  
Mingyue Wang

The density based clustering method Density-Based Spatial Clustering of Applications with Noise (DBSCAN) is a popular method for outlier recognition and has received tremendous attention from many different areas. A major issue of the original DBSCAN is that the time complexity could be as large as quadratic. Most of existing DBSCAN algorithms focus on developing efficient index structures to speed up the procedure in low-dimensional Euclidean space. However, the research of DBSCAN in high-dimensional Euclidean space or general metric spaces is still quite limited, to the best of our knowledge. In this paper, we consider the metric DBSCAN problem under the assumption that the inliers (excluding the outliers) have a low doubling dimension. We apply a novel randomized k-center clustering idea to reduce the complexity of range query, which is the most time consuming step in the whole DBSCAN procedure. Our proposed algorithms do not need to build any complicated data structures and are easy to implement in practice. The experimental results show that our algorithms can significantly outperform the existing DBSCAN algorithms in terms of running time.


Author(s):  
Martin Lotz

We show that recent results on randomized dimension reduction schemes that exploit structural properties of data can be applied in the context of persistent homology. In the spirit of compressed sensing, the dimension reduction is determined by the Gaussian width of a structure associated with the dataset, rather than its size, and such a reduction can be computed efficiently. We further relate the Gaussian width to the doubling dimension of a finite metric space, which appears in the study of the complexity of other methods for approximating persistent homology. We can, therefore, literally replace the ambient dimension by an intrinsic notion of dimension related to the structure of the data.


Author(s):  
Hu Ding ◽  
Mingquan Ye

In real-world, many problems can be formulated as the alignment between two geometric patterns. Previously, a great amount of research focus on the alignment of 2D or 3D patterns, especially in the field of computer vision. Recently, the alignment of geometric patterns in high dimension finds several novel applications, and has attracted more and more attentions. However, the research is still rather limited in terms of algorithms. To the best of our knowledge, most existing approaches for high dimensional alignment are just simple extensions of their counterparts for 2D and 3D cases, and often suffer from the issues such as high complexities. In this paper, we propose an effective framework to compress the high dimensional geometric patterns and approximately preserve the alignment quality. As a consequence, existing alignment approach can be applied to the compressed geometric patterns and thus the time complexity is significantly reduced. Our idea is inspired by the observation that high dimensional data often has a low intrinsic dimension. We adopt the widely used notion “doubling dimension” to measure the extents of our compression and the resulting approximation. Finally, we test our method on both random and real datasets; the experimental results reveal that running the alignment algorithm on compressed patterns can achieve similar qualities, comparing with the results on the original patterns, but the running times (including the times cost for compression) are substantially lower.


Sign in / Sign up

Export Citation Format

Share Document