Author | Andersen, Reid ♦ Gharan, Shayan Oveis ♦ Peres, Yuval ♦ Trevisan, Luca |
Source | ACM Digital Library |
Content type | Text |
Publisher | Association for Computing Machinery (ACM) |
File Format | |
Copyright Year | ©2016 |
Language | English |
Subject Domain (in DDC) | Computer science, information & general works ♦ Data processing & computer science |
Subject Keyword | Cheeger’s inequality ♦ Local graph clustering ♦ Random walks ♦ Sparsest cut |
Abstract | Spectral partitioning is a simple, nearly linear time algorithm to find sparse cuts, and the Cheeger inequalities provide a worst-case guarantee for the quality of the approximation found by the algorithm. A local graph partitioning algorithm finds a set of vertices with small conductance (i.e., a sparse cut) by adaptively exploring part of a large graph $\textit{G},$ starting from a specified vertex. For the algorithm to be local, its complexity must be bounded in terms of the size of the set that it outputs, with at most a weak dependence on the number $\textit{n}$ of vertices in $\textit{G}.$ Previous local partitioning algorithms find sparse cuts using random walks and personalized PageRank [Spielman and Teng 2013; Andersen et al. 2006]. In this article, we introduce a simple randomized local partitioning algorithm that finds a sparse cut by simulating the volume-biased evolving set process, which is a Markov chain on sets of vertices. We prove that for any ε > 0, and any set of vertices $\textit{A}$ that has conductance at most φ, for at least half of the starting vertices in $\textit{A}$ our algorithm will output (with constant probability) a set of conductance $\textit{O}(&sqrt;φ$ /ε). We prove that for a given run of the algorithm, the expected ratio between its computational complexity and the volume of the set that it outputs is $vol(A)^{ε}φ^{&mins;1/2}polylog(n),$ where $vol(\textit{A})$ = $∑\textit{v}$ ∈ $\textit{A}\textit{d}(\textit{v})$ is the volume of the set $\textit{A}.$ This gives an algorithm with the same guarantee (up to a constant factor) as the Cheeger’s inequality that runs in time slightly superlinear in the size of the output. This is the first sublinear (in the size of the input) time algorithm with almost the same guarantee as the Cheeger’s inequality. In comparison, the best previous local partitioning algorithm, by Andersen et al. [2006], has a worse approximation guarantee of $\textit{O}(&sqrt;φ$ log $\textit{n})$ and a larger ratio of $φ^{&mins;1}$ $polylog(\textit{n})$ between the complexity and output volume. As a by-product of our results, we prove a bicriteria approximation algorithm for the expansion profile of any graph. For 0 < $\textit{k}$ ⩽ $vol(\textit{V})/2,$ let $φ(\textit{k})$ ≔ min $\textit{S}:$ $vol(\textit{S})$ ⩽ $\textit{k}φ(\textit{S}).$ There is a polynomial time algorithm that, for any $\textit{k},$ ε > 0, finds a set $\textit{S}$ of volume $vol(\textit{S})$ ⩽ $\textit{O}(\textit{k}1$ + ε) and expansion $φ(\textit{S})≤$ $\textit{O}(&sqrt;φ$ $(\textit{k})/ε).$ As a new technical tool, we show that for any set $\textit{S}$ of vertices of a graph, a lazy $\textit{t}-step$ random walk started from a randomly chosen vertex of $\textit{S}$ will remain entirely inside $\textit{S}$ with probability at least (1 ™ $φ(S)/2)^{t}.$ This itself provides a new lower bound to the uniform mixing time of any finite state reversible Markov chain. |
ISSN | 00045411 |
Age Range | 18 to 22 years ♦ above 22 year |
Educational Use | Research |
Education Level | UG and PG |
Learning Resource Type | Article |
Publisher Date | 2016-05-04 |
Publisher Place | New York |
e-ISSN | 1557735X |
Journal | Journal of the ACM (JACM) |
Volume Number | 63 |
Issue Number | 2 |
Page Count | 31 |
Starting Page | 1 |
Ending Page | 31 |
Ministry of Human Resource Development (MHRD) under its National Mission on Education through Information and Communication Technology (NMEICT) has initiated the National Digital Library of India (NDLI) project to develop a framework of virtual repository of learning resources with a single-window search facility. Filtered and federated searching is employed to facilitate focused searching so that learners can find out the right resource with least effort and in minimum time. NDLI is designed to hold content of any language and provides interface support for leading vernacular languages, (currently Hindi, Bengali and several other languages are available). It is designed to provide support for all academic levels including researchers and life-long learners, all disciplines, all popular forms of access devices and differently-abled learners. It is being developed to help students to prepare for entrance and competitive examinations, to enable people to learn and prepare from best practices from all over the world and to facilitate researchers to perform inter-linked exploration from multiple sources. It is being developed at Indian Institute of Technology Kharagpur.
NDLI is a conglomeration of freely available or institutionally contributed or donated or publisher managed contents. Almost all these contents are hosted and accessed from respective sources. The responsibility for authenticity, relevance, completeness, accuracy, reliability and suitability of these contents rests with the respective organization and NDLI has no responsibility or liability for these. Every effort is made to keep the NDLI portal up and running smoothly unless there are some unavoidable technical issues.
Ministry of Human Resource Development (MHRD), through its National Mission on Education through Information and Communication Technology (NMEICT), has sponsored and funded the National Digital Library of India (NDLI) project.
Phone: +91-3222-282435
For any issue or feedback, please write to ndl-support@iitkgp.ac.in