Emergent Mind

A Distributed Palette Sparsification Theorem

(2301.06457)
Published Jan 16, 2023 in cs.DC and cs.DS

Abstract

The celebrated palette sparsification result of [Assadi, Chen, and Khanna SODA'19] shows that to compute a $\Delta+1$ coloring of the graph, where $\Delta$ denotes the maximum degree, it suffices if each node limits its color choice to $O(\log n)$ independently sampled colors in ${1, 2, \dots, \Delta+1}$. They showed that it is possible to color the resulting sparsified graph -- the spanning subgraph with edges between neighbors that sampled a common color, which are only $\tilde{O}(n)$ edges -- and obtain a $\Delta+1$ coloring for the original graph. However, to compute the actual coloring, that information must be gathered at a single location for centralized processing. We seek instead a local algorithm to compute such a coloring in the sparsified graph. The question is if this can be achieved in $\operatorname{poly}(\log n)$ distributed rounds with small messages. Our main result is an algorithm that computes a $\Delta+1$-coloring after palette sparsification with $O(\log2 n)$ random colors per node and runs in $O(\log2 \Delta + \log3 \log n)$ rounds on the sparsified graph, using $O(\log n)$-bit messages. We show that this is close to the best possible: any distributed $\Delta+1$-coloring algorithm that runs in the LOCAL model on the sparsified graph, given by palette sparsification, for any $\operatorname{poly}(\log n)$ colors per node, requires $\Omega(\log \Delta / \log\log n)$ rounds. This distributed palette sparsification result leads to the first $\operatorname{poly}(\log n)$-round algorithms for $\Delta+1$-coloring in two previously studied distributed models: the Node Capacitated Clique, and the cluster graph model.

We're not able to analyze this paper right now due to high demand.

Please check back later (sorry!).

Generate a summary of this paper on our Pro plan:

We ran into a problem analyzing this paper.

Newsletter

Get summaries of trending comp sci papers delivered straight to your inbox:

Unsubscribe anytime.