Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 177 tok/s
Gemini 2.5 Pro 43 tok/s Pro
GPT-5 Medium 26 tok/s Pro
GPT-5 High 25 tok/s Pro
GPT-4o 119 tok/s Pro
Kimi K2 202 tok/s Pro
GPT OSS 120B 439 tok/s Pro
Claude Sonnet 4.5 38 tok/s Pro
2000 character limit reached

Parameterized (Approximate) Defective Coloring (1801.03879v1)

Published 11 Jan 2018 in cs.DS and cs.CC

Abstract: In Defective Coloring we are given a graph $G = (V, E)$ and two integers $\chi_d, \Delta*$ and are asked if we can partition $V$ into $\chi_d$ color classes, so that each class induces a graph of maximum degree $\Delta*$. We investigate the complexity of this generalization of Coloring with respect to several well-studied graph parameters, and show that the problem is W-hard parameterized by treewidth, pathwidth, tree-depth, or feedback vertex set, if $\chi_d = 2$. As expected, this hardness can be extended to larger values of $\chi_d$ for most of these parameters, with one surprising exception: we show that the problem is FPT parameterized by feedback vertex set for any $\chi_d \ge 2$, and hence 2-coloring is the only hard case for this parameter. In addition to the above, we give an ETH-based lower bound for treewidth and pathwidth, showing that no algorithm can solve the problem in $n{o(pw)}$, essentially matching the complexity of an algorithm obtained with standard techniques. We complement these results by considering the problem's approximability and show that, with respect to $\Delta*$, the problem admits an algorithm which for any $\epsilon > 0$ runs in time $(tw/\epsilon){O(tw)}$ and returns a solution with exactly the desired number of colors that approximates the optimal $\Delta*$ within $(1 + \epsilon)$. We also give a $(tw){O(tw)}$ algorithm which achieves the desired $\Delta*$ exactly while 2-approximating the minimum value of $\chi_d$. We show that this is close to optimal, by establishing that no FPT algorithm can (under standard assumptions) achieve a better than $3/2$-approximation to $\chi_d$, even when an extra constant additive error is also allowed.

Citations (16)

Summary

We haven't generated a summary for this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube