Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 27 tok/s
Gemini 2.5 Pro 46 tok/s Pro
GPT-5 Medium 23 tok/s Pro
GPT-5 High 29 tok/s Pro
GPT-4o 70 tok/s Pro
Kimi K2 117 tok/s Pro
GPT OSS 120B 459 tok/s Pro
Claude Sonnet 4 34 tok/s Pro
2000 character limit reached

Online Multistage Subset Maximization Problems (1905.04162v1)

Published 10 May 2019 in cs.DS and cs.CC

Abstract: Numerous combinatorial optimization problems (knapsack, maximum-weight matching, etc.) can be expressed as \emph{subset maximization problems}: One is given a ground set $N={1,\dots,n}$, a collection $\mathcal{F}\subseteq 2N$ of subsets thereof such that $\emptyset\in\mathcal{F}$, and an objective (profit) function $p:\mathcal{F}\rightarrow\mathbb{R}_+$. The task is to choose a set $S\in\mathcal{F}$ that maximizes $p(S)$. We consider the \emph{multistage} version (Eisenstat et al., Gupta et al., both ICALP 2014) of such problems: The profit function $p_t$ (and possibly the set of feasible solutions $\mathcal{F}_t$) may change over time. Since in many applications changing the solution is costly, the task becomes to find a sequence of solutions that optimizes the trade-off between good per-time solutions and stable solutions taking into account an additional similarity bonus. As similarity measure for two consecutive solutions, we consider either the size of the intersection of the two solutions or the difference of $n$ and the Hamming distance between the two characteristic vectors. We study multistage subset maximization problems in the \emph{online} setting, that is, $p_t$ (along with possibly $\mathcal{F}_t$) only arrive one by one and, upon such an arrival, the online algorithm has to output the corresponding solution without knowledge of the future. We develop general techniques for online multistage subset maximization and thereby characterize those models (given by the type of data evolution and the type of similarity measure) that admit a constant-competitive online algorithm. When no constant competitive ratio is possible, we employ lookahead to circumvent this issue. When a constant competitive ratio is possible, we provide almost matching lower and upper bounds on the best achievable one.

Citations (21)

Summary

We haven't generated a summary for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-Up Questions

We haven't generated follow-up questions for this paper yet.