Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 42 tok/s
Gemini 2.5 Pro 53 tok/s Pro
GPT-5 Medium 17 tok/s Pro
GPT-5 High 13 tok/s Pro
GPT-4o 101 tok/s Pro
Kimi K2 217 tok/s Pro
GPT OSS 120B 474 tok/s Pro
Claude Sonnet 4 36 tok/s Pro
2000 character limit reached

Sufficient Conditions on the Optimality of Myopic Sensing in Opportunistic Channel Access: A Unifying Framework (1309.3692v1)

Published 14 Sep 2013 in cs.IT and math.IT

Abstract: This paper considers a widely studied stochastic control problem arising from opportunistic spectrum access (OSA) in a multi-channel system, with the goal of providing a unifying analytical framework whereby a number of prior results may be viewed as special cases. Specifically, we consider a single wireless transceiver/user with access to $N$ channels, each modeled as an iid discrete-time two-state Markov chain. In each time step the user is allowed to sense $k\leq N$ channels, and subsequently use up to $m\leq k$ channels out of those sensed to be available. Channel sensing is assumed to be perfect, and for each channel use in each time step the user gets a unit reward. The user's objective is to maximize its total discounted or average reward over a finite or infinite horizon. This problem has previously been studied in various special cases including $k=1$ and $m=k\leq N$, often cast as a restless bandit problem, with optimality results derived for a myopic policy that seeks to maximize the immediate one-step reward when the two-state Markov chain model is positively correlated. In this paper we study the general problem with $1\leq m\leq k\leq N$, and derive sufficient conditions under which the myopic policy is optimal for the finite and infinite horizon reward criteria, respectively. It is shown that these results reduce to those derived in prior studies under the corresponding special cases, and thus may be viewed as a set of unifying optimality conditions. Numerical examples are also presented to highlight how and why an optimal policy may deviate from the otherwise-optimal myopic sensing given additional exploration opportunities, i.e., when $m<k$.

Citations (12)

Summary

We haven't generated a summary for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Lightbulb On Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.