Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 116 tok/s
Gemini 2.5 Pro 50 tok/s Pro
GPT-5 Medium 24 tok/s Pro
GPT-5 High 22 tok/s Pro
GPT-4o 59 tok/s Pro
Kimi K2 199 tok/s Pro
GPT OSS 120B 437 tok/s Pro
Claude Sonnet 4.5 38 tok/s Pro
2000 character limit reached

Design of Binary Quantizers for Distributed Detection under Secrecy Constraints (1410.8100v2)

Published 29 Oct 2014 in cs.IT, cs.CR, cs.SY, and math.IT

Abstract: In this paper, we investigate the design of distributed detection networks in the presence of an eavesdropper (Eve). We consider the problem of designing binary quantizers at the sensors that maximize the Kullback-Leibler (KL) Divergence at the fusion center (FC), subject to a tolerable constraint on the KL Divergence at Eve. In the case of i.i.d. received symbols at both the FC and Eve, we prove that the structure of the optimal binary quantizers is a likelihood ratio test (LRT). We also present an algorithm to find the threshold of the optimal LRT, and illustrate it for the case of Additive White Gaussian Noise (AWGN) observation models at the sensors. In the case of non-i.i.d. received symbols at both FC and Eve, we propose a dynamic-programming based algorithm to find efficient quantizers at the sensors. Numerical results are presented to illustrate the performance of the proposed network design.

Citations (25)

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.