Result: On Optimal Data Compression in Multiterminal Statistical Inference

Title:
On Optimal Data Compression in Multiterminal Statistical Inference
Authors:
Source:
IEEE transactions on information theory. 57(9):5577-5587
Publisher Information:
New York, NY: Institute of Electrical and Electronics Engineers, 2011.
Publication Year:
2011
Physical Description:
print, 12 ref
Original Material:
INIST-CNRS
Document Type:
Academic journal Article
File Description:
text
Language:
English
Author Affiliations:
RIKEN Brain Science Institute, Hirosawa 2-1, Wako-shi, Saitama 351-0198, Japan
ISSN:
0018-9448
Rights:
Copyright 2015 INIST-CNRS
CC BY 4.0
Sauf mention contraire ci-dessus, le contenu de cette notice bibliographique peut être utilisé dans le cadre d’une licence CC BY 4.0 Inist-CNRS / Unless otherwise stated above, the content of this bibliographic record may be used under a CC BY 4.0 licence by Inist-CNRS / A menos que se haya señalado antes, el contenido de este registro bibliográfico puede ser utilizado al amparo de una licencia CC BY 4.0 Inist-CNRS
Notes:
Telecommunications and information theory
Accession Number:
edscal.24554445
Database:
PASCAL Archive

Further Information

The multiterminal theory of statistical inference deals with the problem of estimating or testing the correlation of letters generated from two (or many) correlated information sources under the restriction of a certain transmission rate for each source. A typical example is two binary sources with joint probability p(x, y) where the correlation of x and y is to be tested or estimated. Given n iid observations xn = x1 ··· xn and yn = y1 ··· yn, only k = rn (0 < r < 1) bits each can be transmitted to a common destination. What is the optimal data compression for statistical inference? A simple idea is to send the first k letters of xn and yn. A simpler problem is the helper case where the optimal data compression of xn is searched for under the condition that all of yn are transmitted. It is a long standing problem to determine if there is a better data compression scheme than this simple scheme of sending first k letters. The present paper searches for the optimal data compression under the framework of linear-threshold encoding and shows that there is a better data compression scheme depending on the value of correlation. To this end, we evaluate the Fisher information in the class of linear-threshold compression schemes. It is also proved that the simple scheme is optimal when x and y are independent or their correlation is not too large.