Academia.eduAcademia.edu

Successively Structured Gaussian Two-terminal Source Coding

2008, Wireless Personal Communications

Wireless Pers Commun (2009) 48:485–510 DOI 10.1007/s11277-008-9534-x Successively Structured Gaussian Two-terminal Source Coding Hamid Behroozi · M. Reza Soleymani Published online: 20 June 2008 © Springer Science+Business Media, LLC. 2008 Abstract Multiterminal source coding refers to separate encoding and joint decoding of multiple correlated sources. Joint decoding requires all the messages to be decoded simultaneously which is exponentially more complex than a sequence of single-message decodings. Inspired by previous work on successive coding, we apply the successive Wyner-Ziv coding, which is inherently a low complexity approach of obtaining a prescribed distortion, to the two-terminal source coding scheme. First, we consider 1-helper problem where one source provides partial side information to the decoder to help the reconstruction of the main source. Our results show that the successive coding strategy is an optimal strategy in the sense of achieving the rate-distortion function. By developing connections between source encoding and data fusion steps, it is shown that the whole rate-distortion region for the 2-terminal source coding problem is achievable using the successive coding strategy. Comparing the performance of the sequential coding with the performance of the successive coding, we show that there is no sum-rate loss when the side information is not available at the encoder. This result is of special interest in some applications such as video coding where there are processing and storage constraints at the encoder. Finally, we provide an achievable rate-distortion region for the m-terminal source coding. Keywords Multiterminal source coding · m-helper problem · Partial side information · Rate-distortion region · Successive coding · Sequential coding This paper was presented in part at the IEEE International Symposium on Information Theory (ISIT), July 2006 and in part at the IEEE Vehicular Technology Conference (VTC), September 2006. H. Behroozi (B) Department of Mathematics and Statistics, Queen’s University, Kingston, ON, Canada K7L 3N6 e-mail: behroozi@mast.queensu.ca M. Reza Soleymani Electrical and Computer Engineering Department, Concordia University, Montreal, QC, Canada H3G 1M8 e-mail: msoleyma@ece.concordia.ca 123 486 H. Behroozi, M. R. Soleymani 1 Introduction The behavior of many wireless networks such as wireless sensor networks can be modeled by the Chief Executive Officer (CEO) problem [1–4]. This problem, which is closely related to multiterminal source coding problem, is an abstract model for remote sensing and distributed compression in wireless networks. In the CEO problem, a firm’s CEO is interested in an underlying source X which cannot be observed directly. The CEO deploys a team of L agents (sensors) to observe the source data sequence. The agents observe independently correlated (noisy) versions of the source and then communicate information about their observations to a single fusion center (FC) for further processing. The FC aims to form an optimal estimate of X by fusing information sequences received from the agents. The overall system is depicted in Fig. 1. The encoders are distributed and cannot cooperate to exploit the correlation between their inputs. Each encoder is subject to a transmission rate constraint. This constraint comes from the restrictions on the resources such as bandwidth and power that are available at each agent (e.g. sensor node). Thus, the key challenge in such a data gathering model is conserving the energy of the distributed nodes and maximizing their lifetime. It is shown that among the three power consumption tasks by a sensor node, i.e., sensing, data processing and communication, the most energy consumption part is related to the communication tasks. Since the sensor measurements are correlated, the correlation should be exploited to avoid redundant transmission. In fact, since the bit-rate directly impacts power consumption at a sensor node, by eliminating the data redundancy and reducing the communication load, the energy resources can be managed in an optimal manner. Therefore, an efficient source coding technique is the main requirement of a distributed network. In general, multiterminal source coding or distributed data compression refers to separate encoding and joint decoding of multiple correlated sources [5]. Over the last 30 years, significant effort has been made on finding a complete characterization of multiterminal rate-distortion region. But even concrete examples of the problem are hard to analyze. For instance, the complete characterization of the rate region for the 2-terminal source coding in the Gaussian case has been found recently [6]. Nevertheless, today multiterminal source coding is still of special interest; not only because it is an open problem of information theory [7], but also because of its application in wireless communication systems. In fact, the increasing attention given to new applications such as wireless video networks or distributed sensor networks is a reason for new interests in multiterminal source coding schemes. V1 Y1 V2 Y2 X Encoder 1 Encoder 2 Source U1 R1 U2 R2 . . . .. . . VL YL Encoder L Fusion Center (FC) X UL RL Fig. 1 The CEO model. The target data X is observed by L wireless agents (sensors) as Yi . Agents encode and transmit their observations through rate constrained noiseless channels to a FC. The FC obtains an estimate of the source X within an acceptable degree of fidelity 123 Successively Structured Gaussian Two-terminal Source Coding 487 Analog Channel Side Info Encoder Digital Channel Decoder Fig. 2 Digitally enhanced analog transmission [11] Source Coding with Side Information In the point-to-point communication there is only a single source of information at the encoder and decoder, i.e., the message at the encoder, and the received signal at the decoder. If any type of useful information is added to the encoder or decoder’s information, it would be called “side” information [8]. In many distributed applications, the receiver may have some prior information about the source. This prior knowledge which is correlated to the transmitted source is considered as the decoder side information. For instance, given a video sequence, one can predict a forthcoming frame at the decoder based on the previously decoded frames by extrapolation [9]. The predicted frame works as a side information at the decoder in order to reduce the transmission bit-rate. As an another example, assume that we want to upgrade an existing analog TV system [10,11]. Instead of replacing the whole system, a digital channel can be added to the existing system. The scenario is shown in Fig. 2. The output of the analog channel can be considered as a side information to obtain a better reconstruction of the source signal. Side information is often available to improve the rate-distortion performance of the system. Source coding with side information addresses encoding schemes that exploit the side information in order to reduce the transmission rate. 1.1 Multiterminal Source Coding The rate-distortion theory for the multiterminal source coding problem was first studied by Wyner and Ziv [12]. They obtained the rate-distortion function of a single source when the decoder can observe full resolution side information about the source. Today, Wyner-Ziv coding refers to lossy source coding with decoder side information. The important fact about the Wyner-Ziv coding is that it usually suffers rate loss compared to the lossy coding of X when the side information Y , which is correlated to the transmitted source X , is available at both the encoder and the decoder. One exception is when X and Y are jointly Gaussian with a mean squared-error (MSE) distortion measure [13]. There is no rate loss with the Wyner-Ziv coding in this case, which is of special interest in some applications such as video sensor networks since many image and video sources (after mean abstraction) can be modeled as jointly Gaussian [14]. This result is the dual of Costa’s dirty paper theorem for the channel coding with side information at the transmitter only [15]. In this case the side information is knowledge about the state of the channel, available at the encoder, but not the decoder. Costa’s dirty paper analogy could be stated as follow [15]: Consider a sheet of paper covered with independent dirt spots having normally distributed intensity. A writer wants to send a message with a limited amount of ink. He knows the location and intensity of dirt spots and can take advantage of that side information. The dirty paper, with the writer’s message on it, is then sent to the reader. It picks up more normally distributed dirt along its way to the reader. Therefore, the reader cannot distinguish the dirt spots from the ink marks applied by the writer. In [15] Costa shows that the reader can gain the same amount of information as 123 488 H. Behroozi, M. R. Soleymani X Y Decoder X Event Fig. 3 1-helper scenario in a video sensor network. Two sensors are deployed to measure an event. At the decoder, we are only interested in the signal reconstruction of the sensor X which is nearer to the event than other sensor, Y if the original dirt spots were known by the reader, too. In this work, we also focus on the Gaussian sources. The generalized Wyner-Ziv source coding for noisy encoder observations has appeared in [2]. This problem is also known as the remote source coding with decoder side information [5,16]. Similar rate-distortion analysis is presented in [3], and recently in [8]. Systematic lossy source/channel coding with uncoded side information at the decoder is presented in [10]. Oohama [17] gives the solution of the Wyner-Ziv problem with coded side information for the case of two sources. This problem is also known as the 1-helper problem. 1.1.1 One-helper Coding Scheme Consider the following scenario for the event detection: a video sensor is used to monitor an event and transmit limited rate information to a FC for further processing. Also, suppose another sensor is able to monitor the same scene from a different viewing position which is farther from the event than the first sensor and hence it does not worth to reproduce its signal at the FC. The scenario is shown in Fig. 3. Since both sensors monitor the same scene, their information is correlated. Therefore, the second sensor can provide partial side information to reduce the rate required at the first sensor and hence save transmission energy for a given distortion. In fact, the information of the second sensor can be used at the FC to obtain higher reliability and lower probability of detection error. We refer to the sensor providing side information as the helper while the sensor whose measurement is going to be reproduced is referred to as the main source [18]. This problem, which is also called the 1-helper problem, for the special case of correlated memoryless Gaussian sources and squared distortion measures is investigated by Oohama [17]. In the m-helper problem only the main sources is reconstructed while other m sources work as helpers [18–21]. 1.1.2 Two-Terminal Source Coding Scheme In this scheme, two correlated sources X and Y are separately encoded and sent to a single joint decoder, where the decoder reconstruct both sources. The corresponding coding scheme is shown in Fig. 4. X is encoded at rate R0 and Y is encoded at rate R1 . The joint decoder aims to obtain estimates of X and Y with average distortions D0 and D1 , respectively. In the previous video sensor network example, if FC aims to reconstruct both sensors’ observations, the scenario can be modeled by the 2-terminal source coding problem. Berger and Yeung [22] solve the rate-distortion problem for the situation in which the reconstruction of X must be perfect, while that of Y is subject to a distortion criterion. Oohama derives an outer region for the rate-distortion region of the 2-terminal source coding problem and demonstrates that the inner region obtained by Berger [5] and Tung [23] is partially tight [17]. Wagner et al. 123 Successively Structured Gaussian Two-terminal Source Coding n n X X ϕ0 Y ϕ0 (X ) n n Y 489 ϕ1 ϕ1 (Y ) ψ = (ψ0 ,ψ1) X Y n n Fig. 4 2-terminal source coding scheme. X and Y are two correlated sources with correlation coefficient ρ. Both sources are to be reconstructed at the decoder [6] complete the characterization of the rate region for the 2-terminal source coding problem by showing that the inner bound of Berger and Tung is also tight in the sum-rate. Oohama extends his results to more than two sources for a certain class of m + 1 correlated sources, where m source signals are independent noisy versions of the main source, i.e., X i = X 0 + Ni , i ∈ {1, 2, . . . , m}. In other words, sources X 1 to X m are conditionally independent given the source X 0 [24]. In this work, we consider multiterminal source coding for the case of two terminals. We use the successive coding strategy to obtain an achievable rate-distortion region for the 2-terminal source coding problem. 1.2 Successive Coding Strategy Previous research on coding of multiterminal schemes are based on joint decoding of all messages. Joint decoding requires all the messages to be decoded simultaneously which is exponentially more complex than a sequence of single-message decodings [8]. The decoding process cannot begin before all messages have been received at the joint decoder. Inspired by previous work on successive coding strategy, we apply the successive Wyner-Ziv coding to the Gaussian multiterminal source coding schemes. As it is described in [8], the encoder and decoder perform as follows: Encoder When an encoder encodes a message, it considers two things. First, its observations and second, its statistical knowledge about the messages that the decoder has already received from other nodes in the network. Therefore, previously decoded messages that are available at the decoder act as the decoder’s side information. Decoder At the decoder, instead of joint decoding, messages from sources are decoded sequentially in order to increase the fidelity of estimation at each decoding step. In other words, by obtaining more data at each stage, the decoder has additional side information to use in the next decoding step. This is a decentralized strategy because at each stage, only the knowledge sharing between each encoder and decoder is needed [8]. Its successive structuring provides flexibility to deal with distributed signal processing. By applying the successive coding, the available practical Wyner-Ziv coding techniques are applicable to more general distributed and multiterminal source coding problems. Hence, this strategy simplifies the analysis of the multiterminal source coding problem by decomposing the problem into some successive Wyner-Ziv coding cases. It allows us to derive an achievable rate region for multiterminal schemes. The successive coding approach is well-known in the research community. Some of the first applications of the successive coding are as follows: 123 490 H. Behroozi, M. R. Soleymani X Encoder 0 Dec 0 X Encoder 1 Dec 1 X Fig. 5 Successive refinement coding 1.2.1 Successive Refinement Source Coding In the successive refinement source coding [25–27] at first, the source will be described by a few bits of information. The description of the source can be improved when more bits of information are available. The two-stage successive refinement scenario is depicted in Fig. 5. Source X is encoded and transmitted through a rate-constraint channel. With rate R0 , decoder 0 reproduces X with distortion level D0 . At the next stage, encoder 1 sends a sequence at rate R1 to decoder 1. With both received sequences, decoder 1 will produce a more accurate reconstruction of X at distortion level D1 . The source is called successively refinable if the two rates simultaneously lie on the rate-distortion curve, i.e., R0 = R X (D0 ), R0 + R1 = R X (D1 ) (1) where R X (D) is the rate-distortion function of the source X at distortion level D. In general, a source is successively refinable if encoding in multiple stages incurs no rate loss as compared with optimal rate-distortion encoding. It is shown that Gaussian sources with the squared error distortion measure are successively refinable [26]. Steinberg and Merhav [28] consider the problem of successive refinement in the presence of side information. They answer the question whether such a progressive encoding causes rate loss as compared with a single stage Wyner- Ziv coding. It is shown that the jointly Gaussian sources (with the squared error distortion measure) are successively refinable in the Wyner-Ziv setting. It is also shown that there is no rate loss when the difference between the source and the side information is Gaussian and independent of the side information [29]. The characterization of the rate-distortion region for successive refinability for more than two-stage systems is presented in [30]. 1.2.2 CEO Problem The objective of the CEO problem, briefly presented in Sect. 1, is to determine the minimum achievable distortion under a sum-rate constraint. By sum-rate, we mean the total rate at which the agents may communicate information about their observations to the CEO. The special case of Gaussian source and noise statistics with a MSE distortion is called the quadratic Gaussian CEO problem [31]. For this case, the rate region is known [19,20,32,33]. Applications of the successive coding strategy in the CEO problem are presented in [8,34,35]. It is shown that the successive coding is an optimal coding strategy in the Gaussian CEO problem in the sense of achieving the sum-rate distortion function [36]. Since the application of successive coding strategy into the 2-terminal source coding problem is similar to the sequential coding of correlated sources [37], this coding scheme is described as follows. 123 Successively Structured Gaussian Two-terminal Source Coding 491 X Encoder 0 Dec 0 X Y Encoder 1 Dec 1 Y Fig. 6 Sequential coding of correlated sources 1.2.3 Sequential Coding of Correlated Sources Assume that we have two correlated sources X and Y . In the sequential coding scheme, as described in [37], the first encoder observes only the source X and encodes it at rate R0 such that the first decoder is able to reconstruct the source X with distortion D0 . Then, the second encoder, who has access to the source X , observes source Y and encodes it (with the encoding allowed to depend on X ) at rate R1 so that the second decoder, with the available side information from the first decoder, can reconstruct the source Y with distortion D1 . The coding scheme is depicted in Fig. 6. When the sources X and Y are the same source, the above coding scheme reduces to the successive refinement source coding scheme [37]. In this work, we evaluate the performance of the successive coding strategy for the problem of multiterminal lossy coding of correlated Gaussian sources. We address the problem from an information theoretic perspective and determine the rate region for 1-helper source coding scheme and 2-terminal source coding problem. Our main contributions can be summarized as follows: we show that for both the Gaussian 1-helper problem and the Gaussian 2-terminal source coding problem, successively structured Wyner-Ziv codes can achieve the rate-distortion bounds. Therefore, the high complexity optimal source code can be decomposed into a sequence of low complexity Wyner-Ziv codes. We support our derivatives by geometric sphere-packing arguments [38,39]. By comparing the results of the successive coding strategy for the 2-terminal source coding and the sequential coding of correlated Gaussian sources, we demonstrate that there is no sum-rate loss when the output of the first encoder is not available at the second encoder. The result is of special interest in some applications such as video coding where there are storage constraints at the encoder. This successive coding approach leads us to derive an inner bound for the rate region of the m-terminal source coding where all m correlated sources are reconstructed at the FC with specified distortions. The rest of this paper is organized as follows: Section 2 presents the system model and problem formulation. In Sect. 3, we use the successive coding strategy in order to obtain the rate-distortion functions of both 1-helper and 2-terminal source coding systems. The achievable rate-distortion region of the successively structured m-terminal source coding system is derived in Sect. 4. Section 5 concludes this paper. 2 Problem Formulation Let X and Y be correlated Gaussian random variables such that {(X t , Yt )}∞ t=0 are jointly stationary Gaussian memoryless sources. For each observation time t = 1, 2, 3, . . . , the random pair (X t , Yt ) takes a value in real space X × Y and has a probability density function (pdf) p X,Y (x, y) of N ∼ (0, ) where the covariance matrix  is given by 123 492 H. Behroozi, M. R. Soleymani =  σ X2 ρσ X σY ρσ X σY σY2  , −1 < ρ < 1. (2) ∞ We represent n independent instances of {X t }∞ t=1 and {Yt }t=1 by data sequences n n X = {X 1 , X 2 , . . . , X n } and Y = {Y1 , Y2 , . . . , Yn }, respectively. The 2-terminal coding system is shown in Fig. 4. The correlated sources are not co-located and cannot cooperate to directly exploit their correlation. Data sequences X n and Y n are separately encoded to ϕ0 (X n ) and ϕ1 (Y n ). The encoder functions are defined by ϕ0 : X n → C0 = {1, 2, . . . , C0 } , ϕ1 : Y n → C1 = {1, 2, . . . , C1 } . (3) The coded sequences are sent to the FC with the rate constraints 1 log Ci ≤ Ri + δ, i = 0, 1 n (4) where δ is an arbitrary prescribed positive number. The decoder observes (ϕ0 (X n ) , ϕ1 (Y n )), decodes all messages, and makes estimates of all sources. The decoder function is given by ψ = (ψ0 , ψ1 ) where its components are defined by ψ0 : C0 × C1 → X n , ψ1 : C0 × C1 → Y n . (5) Let d0 : X 2 → [0, ∞) and d1 : Y 2 → [0, ∞) be the squared distortion measures, i.e., ) = (Y − Y )2 . For the reconstructed signals  d0 (X,  X ) = (X −  X )2 and d1 (Y, Y Xn = n n n n n  = ψ1 (ϕ0 (X ) , ϕ1 (Y )), the average distortions 0 , 1 can ψ0 (ϕ0 (X ) , ϕ1 (Y )) and Y be defined by  n  1 0 = E d0 (X t ,  Xt ) , n t=1  n  1  1 = E d1 (Yt , Yt ) . n t=1 For given distortion levels D0 and D1 , a rate pair (R0 , R1 ) is admissible if for any δ > 0 and any n ≥ n 0 (δ) there exists a triple (ϕ0 , ϕ1 , ψ) satisfying (3)–(5) such that i ≤ Di + δ for i = 0, 1. The rate-distortion region R(D0 , D1 ) can be defined as R(D0 , D1 ) = {(R0 , R1 ) : (R0 , R1 ) is admissible} . (6) 2.1 One-helper Problem Formulation Assume that distortion level D1 is sufficiently large, i.e., the goal is only to reconstruct X , and the other source, Y , is used as a helper. In other words, the helper will not be reconstructed, and it is just used as auxiliary information to reconstruct the main source. This problem is called the 1-helper problem. The associated coding system is depicted in Fig. 7. The encoder functions (ϕ0 , ϕ1 ) and the decoder function ψ0 for the 1-helper scenario are as follows: 123 ϕ0 : X n → C0 = {1, 2, . . . , C0 } , ϕ1 : Y n → C1 = {1, 2, . . . , C1 } , (7) ψ0 : C0 × C1 → X n , (8) Successively Structured Gaussian Two-terminal Source Coding n n X X ϕ0 Y ϕ0 (X ) n n Y 493 ϕ1 ϕ1 (Y ) ψ0 X n Fig. 7 1-helper coding system where 1 log Ci ≤ Ri + ǫ, i = 0, 1 n (9) and ǫ is an arbitrary prescribed positive number. Here, the decoder makes an estimate of the main source X n as  X n rather than estimates of both sources. There is no distortion constraint for the helper and any available rate may be used for coding of the helper information. The distortion measure for the main source X is defined as d0 : X 2 → [0,∞) which is a MSE distortion measure. The average distortion can be defined by 0 = E n1 nt=1 d0 (X t ,  Xt ) where the reconstructed signal can be represented by  X n = ψ0 (ϕ0 (X n ) , ϕ1 (Y n )). A ratedistortion pair (R0 , D0 ) is admissible if for any ǫ > 0 and any n ≥ n 0 (ǫ) there exists a pair (ϕ0 , ϕ1 , ψ0 ) satisfying (7)–(9) such that 0 ≤ D0 + ǫ. The rate region does not depend on the distortion level D1 , and it can be represented by R0 (D0 ), i.e., R0 (D0 ) = {(R0 , R1 ) : (R0 , R1 ) ∈ R(D0 , D1 ) for some D1 > 0}. It is clear that R0 (D0 ) is an outer region of R(D0 , D1 ). In [6,17] it is shown that the whole rate-distortion region of the Gaussian 2-terminal coding system can be characterized by R(D0 , D1 ) = R∗0 (D0 ) ∩ R∗1 (D1 ) ∩ R01 (D0 , D1 ), (10) where   2 1 + σX 2 2 −2R1 1−ρ +ρ 2 = (R0 , R1 ) : R0 ≥ log , 2 D0   2 1 2 2 −2R0 + σY ∗ 1−ρ +ρ 2 R1 (D1 ) = (R0 , R1 ) : R1 ≥ log , 2 D1   2 2   1 + σ X σY βmax 2 R01 (D0 , D1 ) = (R0 , R1 ) : R0 + R1 ≥ log 1−ρ , 2 D0 D1 2 R∗0 (D0 ) βmax = 1 +  1+ D0 D1 4ρ 2 × 2 2 , (1 − ρ 2 )2 σ X σY (11) (12) (13) (14) and log+ x = max{log x, 0}. Our goal is to show that the whole rate-distortion region of the 2-terminal coding system can be achieved using the successive coding strategy. There are two steps to reach this 123 494 H. Behroozi, M. R. Soleymani goal: (i) Obtaining two curved parts of the R(D0 , D1 ), i.e., (11) and (12); (ii) Obtaining the straight-line segment of R(D0 , D1 ), i.e., the sum-rate limit of (13). Note that by applying the successive coding to a multiterminal problem, the encoding and successive decoding functions can be visualized as a succession of some Wyner-Ziv stages [12]. Thus, the encoding and decoding functions for any stage of the successive coding can be described as follows. Assume that at stage l (l > 1) , Yl is the encoder observation, and Z l−1 is the side information which is received from the decoder at stage l − 1. The encoding and decoding functions are as follows: ϕl : Yln → Cl = {1, 2, . . . , Cl } , (15) n ψl : Zl−1 × Cl → Yln . (16) The coded sequences are sent to the decoder with the rate Rl satisfying 1 log2 Cl ≤ Rl + γ , n (17) where γ is an arbitrary prescribed positive number. The decoder output will be given to the next decoder at stage l + 1 as the side information. 3 Two-terminal Source Coding 3.1 Two-curved Portions of the Rate-distortion Region The two-curved portions of the rate region for the 2-terminal coding system are in fact the rate-distortion regions of the 1-helper scenarios, where one source provides partial side information to the decoder to help reconstruction of the other source signal. We determine the rate-distortion performance of the successive coding strategy in the 1-helper problem. By applying the successive coding/decoding strategy in the 1-helper problem, the problem can be decomposed into two successive coding stages. The scenario is presented in Fig. 8. Each source encodes its message while previously decoded message that is available at the decoder acts as the decoder side information. At the FC, instead of joint decoding, messages from encoders are decoded sequentially. Fusion Center n X X ϕ0 Dec 0 Z1 Dec 1 n Y Y ϕ1 Fig. 8 1-helper coding system with the successive coding strategy 123 X n Successively Structured Gaussian Two-terminal Source Coding 495 Lemma 1 Successively structured Wyner-Ziv codes can achieve the rate-distortion function of the 1-helper coding system. In other words, for every D0 > 0, the achievable rate region of the successive coding strategy for the 1-helper scenario can be represented by   2 1 + σX 2 2 −2R1 R0 (D0 ) = (R0 , R1 ) : R0 ≥ log 1−ρ +ρ 2 . (18) 2 D0 Proof Consider the encoder of source Y . The encoder observation is the source signal and there is no side information available at decoder 1. Although we intend to encode the source Y , the decoder aims to obtain an estimate of the source X . We denote the estimate of the source X at decoder 1 as  X 1 . Based on the results of [8] for Wyner-Ziv coding of noisy observations (Page 38 of [8] Eq. 3.5), the minimum required rate for transmission an of Y in order to achieve average distortion D X 1 in estimating X at the decoder 1, i.e., E{ n1 nk=1 (X k −  X 1,k )2 } ≤ D X 1 would be as follows:   2 σ X − σ X2 |Y 1 (19) R1 (D X 1 ) = log , σ X2 |Y ≤ D X 1 ≤ σ X2 2 D X 1 − σ X2 |Y   where σ X2 |Y = σ X2 1 − ρ 2 . In the second step, we consider the encoder of source X . Let Z 1 denote the output signal generated by the helper at the decoder. Then, X can be encoded at the Wyner-Ziv rate  2  σ X |Z 1 1 R0 (D0 ) = log . (20) 2 D0 To obtain σ X2 |Z 1 , we use an innovation form, given in [40], to rewrite the relationship between X 1 = (X + V )α the estimate of the source X at decoder 1, i.e.,  X 1 , and the source X as  σ2 D X1 X where V ∼ N (0, σ 2 −D ) and α = (1 − X1 X DX 1 ). σ X2 In fact, this relationship is presented in [40, p. 370] as the forward channel realization of the Gaussian rate-distortion function. This channel realization that achieves the rate-distortion function is also called the optimum test channel. This relationship gives us the distortion of E (X −  X 1 )2 = D X 1 . The message Z 1  can be considered as Xα1 = X + V , i.e., Z 1 can be considered as the source X in an additive white Gaussian noise (AWGN), V . Thus, σ X2 |Z 1 = D X 1 . By substituting this result in (20) we obtain   DX1 1 R0 (D0 ) = log , 0 ≤ D0 ≤ D X 1 . (21) 2 D0 Rewriting Eq. 19 as D X 1 = σ X2 1 − ρ 2 + ρ 2 2−2R1 , (22) and then substituting (22) in (21) results in (18). This is the achievable rate-distortion function by the successive coding strategy. Comparing our result with the results of [17] and [6] in (11) shows that by applying the successive coding strategy, the rate-distortion function for the 1-helper coding system is achievable. This completes the proof. ⊔ ⊓   2 −2R 2 2 −2R 0 1 By rewriting (18) as D0 = σ X 2 1−ρ +ρ 2 it can be seen that each additional bit of the main source description reduces the average final distortion by a factor of 4; however, each additional bit of the helper description can reduce the average distortion by at most a factor of 4. This maximum reduction occurs when the correlation coefficient is one. Also, under a constant rate of the helper, D0 is a decreasing function of the correlation coefficient. 123 496 H. Behroozi, M. R. Soleymani Remark 1 We can also obtain the result of Lemma 1 when decoder 1 decodes the associated source signal Y instead of X . Then, this decoded signal is used as the side information for decoder 0. This could be shown as follows. The encoder of Y encodes and transmits its signal with the rate   σY2 1 R1 (D1 ) = log (23) , 0 ≤ D1 ≤ σY2 2 D1 where D1 is the average distortion in estimating Y at decoder 1. The encoder of X can encode the source X by considering its statistical knowledge about the decoder’s data, i.e., Z 1 = Y , with the rate ⎛ ⎞  2  σ X2 1 − ρ X2 Y σ X |Y 1 1 ⎠, (24) R0 (D0 ) = log = log ⎝ 2 D0 2 D0  D1 σY2 where Z 1 = Y = Y + V with V ∼ N (0, σ 2 −D ). Therefore, ρ X Y = ρ σY2 − D1 /σY . By Y 1 doing some manipulations, we will see that the rate-distortion tradeoff in (24) will be the same as (18). 3.1.1 Geometric Sphere-Packing argument In this subsection, we show that the rate distortion function of the successively structured 1-helper problem can be derived from geometric sphere-packing arguments [38,39]. Barron et al. [38] use these arguments to develop a geometric interpretation of the duality between the information-embedding problem and the Wyner-Ziv problem. In particular, the authors in [38] show that the rate-distortion function of Wyner-Ziv source coding and the channel capacity in the quadratic-Gaussian case can be derived geometrically. Similar arguments are also developed in [39] to illustrate the duality between channel coding and rate-distortion with side information. These arguments for the Wyner-Ziv source coding with noisy observation are presented in [8]. The basic geometric interpretations, presented in [8,38,39], for deriving the rate-distortion function of the Wyner-Ziv scheme can be described as follows. Given a side information Y at the decoder, a minimum mean-squared error (MMSE) estimate of the source is E[X |Y ] and the associated estimation error is σ X2 |Y . Before exploiting the message from the encoder, thedecoder knows that the source vector X lies within an uncertainty sphere of radius ru = nσ X2 |Y centered at its source estimate based on the side information Y only, i.e., E[X |Y ]. This sphere is depicted by a large dottedcircle in Fig. 9. Within this sphere, we pack spherical quantization regions of radius rq = n D − σ X2 |Y . The encoder maps X to the label of the quantization region in which it falls. The decoder picks out the quantization vector from the uncertainty sphere by referring to the label of the quantization region. We apply these geometric interpretations into the successively structured 1-helper problem. Lemma 2 The rate-distortion function of the successively structured 1-helper problem can be derived using geometric sphere-packing arguments. Proof Consider the successively structured 1-helper problem depicted in Fig. 8. At the first step, we obtain the transmission rate of the side information Y . Before receiving any information from the encoder, the decoder 1 only knows the transmitted vector lies in an uncertainty 123 Successively Structured Gaussian Two-terminal Source Coding 497 ru X rq Fig. 9 Sphere-packing argument to determine the rate-distortion function. The source estimate lies within the dotted circle with the radius of ru . The solid small circles with the radius rq correspond to quantization regions at the encoder    sphere of radius ru 1 = n σ X2 + ε11 which is centered at the source estimate E [X ]. The encoder of the helper source maps Y to the label of the quantization region in which it lies. Since our final goal is to reconstruct X at the decoder, we must consider how much the encoded message once decoded will help to reconstruct X . Therefore, the radius of each spherical  quantization region is equal to rq1 = σ2 n σ 2 −σX 2 X X |Y D X 1 − σ X2 |Y − ε22 . If no two quantiza- tion regions share the same label in each large sphere of radius ru 1 , the quantization region in which Y is located can be determined without error. The scenario is depicted in Fig. 9. We determine the minimum number of quantization spheres required to cover the uncertainty   n sphere. This number is lower bounded by the ratio of volumes of spheres: M1 ≥ k(n) ru 1  n k(n) rq1 where k(n) is a coefficient that is a function of the dimension. Thus, the lower bound on the rate-distortion function R1 (D X 1 ) can be obtained as ⎞ ⎛  2  2 σ X − σ X2 |Y σ X + ε11 1 1 ⎟ 1 ⎜ R1 = log2 M1 ≥ log ⎝ . ⎠ > log σ X2 n 2 2 D X 1 − σ X2 |Y 2 −ε D − σ X1 22 X |Y σ 2 −σ 2 X X |Y (25) In the next step, based solely on the available side information at the decoder 0, Z 1 , the decoder can have a MMSE equal to σ X2 |Z 1 = D X 1 . Therefore, the transmitted vector X falls 123 498 in an uncertainty sphere of radius ru = H. Behroozi, M. R. Soleymani  √ n σ X2 |Z 1 + ε1 = n (D X 1 + ε1 ) which is cen- tered at the source estimate E [X |Z 1 ]. This sphere is shown by the dotted circle in Fig. 9. The encoder 0 maps X to the label of the quantization region in which it lies, where the radius √ of each quantization sphere equals to rq = n (D0 − ε2 ). Thus, the minimum number of quantization spheres of radius rq required to cover the larger sphere of uncertainty is lower bounded by √ n k(n) n (D X 1 + ε1 ) M≥ (26) √ n . k(n) n (D0 − ε2 ) As a result, the lower bound on the rate-distortion function of the 1-helper problem can be derived as     D X 1 + ε1 DX1 1 1 1 R0 = log2 M ≥ log > log . (27) n 2 D0 − ε2 2 D0 By substituting D X 1 from (25) into (27), the same result as (18) will be obtained. ⊔ ⊓ To apply the successive coding strategy into the 2-terminal coding system, there are two possible orderings for the coding/decoding process: (a) the message from source Y is decoded first, and (b) the message from source X is decoded first. Ordering (a) determines the ratedistortion function for the 1-helper problem when Y is the helper and therefore the distortion for reconstructing X is minimized. Ordering (b) determines the rate-distortion function for the 1-helper problem when X is the helper and therefore the distortion for reconstructing Y is minimized. These two rate-distortion functions determine two curved portions of the rate region R(D0 , D1 ) for the 2-terminal source coding problem. 3.2 Sum-rate Limit for the 2-terminal Rate Region Now we want to obtain the last part of the R(D0 , D1 ) which is defined by the boundary of the sum-rate. Using (23) and (24), ordering (a) gives the minimum sum-rate of   2 ρ2  σ X2 σY2  σ 1 R0 + R1 = log 1 − ρ2 + X . (28) 2 D0 D1 D0 By symmetry, ordering (b) gives the minimum sum-rate of    σY2 ρ 2 σ X2 σY2  1 2 1−ρ + R0 + R1 = log . 2 D0 D1 D1 (29) By allowing time-sharing between ordering (a) and (b), the achievable sum-rate can be expressed as     α   σ X2 ρ 2 σ X2 1 − ρ 2 + D0 ρ 2 σY2 σ X2 σY2  1 1 2   1−ρ + , (30) + log R0 + R1 = log 2 D0 D1 D0 2 σY2 1 − ρ 2 + D1 ρ 2 σ X2 where 0 ≤ α ≤ 1. Since the sum-rate of (30) is achievable, it is an upper bound for the sum-rate part in the rate region of 2-terminal source coding problem. By subtracting (30) from (13) and doing some manipulations, we can show that the sum-rate distortion function of the 2-terminal source coding is achievable if 123 Successively Structured Gaussian Two-terminal Source Coding max ρ2 = min   D0 D1 , σ X2 σY2 D0 D1 , σ X2 σY2   499 −1 (31) . −1 In this case, the successive coding strategy degrades to the no-helper problem where only the source with the minimum average distortion Di (i = 0, 1) should be encoded. In our analysis, so far, we have assumed that each decoder desires to reconstruct the corresponding source within the given fidelity, Di . But we can develop connections between the data fusion and the source coding with side information. In this scenario, the encoder quantizes and encodes the source to a degree of fidelity less than Di to reduce the transmission rate. But this reduction does not affect the quality of the reconstructed signals at the FC. In fact, the remaining correlation among the decoded signals enables the FC to reconstruct the sources into desired degrees of fidelity. By using a linear estimator after the source decoder, both source signals can be reproduced at the desired degrees of fidelity, Di ’s. The joint design of the source encoding and the data fusion will yield substantial performance gain over decoupled designs. Based on this scenario, we demonstrate that the successive coding strategy can achieve all the rate-distortion tuples (R0 , R1 , D0 , D1 ) belonging to the inner bound of Berger and Tung [5,23]. In fact, the whole rate-distortion region of the 2-terminal source coding problem can be characterized by applying the successive coding strategy. Theorem 1 Successive coding strategy can achieve the sum-rate distortion function of the 2-terminal source coding problem. Proof Our proof is similar to the proof of source-splitting method presented in [41,42]. Consider the ordering (a) for the 2-terminal source coding, depicted in Fig. 10. The source encoder ϕ1 , quantizes Y n and then compresses the quantized signal. The output message W1 is transmitted at rate R1 . At the decoder side, W1 is decompressed and used to reconstruct Y n as Y n . By exploring the remained correlation between Y n and X n , the encoder ϕ0 compresses the quantized version of X n at rate R0 . Using Y n as the side information, the decoder ψ0 decodes the received signal to X n . Given X n and Y n , the linear estimator reproduces X n and Y n using linear combination of the inputs:  X n = α0 Y n + β0 X n ,  Y n = α1 Y n + β1 X n . X X n ϕ0 n Y Y ϕ1 R0 R1 ψ0 n ~ X Linear ~n ψ1 (32) Y X Estimator (FC) Y n n Fig. 10 Block diagram of the successive coding strategy (with decoding ordering (a)) in the 2-terminal source coding problem. The linear estimator fuses both received signals to produce the estimates of both sources 123 500 H. Behroozi, M. R. Soleymani Define the quantization errors as E 0n = X n − X n and E 1n = Y n − Y n . We first derive the estimator coefficients. The average quantization distortions 0 , 1 can be defined by  n   n   1 1 2 2 (X t −  Xt ) = E E 0,t , 0 = E n n t=1 t=1  n   n  1 1 2 2 (Yt − Yt ) = E E 1,t . 1 = E n n t=1 t=1 Due to the orthogonal properties for optimal estimations, applying the projection theorem results in:   E (X −  X ) X = 0, E (X −  X )Y = 0,    )Y = 0. E (Y − Y ) X = 0, E (Y − Y (33) As a result, bydoing some manipulation, the estimator coefficients that minimize the average  )2 can be obtained as distortions E (X −  X )2 and E (Y − Y   2 σY + 1 σ X2 − ρ 2 σ X2 σY2 ρσ X σY 0 α0 = , β0 = ∗ ∗  2  2 2 2 2 σ + 0 σY − ρ σ X σY ρσ X σY 1 α1 = X , β1 = , (34) ∗ ∗    where ∗ = σ X2 + 0 σY2 + 1 − ρ 2 σ X2 σY2 . Based on the rate-distortion theory and Wyner-Ziv results, the transmission rates of this scheme can be expressed as n R0 ≥ I (X n ; X n ) − I ( X n ; Y n ), n R1 ≥ I (Y n ; Y n ). (35) (36)   2 For jointly Gaussian random variables X and Y , I (X ; Y ) = − 21 log 1 − ρ . We just need to obtain the correlation coefficients ρ X X , ρ X Y , and ρY Y . We can show ρ X2 X = σ X2 σ X2 + 0 , ρY2 Y = σY2 ρ 2 σ X2 σY2 2   . , ρ = XY σY2 + 1 σ X2 + 0 σY2 + 1 Thus, (35) and (36) can be computed as     σ X2 + 0 σY2 + 1 − ρ 2 σ X2 σY2 1   , R0 ≥ log 2 0 σY2 + 1   σY2 + 1 1 R1 ≥ log . 2 1 (37) (38) (39) The overall average distortions can be expressed as   X )2 = E (X − α0 Y − β0 X )2 D0∗ = E (X −      = σ X2 + α02 σY2 + 1 + β02 σ X2 + 0 − 2α0 ρσ X σY − 2β0 σ X2 + 2α0 β0 ρσ X σY , (40)   2 ∗ 2  D1 = E (Y − Y ) = E (Y − α1 Y − β1 X )     = σY2 + α12 σY2 + 1 + β12 σ X2 + 0 − 2α1 σY2 − 2β1 ρσ X σY + 2α1 β1 ρσ X σY . (41) 123 Successively Structured Gaussian Two-terminal Source Coding By substituting (34) in (40) and (41), the average distortions can be simplified as  2   σY 1 − ρ 2 + 1 σ X2 0 ∗ D0 = , ∗ D1∗    σ X2 1 − ρ 2 + 0 σY2 1 . = ∗  501 (42) (43) As we mentioned in the last paragraph in Sect. 3.1.1, there are two possible message orderings in the successively structured 2-terminal coding system: (a) the message from source Y is designed to be decoded first, and (b) the message from source X is designed to be decoded first. So far, we have analyzed the message ordering (a). We show that one of the corner points on the sum-rate bound of (13) is achievable with the above-mentioned coding scheme, depicted in Fig. 10. We prove this result by construction. Let   2D0 σ X2 σY2 1 − ρ 2   0 = , (44) βmax σ X2 σY2 1 − ρ 2 − 2D0 σY2 1 =   2D1 σ X2 σY2 1 − ρ 2   . βmax σ X2 σY2 1 − ρ 2 − 2D1 σ X2 Substituting (44) and (45) in (38) and (39) reveals that  2 βmax σ X2 σY2 1 − ρ 2    , R0 A = D0 βmax σY2 1 − ρ 2 − 2D1 ρ 2   βmax σY2 1 − ρ 2 − 2D1 ρ 2   . R1A = 2D1 1 − ρ 2 (45) (46) (47) which satisfy the sum-rate distortion function of (13). The achievable average distortions in (42) and (43) can be calculated as D0∗ = D0 and D1∗ = D1 . Therefore, the sum-rate distortion function of (13) is achievable. In fact, the successively structured 2-terminal coding scheme with coding/decoding of ordering (a) can achieve one corner point of the sum-rate distortion function. Analyzing decoding ordering (b) gives the same rate-distortion points as those of decoding ordering (a) in (46) and (47) with the subscripts interchanged, i.e., subscripts (0, 1, X, Y ) are replaced by (1, 0, Y, X ), respectively. As a result,   βmax σ X2 1 − ρ 2 − 2D0 ρ 2   R0B = , (48) 2D0 1 − ρ 2  2 βmax σ X2 σY2 1 − ρ 2    . (49) R1B = D1 βmax σ X2 1 − ρ 2 − 2D0 ρ 2 Therefore, the optimal rate allocation scheme that achieve any distortion pair (D0 , D1 ) using minimum sum-rate R of (13) corresponds to the points on line AB. Each corner point of this sum-rate line is achievable using one of two possible decoding orderings in the successively structured 2-terminal coding system. Allowing time-sharing between both coding schemes, i.e., the 2-terminal coding system with decoding ordering (a) and the 2-terminal coding system with decoding ordering (b), achieves all the intermediate points of line AB. ⊓ ⊔ 123 502 H. Behroozi, M. R. Soleymani 3.3 Comparison with Sequential Coding of Correlated Sources In this part, we make a comparison between the minimum sum-rate of the sequential coding [37] and the sum-rate of the successive coding strategy for coding of correlated Gaussian sources. In fact, the scenario of successive coding in the 2-terminal source coding problem is similar to the scenario of sequential coding for two correlated sources, illustrated in Sect. 1.2.3. The sequential coding scheme is depicted in Fig. 6. In this coding system, the encoding is performed sequentially, i.e., the encoding of Y depends on the source X but not vice versa. In [37], it is shown that the minimum sum-rate for the sequential coding of correlated Gaussian sources is given by     σ X2 σY2 1 D0 2 R0 + R1 = log . (50) 1−ρ 1− 2 2 D0 D1 σX Now, consider the ordering (b) of the successive coding strategy. The minimum sum-rate of the successive coding strategy for this scheme is given in (29). By rewriting (29), we observe that this is the same as the minimum sum-rate of sequential coding in (50). Therefore, there is no sum-rate loss with the successive coding compared with the sequential coding of correlated Gaussian sources. This means that the availability of side information at the encoder does not improve the sum-rate distortion function. This observation is of special interest in some applications such as video coding which is described as follows. 3.3.1 Video Coding A video source can be considered as a sequence of video frames [37,43]. Each frame of the video may be viewed as a source and a sequence of frames corresponds to a sequence of correlated sources. The video coding scenario is illustrated in Fig. 11. Video frame j, consists of a two-dimensional array of pixels, correspond to the source X j . Specifically, X j = {X j (k)}nk=1 for j = 1, 2, . . . , T represent T video frames where k denotes the spatial location of a pixel with respect to a particular spatial scan order (e.g., zig-zag), and X j (k) represents luminance intensity of the pixel at spatial location k in frame number j. This statistical structure implies that the sources are spatially independent but temporally dependent. With this perspective the video coding can be modeled as sequential encoding of correlated sources. space time Frame 1 Frame 2 Frame 3 X1 X2 X3 Fig. 11 Video coding scenario. A video source can be considered as a sequence of video frames which are spatially independent but temporally dependent. Each frame, consisting of a two-dimensional array of pixels, can be modeled as a source and a sequence of frames corresponds to a sequence of correlated sources 123 Successively Structured Gaussian Two-terminal Source Coding 503 By determining the minimum total rate R(D0 , D1 ) = R0 + R1 required to achieve distortion levels D0 and D1 in the successively structured 2-terminal source coding, we showed that there is no sum-rate loss with the successive coding compared with the sequential coding of correlated Gaussian sources. Consequently, efficient compression can be achieved by exploiting source statistics at the decoder side only. Furthermore, in most practical video codecs there are storage constraints. In particular, the encoder for encoding a frame can retain a copy of the previous frame but not earlier frames [37]. This suggests that we should consider the successive coding as a promising technique to achieve the minimum sum-rate with no storage constraint at the encoder. 4 M-terminal Source Coding In the multiterminal source coding, the FC decodes messages of all sources. The multitern minal coding system  n  with m-terminal is shown in Fig. 12. Data sequences Yi are separately encoded to ϕi Yi where the encoder functions are defined by ϕi : Yin → Ci = {1, 2, . . . , Ci } , i = 1, 2, . . . , m. (51) The coded sequences are sent to the FC with the rate constraints 1 log Ci ≤ Ri + δ, i = 1, 2, . . . , m (52) n where  δ is an arbitrary   prescribed positive number. The decoder observes the m-tuple ϕ1 Y1n , . . . , ϕm Ymn , decodes all the messages, and makes estimates of all the sources. The decoder function is given by ψ = (ψ1 , ψ2 , . . . , ψm ) where its components are defined by ψi : C1 × C2 × · · · × Cm → Yin , i = 1, 2, . . . , m. (53) i )2 . Let di : Yi2 → [0, ∞) be the squared distortion measures, i.e., di (Yi , Yi ) = (Yi − Y   n  n  n  For the reconstructed signals Yi = ψi ϕ1 Y1 , . . . , ϕm Ym , the average distortions 1 , 2 , . . . , m can be defined by   n 1 it ) . di (Yit , Y (54) i = E n t=1 n n Y1 Y1 ϕ1 ϕ1 (Y1 ) n n n Y2 Y2 ϕ2 .. . ... n Ym Ym ϕm ϕ2 (Y2 ) ψ=(ψ1 ,ψ2 ,...,ψm ) .. . Y1 .. .n Ym n ϕm(Ym) Fig. 12 m-terminal source coding scheme. Yi ’s for i = 1, . . . , m are m correlated sources that are separately encoded. The joint decoder aims to obtain estimates of all the sources 123 504 H. Behroozi, M. R. Soleymani For given distortion levels (D1 , D2 , . . . , Dm ), an m-tuple set of rates (R1 , R2 , . . . , Rm ) is admissible if for any δ > 0 and any n ≥ n 0 (δ) there exists a (m+1)−tuple (ϕ1 , ϕ2 , . . . , ϕm , ψ) satisfying (51)–(53) such that i ≤ Di + δ for i = 1, 2, . . . , m. The rate-distortion region R (D1 , D2 , . . . , Dm ) can be defined as all the m-tuple sets of rates (R1 , R2 , . . . , Rm ) that are admissible. Similar to the previous section, we first consider the special case of the m-terminal coding system, where the goal is to estimate one of these sources while other sources provide partial side information to the decoder to help reconstruction of the main source. Based on the result of the following section, we can obtain an inner region for the m-terminal coding system. 4.1 Achievable Rate Region of a Special case: m-helper Problem The m-helper system and the system based on the successive coding strategy are shown in Figs. 13 and 14, respectively. By generalizing the results of Sect. 3 and considering the fact that each decoder decodes its corresponding source, the rates of encoders can be computed as follows: 2 σY 1 log m 2 Dm 2 σ 1 Y |Z = log m−1 m 2 Dm−1 Rm = Rm−1 2 Rm−2 = .. . X ϕ0 Y1 ϕ1 .. . .. . n Ym Ym ϕ0 (X ) n n Y1 (55) n n X σY |Z ,Z 1 log m−2 m m−1 2 Dm−2 .. . ϕm ϕ1 (Y1 ) ψ0 X n n ϕm(Ym) Fig. 13 m-helper coding system. X and Yi ’s for i = 1, . . . , m are m + 1 correlated sources that are separately encoded. The joint decoder wants to obtain an estimate of the main source X while other m sources, called the helpers, play the role of partial side information to help the decoder to reconstruct the transmitted sequence of the main source 123 Successively Structured Gaussian Two-terminal Source Coding 505 Fusion Center n X X ϕ0 R0 ϕ1 R1 n Y1 Y1 Y2 ϕ2 n Ym n ~n ~ ~n ~n Ym ,Ym−1 ,..., Y2 , Y1 R2 .. . .. . Ym n Dec 1 n ~n ~ ~n Ym ,Ym−1 ,..., Y2 n Y2 X Dec 0 ϕm Dec 2 .. .~ n Ym Rm Dec m Fig. 14 Successively structured m-helper source coding scheme. By applying the successive coding/decoding strategy in the m-helper problem, the problem is decomposed into m successive coding stages 2 σY |Z ,Z ,...,Z 2 1 log 1 m m−1 2 D1 2 σ 1 X |Z m ,Z m−1 ,...,Z 2 ,Z 1 R0 = log . 2 D0   Di i where Z i = Yi = Yi + Vi , Vi ∼ N (0, D ), α = 1 − and Di is the average distortion i αi σ2 R1 = Yi in estimating source Yi at decoder i for i = 1, . . . , m. To obtain the final achievable ratedistortion region R0 (D0 ), we should obtain the conditional variances in (55) in terms of the source variances and the correlation coefficients between sources. If X i is a zero mean Gaussian random variable, X i ∼ N (0, σ X2 ), and X i ’s are jointly Gaussian, then the conditional pdf, p X n+1 |X 1 ,X 2 ,...,X n (xn+1 |x1 , x2 , . . . , xn ) can be obtained as follows [44]: 1 p X n+1 |X 1 ,X 2 ,...,X n (xn+1 |x1 , x2 , . . . , xn ) = √ 2π   (xn+1 − a1 x1 − a2 x2 − · · · − an xn )2 exp − , 2 (56) where  = σ X2 n+1 |X 1 ,X 2 ,...,X n  = Rn+1,n+1 − a1 R1,n+1 − a2 R2,n+1 − · · · − an Rn,n+1 (57) and Ri, j = E X i X j = ρ X i X j σ X i σ X j which are called the correlation functions. Coefficients a1 , a2 , . . . , an can be computed from the following equation:  E (X n+1 − a1 X 1 − a2 X 2 − · · · − an X n ) X i = 0, (58) for i = 1, 2, . . . , n. This equation can be represented as the series of equations as follows: 123 506 H. Behroozi, M. R. Soleymani ⎧ a1 R11 + a2 R21 + · · · + an Rn,1 = Rn+1,1 ⎪ ⎪ ⎪ ⎨ a1 R12 + a2 R22 + · · · + an Rn,2 = Rn+1,2 .. .. ⎪ ⎪ . . ⎪ ⎩ a1 R1,n + a2 R2,n + · · · + an Rn,n = Rn+1,n (59) To solve (59), consider it as the matrix equation AR X X = R X n+1 X where A = [a1 a2 . . . an ], X = [X 1 . . . X n ], R X X = E Xt X , and R X n+1 X = [Rn+1,1 Rn+1,2 . . . Rn+1,n ]. Therefore, the coefficient vector of A can be obtained by A = R X n+1 X R−1 XX. (60) As a result, using (57) and (60), the conditional variances required in (55) can be obtained versus the correlation functions. Since the final goal is to obtain R0 (D0 ) in terms of source variances and correlation coefficients between sources, we just need to derive the relation between   correlation coefficients ρYi Z j , ρ Z i Z j , ρ X Z i and correlation coefficients (ρ X Yi , ρYi Y j ). These coefficients can be expressed as follows: ρYi Y j  2 ρYi Z j = σY j − D j , σY j ρ X Yi  2 ρ X Zi = σYi − Di , (61) σYi  ρYi Y j  2 σYi − Di σY2j − D j . ρZi Z j = σYi σY j 4.2 Special Case: Two-helper Problem From Eq. 55, we know that  2  σY2 1 R2 (D2 ) = log , 2 D2  2  σY1 |Z 2 1 , R1 (D1 ) = log 2 D1  2  σ X |Z 1 ,Z 2 1 R0 (D0 ) = log , 2 D0 (62) where   σY21 |Z 2 = σY21 1 − ρY21 Z 2 , σ X2 |Z 1 ,Z 2 = σ X2 1 − ρ12 1 − ρ Z2 1 Z 2 (63) , ρ12 = ρ X2 Z 1 + ρ X2 Z 2 + ρ Z2 1 Z 2 − 2ρ X Z 1 ρ X Z 2 ρ Z 1 Z 2 . Combining (63) with (61) results in ⎡   ⎤ σ X2 1 − ρ12 1 ⎦, R0 (D0 ) = log ⎣ 2 D 1 − ρ2 0 123 Z1 Z2 (64) Successively Structured Gaussian Two-terminal Source Coding 507 where ρ12 and ρ Z2 1 Z 2 in terms of the correlation coefficients among X, Y1 and Y2 are as follows: ρ12 = ρ X2 Y1 + ρ X2 Y2 + ρY21 Y2 − 2ρ X Y1 ρ X Y2 ρY1 Y2   + 2ρ X Y1 ρ X Y2 ρY1 Y2 − ρ X2 Y1 − ρY21 Y2 × 1 − ρY21 Y2 + ρY21 Y2 2−2R2 2−2R1   + 2ρ X Y1 ρ X Y2 ρY1 Y2 − ρ X2 Y2 − ρY21 Y2 2−2R2   + ρY21 Y2 − 2ρ X Y1 ρ X Y2 ρY1 Y2 × 1 − ρY21 Y2 + ρY21 Y2 2−2R2 2−2R1 2−2R2 , (65) and ρ Z2 1 Z 2 = ρY21 Y2 1 − 2−2R2 × 1 − 2−2R1 1 − ρY21 Y2 + ρY21 Y2 2−2R2 . (66) Therefore, (64) defines an achievable rate-distortion region for the 2-helper problem based on the successive coding strategy. If R1 → ∞ and R2 → ∞, i.e., the case of full resolution side information, (64) can be simplified to ⎤ ⎡ σ X2 1 − ρ X2 Y1 − ρ X2 Y2 − ρY21 Y2 + 2ρ X Y1 ρ X Y2 ρY1 Y2 1 ⎦, (67) R0 (D0 ) = log ⎣ 2 D 1 − ρ2 0 Y1 Y2 which is equal to the conditional rate-distortion function, R X |Y1 ,Y2 , when both encoder and decoder have access to Y1 and Y2 . Therefore, there is no rate loss with the successive coding when compared to lossy coding of X with the side information (Y1 and Y2 ) available at both the encoder and the decoder. Example 1 Consider the two-helper problem with R2 = 0. Since the rate of the encoder 2 is zero, there is no help from Y2 . Therefore, we expect to obtain the rate-distortion for the 1-helper problem. By substituting R2 = 0 in (64), we obtain   σ X2 1 2 2 −2R1 1 − ρ X Y1 + ρ X Y1 2 R0 (D0 ) = log , (68) 2 D0 which is the rate-distortion function of the 1-helper problem. 4.3 An Inner Region for m-terminal Coding System Assume i = (πi1 , πi2 , . . . , πim ) is a permutation of the set Im = {1, 2, . . . , m}. Using the successive coding strategy, there are m! possible orderings for the coding/decoding process in the m-terminal source coding scheme. For a given permutation , the following rate region is achievable:      D(1) , D(2) , . . . , D(m) , R (D(1) , D(2) , . . . , D(m) ) = R∗(m) D(m) ∩ R (69) where R∗( j)   D( j) =   R(1) , R(2) , . . . , R(m) : R( j)  2  σY( j) |Z ( j−1) ,...,Z (1) 1 ≥ log , 2 D( j) (70) 123 508 H. Behroozi, M. R. Soleymani    D(1) , D(2) , . . . , D(m) = R m j=1 R( j) ≥ 1 2 log  )m and Z ( j) = Y( j) + V( j) , V( j) ∼ N (0,   R(1) , R(2) , . . . , R(m) : 2 j=1 σY( j) |Z ( j−1) ,...,Z (1) )m j=1 D( j) D( j) α( j) ), α( j)  = 1− * (71) , D( j) σY2 ( j)  . The inner region for the rate region of the multiterminal source coding can be represented as , , ++ m! min (D1 , D2 , . . . , Dm ) , (72) Rin (D1 , D2 , . . . , Dm ) = ∩i=1 R∗i (m) (Di (m) ) ∩ R where 1≤i≤m!    i Di (1) , Di (2) , . . . , Di (m) . min (D1 , D2 , . . . , Dm ) = min R R  (73) R∗i (m) Di (m) is in fact the rate-distortion tradeoff of the (m − 1)-helper coding scheme using the ith decoding ordering (out of m! possible orderings). 5 Conclusion In this paper, we showed that the successive coding strategy can achieve the rate-distortion function of the 1-helper problem as well as the whole rate-distortion region of the 2-terminal source coding system. Compared with the sequential coding of correlated Gaussian sources, there is no loss of sum-rate for the successive coding where the side information is not available at the encoder. This result is of special interest in some applications such as video coding, where we have processing and storage constraints at the encoder. Finally, we derived an inner bound for the rate region of the m-terminal source coding system. References 1. Gelfand, S. I., & Pinsker, M. S. (1979). Coding of sources on the basis of observations with incomplete information. Problems of Information Transmission, 15(2), 115–125. 2. Yamamoto, H., & Itoh, K. (1980). Source coding theory for multiterminal communication systems with a remote source. Transactions of IECE Japan, E63(10), 700–706. 3. Flynn, T. J., & Gray, R. M. (1987). Encoding of correlated observations. IEEE Transactions on Information Theory, 33(6), 773–787. 4. Berger, T., Zhang, Z., & Viswanathan, H. (1996). The CEO problem. IEEE Transactions on Information Theory, 42(3), 887–902. 5. Berger, T. (1978). Multiterminal source coding. In G. Longo (Ed.). The Information Theory Approach to Communications (CISM Courses and Lectures, no. 229) (pp. 171–231). Vienna, Austria: Springer-Verlag. 6. Wagner, A. B., Tavildar, S., & Viswanath, P. (2008). The rate region of the Quadratic Gaussian two-terminal source-coding problem. IEEE Transactions on Information Theory, 54(5), 1938–1961. 7. Servetto, S. D. The sum-rate of multiterminal source coding with two encoders. IEEE Transactions on Information Theory, submitted for publication, Apr. 2006 (original title: The Region of Achievable Rates for Multiterminal Source Coding, Part I). 8. Draper, S. C. (2002). Successive structuring of source coding algorithms for data fusion, buffering, and distribution in networks, Ph.D. dissertation, Cambridge, MA: Massachusetts Institute of Technology (MIT), Jun. 2002. 9. Natrio, L., Brites, C., Ascenso, J., & Pereira, F. (2005). Extrapolating side information for low-delay pixel-domain distributed video coding. Int. Workshop on Very Low Bitrate Video Coding, Sardinia, Italy, Sept. 2005. 10. Shamai, S., Verdu, S., & Zamir, R. (1998). Systematic lossy source/channel coding. IEEE Transactions on Information Theory, 44(2), 564–579. 123 Successively Structured Gaussian Two-terminal Source Coding 509 11. Girod, B., Aaron, A., Rane, S., & Rebollo-Monedero, D. (2005). Distributed video coding. Proceedings of the IEEE, 93(1), 71–83, Jan. 2005. Invited paper, Special Issue on Video Coding and Delivery. 12. Wyner, A. D., & Ziv, J. (1976). The rate-distortion function for source coding with side information at the decoder. IEEE Transactions on Information Theory, 22(1), 1–10. 13. Wyner, A. D. (1978). The rate-distortion function for source coding with side information at the decoder II: General sources. Information and Control, 38(1), 60–80. 14. Xiong, Z., Stankovic, V., Cheng, S., Liveris, A., & Sun, Y. (2004). Source-channel coding for algebraic multiterminal binning. In Proceedings on IEEE Information Theory Workshop (ITW), San Antonio, Texas, Oct. 2004. 15. Costa, M. (1983). Writing on Dirty Paper. IEEE Transactions on Information Theory, 29(3), 439–441. 16. Eswaran, K. (2006). Remote source coding and AWGN CEO problems. Master’s thesis, EECS Department, University of California, Berkeley, Jan. 2006. 17. Oohama, Y. (1997). Gaussian multiterminal source coding. IEEE Transactions on Information Theory, 43(6), 1912–1923. 18. Pandya, A., Kansal, A., Pottie, G. J., & Srivastava, M. B. (2004). Lossy source coding of multiple Gaussian sources: m-helper problem. In Proceedings on IEEE Information Theory Workshop (ITW), San Antonio, Texas, Oct. 2004. 19. Oohama, Y. (1999). Multiterminal source coding for correlated memoryless Gaussian sources with several side information at the decoder. In Proceedings on IEEE Information Theory Workshop, p. 100, Kruger National Park, South Africa, Jun. 1999. 20. Oohama, Y. (2005). Rate-distortion theory for Gaussian multiterminal source coding systems with several side informations at the decoder. IEEE Transactions on Information Theory, 51(7), 2577–2593. 21. Oohama, Y. (2006). Gaussian multiterminal source coding with several side informations at the decoder. In Proceedings on IEEE Int. Symp. Information Theory (ISIT), Seattle, Washington, Jul. 2006. 22. Berger, T., & Yeung, R. (1989). Multiterminal source encoding with one distortion criterion. IEEE Transactions on Information Theory, 35(2), 228–236. 23. Tung, S. Y. (1978). Multiterminal source coding. Ph.D. Thesis, School of Electrical Engineering, Cornell University, Ithaca, NY, May 1978. 24. Oohama, Y. (2006). Separate source coding of correlated Gaussian remote sources. In Proceedings on 2006 Information Theory and Applications Workshop, San Diego, CA, USA, Jan. 2006. 25. Koshelev, V. N. (1980). Hierarchical coding of discrete sources. Problemy Peredachi Informatsti, 16(3), 31–49. 26. Equitz, W. H. R., & Cover, T. M. (1991). Successive refinement of information. IEEE Transactions on Information Theory, 37(2), 269–275. 27. Rimoldi, B. (1994). Successive refinement of information: Characterization of achievable rates. IEEE Transactions on Information Theory, 40(1), 253–259. 28. Steinberg, Y., & Merhav, N. (2004). On successive refinement for the Wyner-Ziv problem. IEEE Transactions on Information Theory, 50(8), 1636–1654. 29. Cheng, S., & Xiong, Z. (2005). Successive refinement for the Wyner-Ziv problem and layered code design. IEEE Transactions on Signal Processing, 53(8), 3269–3281. 30. Tian, C., & Diggavi, S. (2007). On multistage successive refinement for Wyner-Ziv source coding with degraded side information. IEEE Transactions on Information Theory, 53(8), 2946–2960. 31. Viswanathan, H., & Berger, T. (1997). The Quadratic Gaussian CEO problem. IEEE Transactions on Information Theory, 43(5), 1549–1559. 32. Chen, J., Zhang, X., Berger, T., & Wicker, S. B. (2004). An upper bound on the sum-rate distortion function and its corresponding rate allocation schemes for the CEO problem. IEEE Journal on Selected Areas in Communications, 22(6), 977–987. 33. Prabhakaran, V., Tse, D., & Ramchandran, K. (2004). Rate region of the Quadratic Gaussian CEO problem. In Proc. IEEE Int. Symp. Information Theory (ISIT), June 2004, p. 119. 34. Chen, J., & Berger, T. (2008). Successive Wyner-Ziv coding scheme and its application to the Quadratic Gaussian CEO problem. IEEE Transactions on Information Theory, 54(4), 1586–1603. 35. Behroozi, H., & Soleymani, M. R. (2007). Distortion sum-rate performance of successive coding strategy in quadratic Gaussian CEO problem. IEEE Transactions on Wireless Communications, 6(12), 4361–4365. 36. Behroozi, H., & Soleymani, M. R. (2006). Successively structured Gaussian CEO problem. In Proceedings on IEEE Global Telecommunications Conference (GLOBECOM 2006), San Francisco, CA, Nov. 2006. 37. Viswanathan, H., & Berger, T. (2000). Sequential coding of correlated sources. IEEE Transactions on Information Theory, 46(1), 236–246. 123 510 H. Behroozi, M. R. Soleymani 38. Barron, R. J., Chen, B., & Wornell, G. W. (2003). The duality between information embedding and source coding with side information and some applications. IEEE Transactions on Information Theory, 49(5), 1159–1180. 39. Pradhan, S. S., Chou, J., & Ramchandran, K. (2003). Duality between source coding and channel coding and its extension to the side information case. IEEE Transactions on Information Theory, 49(5), 1181–1203. 40. Cover, T. M., & Thomas, J. A. (1991). Elements of information theory. New York: Wiley. 41. Yang, Y., Stankovic, V., Xiong, Z., & Zhao, W. (2005). On multiterminal source code design. In Proceedings on IEEE Data Compression Conference (DCC), Snowbird, Utah, Mar. 2005. 42. Yang, Y., Stankovic, V., Xiong, Z., & Zhao, W. (2008). On multiterminal source code design. IEEE Transactions on Information Theory, 54(5), 2278–2302. 43. Ma, N., Wang, Y., & Ishwar, P. (2007). Delayed sequential coding of correlated sources. 2nd Int. Workshop on Information Theory and its Applications, San Diego, CA, USA, Jan. 2007. 44. Papoulis, A. (1991). Probability, random variables, and stochastic processes. McGraw- Hill. Author Biographies Hamid Behroozi received the B.Sc. degree from University of Tehran, Tehran, Iran, the M.Sc. degree from Sharif University of Technology, Tehran, Iran, and the Ph.D. degree from Concordia University, Montreal, QC, Canada, all in Electrical Engineering in 2000, 2003, and 2007, respectively. He is currently a Postdoctoral Fellow in the Department of Mathematics and Statistics, Queen’s University. His research interests include information theory, joint source-channel coding, and cooperative communications. Dr. Behroozi is the recipient of several academic awards including Ontario Postdoctoral Fellowship awarded by the Ontario Ministry of Research and Innovation (MRI), Quebec Doctoral Research Scholarship awarded by the Government of Quebec (FQRNT), Hydro Quebec Graduate Award, and Concordia University Graduate Fellowship. M. Reza Soleymani received the B.S. degree from the University of Tehran, Tehran, Iran, in 1976, the M.S. degree from San Jose State University, San Jose, CA, in 1977, and the Ph.D. degree from Concordia University, Montreal, QC, Canada, in 1987, all in electrical engineering. From 1987 to 1990, he was an Assistant Professor with the Department of Electrical and Computer Engineering, McGill University, Montreal. From October 1990 to January 1998, he was with Spar Aerospace Ltd. (currently EMS Technologies Ltd.), Montreal, QC, Canada, where he had a leading role in the design and development of several satellite communication systems. In January 1998, he joined the Department of Electrical and Computer Engineering, Concordia University, as a Professor. His current research interests include wireless and satellite communications, information theory and coding. 123