Skip to main content

Novel solutions for side information generation and fusion in multiview DVC


One of the key problems in distributed video coding is the generation of side information. This task consists of producing an estimate of an image with some neighboring ones, such as those taken by the same camera at different time instants, or, in the case of multiview setups, images taken at the same time instant by different cameras. If both estimates are available, a further problem arises, which is how to merge them in order to create a single side information. This problem is very relevant since a good estimate of the unknown image will require only a few bits to be corrected. Considering a multiview distributed video-coding setup, we propose a novel technique for inter-view interpolation based on occlusion prediction, a new fusion technique from multiple estimates, and finally an adaptive validation step for switching among the three possible side information images: temporal, inter-view, and fusion. We provide a comprehensive set of experimental results, which indicate bit rate reductions of more than 9% in average; moreover, we observe much more consistent results with respect to state-of-the-art techniques.

1 Introduction

Even if the distributed source coding theory is more than 30 years old[1, 2], it is only in the last 10 years that practical distributed video coding (DVC) systems have been proposed[3]. Since then, this topic has gathered much attention from the research community[4, 5] because several applications could benefit from efficient DVC, as for example interactive multiview video streaming[68].

In a typical DVC system, the encoder is not able to jointly encode all the input images. There can be several reasons: there are multiple sources that cannot communicate, as in the classical DVC paradigm; the joint encoding process is too computational expensive, such as in the case of light-weight sensor networks; or, even if the encoder knows all the input images, it does not know which ones have been requested by the user, such as in the case of interactive multiview video streaming. In all these cases, for these unknown images, the encoder only sends some parity bits of an error-correcting code. These bits will be used by the decoder to correct a suitable estimate of the current image, which is produced using the available information such as images from the same camera at different time instants, or, for multiview systems, images from different cameras at the same instant. This estimate is called side information (SI). It is clear that the more the SI is similar to the actual image, the less bits will be requested to correct it[9]. Therefore, side information generation is a crucial step for an efficient DVC system.

In this paper, we provide a system for SI generation in multiview DVC made up of three main components: temporal interpolation, inter-view interpolation, and adaptive fusion. We implement them within a system based on the architecture proposed by Aaron et al.[3], which is also at the basis of the popular distributed coding for video services (DISCOVER)[10] and VISNET I/II[11] DVC systems. In such a scheme (shown in Figure1), the input images are split into key frames (KFs) and Wyner-Ziv frames (WZFs).

Figure 1

The reference DVC system[3].

The KFs are encoded with a still image coding technique; at the decoder, they are used to produce an estimate of the current WZF, i.e., the SI. This estimate is then corrected by requesting a suitable number of bits from a channel code. Therefore, the encoder need not to know the SI to produce these bits: the encoding of KFs and WZFs is actually distributed. The Wyner-Ziv coder usually works in the DCT domain: the WZFs are transformed and quantized, and the channel code produces the parity bits for the coefficients’ bit planes. At the decoder side, the SI is in turn transformed, and the received parity bits are used to correct the DCT bands. Further bits are requested until the estimated bit error rate drops below a given threshold, then each coefficient value is reconstructed using a minimum mean square error estimator, and the IDCT returns the decoded Wyner-Ziv frame.

The technique for SI generation should take into account as much as possible the a priori information about the DVC system. In particular, for a multiview DVC system, different techniques exist for the case of temporal interpolation (SI generated from images of the same camera at different instants) or inter-view interpolation (SI generated from images of different cameras at the same time). When both estimates are available, a further problem arises, which is called fusion: for each pixel, should we use the temporal estimate, the inter-view estimate, or a combination of both? In this paper, we propose new solutions to these problems. The first contribution is a novel solution for the inter-view SI generation, based on occlusion avoidance. For the temporal interpolation, we consider a method we proposed for monoview video, and we adopt it in the context of multiview. We then provide a new method for SI fusion, based on two contributions: an occlusion detection method and an adaptive decision algorithm based on the rate of the channel code. The latter allows to overcome the performance inconsistency that has been often observed in previous fusion algorithms. Finally, we consider a multiview DVC codec that takes advantage of all these improvements, and we compare it with state-of-the-art systems. An extensive experimental validation fully supports the proposed techniques.

The rest of the paper is organized as follows. In Section 2, we describe the state of the art for interpolation and fusion. Then, two sections (Sections 3 and 4) are devoted to the new inter-view interpolation technique and the fusion algorithms. Experimental results (Section 5) are then followed by the conclusion and the outline of future work (Section 6).

2 State of the art for side information generation

Let us introduce some notations. For temporal SI generation, we consider monoview video sequences (or a single view from a multiview setup). Therefore, we only need a single (temporal) index to designate images: with I t , we refer to the t th image of a video sequence. When we deal with multiview systems, we need a second index for the view point. We note as I t,k the image taken at the instant t by the k th camera. In both cases, the corresponding SI is indicated with a hat. When the indexes are not necessary, we will simplify the notation referring to the temporal interpolation as I ̂ T , to the inter-view interpolation as I ̂ V and to their fusion as I ~ .

In this section, we will describe the motion interpolation algorithm proposed within the DISCOVER project that will be the reference method for comparison with our proposed algorithms. Then, we will provide a state of the art about techniques improving the linear interpolation of DISCOVER both in time and in the view domain. Finally, a state of the art on fusion techniques is given.

2.1 Side information generation by temporal interpolation

2.1.1 The DISCOVER algorithm

The most popular method for temporal image interpolation is the one proposed within the DISCOVER project[10]. It consists of the following steps. First, the two KFs,a say I t-1 and I t+1, are low-pass filtered in order to smooth out the noise. Then a block-matching motion estimation from I t+1 to I t-1 is performed. The resulting motion vector field v(·) is split in a couple of fields u(·) and w(·) (pointing from t respectively to t - 1 and t + 1). For the block of the WZF centered in p, this is obtained by looking for the trajectory closest to p: in the hypothesis of linear motion, the block centered in q at time t + 1, is centered inq+ 1 2 v(q) at time t. Then, we select the position q , such that

q (p)=arg min q N ( p ) q + 1 2 v ( q ) - p 2

whereN(p) is the set of the block center positions near p. The vectors v and w are defined asu(p)= 1 2 v( q (p)) andw(p)=- 1 2 v( q (p)). Next, a bidirectional motion refinement is performed. Finally, a weighted median filter is applied on the resulting motion vector fields in order to regularize them. After this process, one ends up with two motion vector fields that are used to compensate the previous and the next key frame. The average of the resulting images is the side information I ̂ t . We observe that, even if DISCOVER uses a simple motion model (linear trajectories), it gives very good results in some cases and is so popular that it will be used as reference to validate the proposed SI generation technique.

2.1.2 Other techniques

Several algorithms have been proposed in order to improve the DISCOVER performance. Huang et al.[12] have proposed to improve the forward motion estimation using also the chroma components of the decoded KFs, as already proposed for wavelet-based video coding in[13]. They further proposed an adaptive weighted overlapped block motion compensation: each compensated block is weighted by the inverse of MSE between the forward and backward compensated blocks. In the work of Macchiavello et al.[14], the authors have slightly modified the DISCOVER algorithm: without shifting vectors, the splitting is performed. The blank areas are filled with a block matching between these partially blank blocks and the adjacent KFs.

Instead, Ascenso et al.[15] have proposed to add a constraint during the motion estimation from t + 1 to t - 1: all the motion vectors have to cross the center of a block in the WZFs. Kubasov et al.[16] have proposed to replace block matching technique for motion estimation by mesh-based motion-compensated interpolation, in order to take into account more complex motion than simple translation. Mys et al.[17] have proposed to use the SKIP mode (like the one used in H.264/AVC) in DVC. Since there are several blocks that do not move between I t-1 and I t+1, their SI does not need to be corrected, similar to what is done in the PRISM codec[5, 18]. In[19, 20], it is proposed to use dense vector field techniques such as the Cafforio-Rocca algorithm and a total variation-based algorithm, for motion estimation and interpolation. The results show an average rate reduction up to 5.9%.

In the two works of Ye et al.[21, 22], a partially decoded WZF (using the parity bits of the DC band) allows the detection of suspicious vectors and to refine the motion vector fields. Finally, an optimal motion compensation mode selection is proposed: the previous, the next frame, the bidirectional motion-compensated average of the previous and the next frame are used for the SI by evaluating which of them gives the smallest matching error. Abou-El-Ailah et al.[23, 24] have proposed to fuse global and local motion estimation for constructing the SI: for global motion estimation, the parameters that model the global motion (translational, affine, or perspective) are estimated by SIFT features and sent to the decoder. Local motion estimation is obtained by DISCOVER algorithm. These two SI are fused during the decoding process using also the partially decoded WZF. Martins et al.[25] have proposed an iterative refinement as in[21, 26] for each band but only for some selected blocks. Let Y be the initial SI and let R b be the partially decoded WZF up to band b. The blocks for which the MAD between Y and R b is larger than a given threshold are refined by searching among the neighboring blocks the one that minimizes the MAD. This will be used as new SI from the correction of the next band. This refinement is performed for each band.

Hash-controlled motion estimation techniques have also been proposed: some additional information, called hash signature, is sent to the decoder[27, 28], for example, some blocks of the original WZF. In this context, Verbist et al.[29] have proposed a probabilistic model for SI construction. An overlapped block motion estimation is performed from I t-1 to I t+1, and a collection of candidate predictor values is available for each pixel. Then, by supposing that the noise between the side information and the real WZF can be modeled as Laplacian, the best candidate can be chosen by maximum likelihood estimation.

Recently, we proposed a temporal interpolation technique based on high-order motion interpolation (HOMI)[3032] that outperforms the DISCOVER RD performance. In this algorithm, the basic idea is to use four frames, I t-3, I t-1, I t+1, and I t+3, to estimate the current WZF I t . At first, the DISCOVER motion interpolation algorithm is performed on the frames I t-1 and I t+1. Then, the obtained vectors are lengthened to the frames I t-3 and I t+3 through a motion estimation criterion in order to estimate the trajectory of each block among the frames. The obtained four positions are interpolated in order to estimate the position of that block in the frame I t . Finally, the motion vectors for that block are obtained as the difference between the position in t and the positions in the next and the previous image. This algorithm will be used in the following temporal interpolation in our proposed algorithms.

Another framework for DVC developed in recent years is VISNET II codec[11]: the motion interpolation step for SI generation is the same as of DISCOVER. The main modifications at the decoder side are the iterative reconstruction of the WZF for each DCT band and an adaptive deblocking filter applied to the decoded WZFs. At the encoder, a CRC is added in order to improve robustness. A comprehensive review and classification of techniques proposed for temporal SI generation can be found in the work by Brites et al.[33].

2.2 Side information generation by inter-view interpolation

In multiview DVC systems, three types of camera are commonly considered: pure key cameras for which all the frames are KFs, pure Wyner-Ziv cameras where all the frames are WZFs, and hybrid cameras for which each second frame is a WZF and the others are KFs. These cameras can be arranged in a variety of configurations, but mainly three schemes have been considered in the literature[34]:

  •  Asymmetric scheme (see Figure2a). Pure key cameras (black) and pure WZ cameras (white) are alternated.

  •  Hybrid 1/2 scheme (see Figure2b). Every second camera is a pure key one, and the other (gray) is hybrid.

  •  Symmetric 1/2 scheme (see Figure2c). All the cameras are hybrid, and the KFs and the WZFs are placed on a quincunx grid in the time-view axes.

Figure 2

Configurations for multiview DVC. (a) Asymmetric scheme, (b) hybrid 1/2 scheme, and (c) symmetric 1/2 scheme. The filled frames are KFs, and the blank ones are WZFs. For simplicity, only three cameras per scheme are shown.

In all these schemes, for a generic WZF I t,n , at least two KFs are available, I t,n-1 and I t,n+1, i.e., two images taken at the same temporal instant k by two neighboring cameras. Moreover, except for the first scheme, two other images are available, namely I t-1,n and I t+1,n , the previous and next frames from the same camera. All these images should be used in order to generate a side information as reliable as possible. This is commonly achieved by three steps: a temporal estimation, an inter-view estimation, and finally a fusion.

In this section, inter-view estimation is considered; it can be seen as a form of image-based rendering (IBR) process. Under this perspective, this problem has been long studied[35], long before the rise of DVC systems. There is a simple approach in using the motion interpolation algorithms among neighboring views: the concept of motion is replaced by that of disparity. The results are acceptable as far as the views are rectified with a viewing axis perpendicular to the baseline[36]. This approach is simple but has the drawback of not taking into account the specificity of inter-view estimation and cannot assure the best performance. Therefore, in the literature, there are several approaches based on increasingly complex models of the multicamera setup. For example, Guo et al.[37] propose a global affine motion model with six parameters for inter-view interpolation. However, global motion models cannot accurately describe the sudden variation in the disparity field associated to object borders. Occlusions are neither taken into account.

Artigas et al.[38] have proposed a method based on depth maps. Indeed, depth maps along with camera parameters allow to create a virtual view point, namely that of the WZ camera. However, in our work, we consider a different problem where depth information is not directly available.

A similar approach is used by Ouaret et al.[39] for a hybrid 1/2 scheme with three cameras. In this case, the homography matrices that needed to map the KF into the WZ camera are estimated using a MMSE criterion. However, local object motion can create outliers, and this invalidates the estimation of the homography matrices. Moreover, this technique can suffer from distortion caused by camera lens.

Areia et al.[40] have proposed a very simple method for the hybrid 1/2 scheme with two cameras (stereoscopic video): the disparity field is computed over the last two decoded frames, and then it is directly applied to the current KF to generate the side information. The authors have recognized that this approach may not achieve the best performance, but it has the advantage of being very simple and of reusing the same algorithms as that of the temporal interpolation step.

2.3 Fusion techniques

When both temporal and inter-view interpolations are available, such as in the case of a hybrid or a symmetric 1/2 scheme (Figure2), they have to be combined in order to create a single side information. In the context of DVC, this operation is referred to as fusion.

A common reference is the ideal fusion[40]: for each pixel p, the value of the fused image I ~ (p) is selected as the temporal estimate I ̂ T (p) value or the inter-view estimate I ̂ V (p) value, according to the error with respect to the actual WZF. Of course, this method cannot be used in practice and serves only as a theoretical bound. Practical methods for image fusion are often based on the differences between the interpolated images and the KFs. Let us refer to the absolute difference between the temporal interpolation and the forward key frame (respectively, to the backward key frame) as e T F (respectively, to e T B ). Likewise, the difference between the inter-view interpolation and the left and right views are referred to as e V L and e V R . In[39], for each pixel p, the inter-view interpolation is chosen if e V L (p)< e T B (p) and e V R (p)< e T F (p). Otherwise, the temporal interpolation is used. In[41], an encoder-driven fusion is proposed. Each WZF is compared with respect to (w.r.t.) the previous and the forward frame. A binary mask is set to 0 or 1 if the value of each pixel of the WZF is closest to the co-located pixel of the previous or the forward frame, respectively. This mask is sent to the decoder by JBIG. At the decoder side, the two SI (temporal and inter-view) are compared w.r.t. the previous frame or the forward one according to the mask sent by the encoder. For each pixel, the closest SI is chosen for decoding. In[42], two fusion techniques are proposed: temporal motion (TM) interpolation projection fusion (IPF) and spatial view (SV) IPF. In TM-IPF, a block matching between the temporal estimate I ̂ T and the two adjacent KFs in time domain is performed. For the pixels where the two compensation errors are larger than a threshold, inter-view interpolation is used because temporal interpolation is supposed to have failed. Otherwise, the temporal interpolation is kept. SV-IPF is the counterpart of TM-IPF: I ̂ V is kept if its error relative to KFs is small; otherwise, I ̂ T is used. Guo et al. in[37] computed the absolute difference between the two motion-compensated frames in time domain. For pixels where this difference is larger than a threshold and if the norm of the two motion vectors is larger than a threshold, inter-view interpolation is chosen; otherwise, temporal interpolation is used. In[43], the decision on which SI has to be used is taken by observing the difference of the two SI on the neighboring, already decoded pixels. The correlation between the temporal/inter-view SI and the real WZF is assumed to be stationary, so they establish a criterion based on the error of the neighboring, already decoded pixels.

In[44], the fusion based on two different error images, referred to as E T and E V . E T , is the absolute difference between the motion-compensated backward reference and the motion-compensated forward reference. Likewise, disparity compensation on the left and right images is used to create E V . Two efficient fusion techniques are proposed: in the binary fusion, the pixel value in p is selected from the inter-view interpolation if E V (p) < E T (p); otherwise, it is selected from the temporal interpolation; in the linear fusion, the coefficientα= E T ( p ) E T ( p ) + E V ( p ) is computed and I ~ (p) is defined as a weighted average of I ̂ V (p) and I ̂ T (p) using, respectively, α and 1 - α as weights.

In[45], a new correlation model for the temporal and inter-view side information is proposed: the smaller the difference between the two SIs, the smaller is the variance of the probability density function (PDF) that models the correlation noise. Finally, a reconstruction with two SIs is performed such as that in[46]. Multihypothesis techniques are used by[47, 48] for fusing different SIs. Three SIs are generated (by block-based techniques, by optical flow, and by overlapped block motion compensation), and three parallel decoders are used in order to decode them. They choose the SI of the decoder that firstly converges. This technique has been extended also for the WZ coding of depth maps.

Recently, support vector machine was proposed by Dufaux[49] for fusion, which is considered as a classification problem with two classes. The features are extracted from the four KFs surrounding the WZF that has to be estimated. This technique is better than the previous methods even though it surpasses binary and linear fusion just slightly. Since, in addition, the latter methods are much simpler, we use them as reference for fusion performance.

3 Inter-view interpolation with priority to large disparity

We introduce a novel inter-view interpolation algorithm for multiview DVC for the case of aligned cameras or rectified sequences. In this case, the inter-view estimation is commonly produced by adapting the motion-compensated temporal interpolation algorithms since the object 'trajectories’ along views are practically straight lines. As a consequence, the DISCOVER algorithm works fairly well in this case, and considering higher order interpolation does not lead to significant gains. However, it does not take into account occlusions, so we can try to improve upon it by addressing this issue.

We took inspiration from the paper by Daribo and Pesquet-Popescu[50], where, in the context of multiview-plus-depth video coding, the original depth information was used to change the patch priority in the inpainting of occlusion areas in synthesized views. The key idea for our algorithm is quite similar: whenever two objects (actually, two blocks of pixels) of the reference KFs have estimated trajectories both passing near the current position, we take into account the object depth, instead of simply selecting the trajectory closer to the current position, as DISCOVER would do (see Equation 1). The foreground objects, having smaller depths, occlude the background and should be preferred by the SI generation algorithm. Therefore we modify Equation 1 by adding a penalization for small disparities (which translates into large depths). We obtain the following equation:

q (p)=arg min q q + 1 2 v ( q ) - p 2 - γ v ( q ) 2

where the penalization parameter γ > 0 is to be chosen experimentally. In conclusion, the proposed method consists in replacing Equation 1 of DISCOVER with Equation 2; the other steps of DISCOVER stay unaltered. We call this method 'interpolation with priority to large disparity’ (IPLD).

In Figure3, we show a schematic example where IPLD allows the selection of the right trajectory. In this figure, the y axis of frames is orthogonal to the drawing plane. After the first disparity estimation, the disparitiesb for blocks centered in q 1 and q 2 are respectively v(q 1) and v(q 2). Both blocks B 1 and B 2 could be interpolated in position p in the WZF. DISCOVER would simply select the block whose trajectory is closer to p, B 1 in this case. On the contrary, we should select the block that occludes the other, B 2 in this case. In order to estimate the occlusion, we penalize blocks with small disparities since they probably belong to the background. Thus, provided that a suitable value for the penalization parameter γ is chosen, IPLD is able to correctly select the block B 2.

Figure 3

A 1D representation for the IPLD algorithm. A 1D representation of a possible situation of two disparity vectors after the forward disparity estimation of DISCOVER. With DISCOVER, we would choose the green vector in the splitting procedure because it is closer to p than the red vector, but arguably, the object in the red block occludes the object in the green block within the frame in the k th view.

This algorithm is well suited for the inter-view estimation because, in this case, the disparity is strongly related to the depth. However, it would not be as much efficient for temporal interpolation since objects with higher velocity do not necessarily occlude objects with smaller velocity, even though if two objects have the same speed, the one closer to the camera will have a larger apparent velocity. This intuition is confirmed by preliminary experiments.

The increase of computational complexity of IPLD with respect to DISCOVER is negligible: when Equation 1 is replaced by Equation 2, for each candidate trajectory, we only need the computation of -γv(q)2 and an additional sum. This computational cost is dominated by two multiplications. Now, since in the splitting step only a small number of candidate trajectories is considered, we estimate the increase of computational complexity of our method in less than 20 multiplications per block. This is to be compared to the much heavier cost of motion estimation in the DISCOVER algorithm.

In Figure4, we show an inter-view-interpolated image from the 'newspaper’ sequence, using DISCOVER (Figure4a) and IPLD (Figure4b): our technique is much more successful in discriminating the background from the foreground.

Figure 4

A detail of the 'newspaper’ sequence. Inter-view interpolation generated (a) by DISCOVER (peak signal-to-noise ratio (PSNR) = 15.1 dB), and (b) by IPLD (PSNR = 28.3 dB). The original image is also shown (c).

4 Proposed fusion method

Since HOMI and IPLD allow the production of good temporal and inter-view interpolations of images, as a first novel contribution, we consider a multiview DVC system that just uses an existing fusion algorithm (such as those proposed in[44]) on these improved SI images; we expect that such a solution outperforms a reference system based on state-of-the-art interpolation methods. However, we can further improve the system performance with a novel fusion scheme based on occlusion detection.

Moreover, as it was observed in the past (e.g., in[44]), even though for the majority of test video sequences fusion improves upon temporal and inter-view interpolations, for some others, the opposite is true. In other words, there are sequences for which temporal or inter-view interpolation is much better than fusion; as a consequence, the rate-distortion (RD) performance comparison between fusion and interpolation is not very consistent even though, on the average, fusion has been reported for being the best solution.

In the second part of this section, we propose a novel Bayesian method for deciding among fusion, temporal interpolation or inter-view interpolation, based on the observation of the parity bit rate needed by the decoder. The resulting adaptive fusion method shows a much more consistent behavior and improved performance with respect to previous techniques.

4.1 Occlusion detection-based fusion

For the fusion problem, we consider a scheme where a WZF I t,k has four neighboring KFs: I t,k-1, I t,k+1, I t-1,k , and I t+1,k . For example, we can consider an hybrid camera in a hybrid 1/2 scheme (Figure2b) or any camera (except the first and the last) in a symmetric 1/2 scheme (Figure2c). For the time being, any temporal interpolation and inter-view interpolation algorithms can be used to produce the partial estimates, I ̂ T and I ̂ V . The proposed fusion between these images is performed as follows.

Let us suppose that we have the disparity fields d k-1,k+1(m,n) (from I t,k-1 to I t,k+1) and d k+1,k-1(m,n) (from I t,k+1 to I t,k-1). They can be obtained as side product of the inter-view interpolation process, or they can be estimated from scratch for the fusion process. In our implementation, we compute the disparity fields at the decoder from scratch using the method proposed by Miled and Pesquet[51] on the received KFs.

The coherence of the disparity fields is estimated by a left-right consistency crosscheck (LR-CC)[5254]: if the pixel p = (m,n) in the image k 1 corresponds to the pixel(m,n+ d k 1 , k 2 (m,n)) in the image k 2, the disparity of the former should be the opposite of the disparity of the latter. Therefore, the absolute sum of these quantities is a measure of the disparity coherence in the direction from k 1 to k 2. We can write

R k 1 , k 2 (m,n)=| d k 1 , k 2 (m,n)+ d k 2 , k 1 (m,n+ d k 1 , k 2 (m,n))|

For a perfectly coherent disparity, we would obtain R = 0 everywhere. However, we take into account a tolerance term τ ≥ 0. In the scientific literature, a common value for τ is 1 pixel[55, 56], at least in the context of stereo matching. Since our target is slightly different (image fusion in the context of multiview DVC), we do not use the value in the literature but rather optimize it experimentally. We define the occlusion map from k 1 to k 2 as follows:

O k 1 , k 2 (m,n)=u( R k 1 , k 2 (m,n)-τ)

where u(·) is the left continuous Heaviside step function. Using Equation 4, we perform an occlusion detection: O k 1 , k 2 shows the points of k 1 occluded (i.e., not visible) in k 2.

Finally, we use the occlusion maps O k-1,k+1 and O k+1,k-1 to improve the fusion process. For each pixel p, we decide whether to use only the temporal interpolation (since an occlusion has been detected) or to use a fused image, such as the linear fusion proposed in[44]. Now, the inter-view estimate I ̂ V (p) has been produced (using DISCOVER or IPLD) as the average of two disparity-compensated values:

I ̂ V (m,n)= 1 2 I t , k - 1 (m,n+e(m,n))+ 1 2 I t , k + 1 (m,n+f(m,n))

where e and f are respectively the estimates of d k,k-1 and d k,k+1, and they have been obtained by splitting.

Looking at Equation 5, we argue that I ̂ V (p) is affected by the possible occlusion of the pixel (m,n + e(m,n)) in the left image and of the pixel (m,n + f(m,n)) in the right image. Therefore, the proposed fusion formula is as follows:

I ~ ( p ) = I ̂ T ( p ) if O k - 1 , k + 1 ( m , n + e ( m , n ) ) = 1 I ̂ T ( p ) if O k + 1 , k - 1 ( m , n + f ( m , n ) ) = 1 I LIN ( p ) otherwise

In other words, we use only the temporal interpolation if one or both of the two pixels that contribute to I ~ (p) is estimated as occluded. Otherwise, we use the linear fusion proposed in[44], indicated as I LIN(p), with the only difference that we merge the estimates obtained by HOMI and IPLD instead of those obtained by DISCOVER.

We observe explicitly that in order to obtain our estimate of the WZF, we use four disparity fields (d k-1,k+1, d k+1,k-1, e, and f), and this feature characterizes our method with respect to existing techniques based on LR-CC.

4.2 Adaptive validation

As shown in the scientific literature and confirmed in our tests, even though fusion improves the quality of side information on the average, for some sequences, this is not true; on the contrary, the image generated by fusion is much worse than the one generated by inter-view interpolation only or temporal interpolation only. This happens for example for sequences where one interpolation is much better than the other (as shown for the 'outdoor’ sequence in[49] and in our experimental section). For these sequences, the correlation along one axis (temporal or inter-view) is much stronger than the correlation along the other.

Therefore, we would need a method to decide which image among I ̂ T , I ̂ V , and I ~ should be used as side information. This process can be seen as a final step of the fusion process, which validates I ~ or goes back to I ̂ T or I ̂ V : we call the proposed method adaptive validation (AV).

We recall that the best side information is the one that requires the smallest bit rate in order to be corrected by the channel decoder. Let us consider a given WZF I t,k and let us call R T , R V , and R F , the bit rate (in bits per pixel) for correcting temporal, inter-view, and fused SIs, respectively. The central idea of the proposed method is very simple: we take the decision among the three strategies based on observing, for a small number of WZFs, the parity bit rate needed to correct the temporal and inter-view interpolations. The rationale behind this idea is that fusion is ineffective only when one of the two interpolations is much worse than the other. Therefore, it suffices to observe the two rates R T and R V to estimate whether fusion is viable or not, while the value of R F is less relevant to this end.

Let us first discuss about the rate overhead associated to the evaluation of R T and R V . This is performed by actually asking the turbo encoder to send the parity bits needed to correct the temporal and the inter-view side information. More precisely, we use the turbo encoder of DISCOVER that consists of two interleaved punctured 1/2-convolutional encoders. The DCT bands are quantized according to the quantization index, and each DCT band is independently coded. Let P be the puncturing period (usually set to 48). Then, for each DCT band, the parity bit stream is structured into sub-blocks of size MN 16 P bits (one for each convolutional encoder)[57], where M × N is the spatial resolution of the frames and 16 is the number of pixels per block. At the first step, only a sub-block for each DCT band and for each bit plane is sent. Then, if needed, the decoder requests (via the feedback channel) a certain number of sub-blocks for the b th band for the p th bit plane. We call this number N j (p,b), where the index j can be T or V, in order to identify respectively the case of temporal or inter-view SI.

The total bit rate per pixel (R j with j {T,V}) would be

R j = 1 MN b = 1 16 p = 1 P ( q i , b ) 2 MN 16 P + 2 N j ( p , b ) MN 16 P

where P(q i ,b) is the number of bits used for quantizing the b th band depending on the quantization index q i . The number of parity bits sent to the decoder in order to correct the side information is given by two contributions: an amount depending only on the image resolution and on the quantization index that we will call R 0 and a variable amount depending on the SI quality expressed in terms of parity bit requests per band per bit plane.

Then, for two different SIs (temporal and inter-view), the value of N j (b,p) may be different for each bit plane and band. In the worst case, the sets of parity bits are disjoint (except for the common part R 0), so the rate needed to send both sets is R T  + R V  - R 0. In the most favorable case, one set of parity bits is a subset of the other; then, it suffices to send the parity bits needed for the worst case: the rate is max(R T ,R V ). In the general case, we need a bit rate between these two extreme cases.

Let us show the operation of the proposed system with an example taken from the encoding of the 'book arrival’ sequence. We use a symmetric 1/2 scheme and consider image number 1 from view 1, encoded as a WZF. The key frames have been encoded with QP = 31. In Table1, the values of N T (p,b) and N V (p,b) for the corresponding SI images I ̂ T and I ̂ V , generated using images 0 and 2 of view 1 and image 1 of views 0 and 2, respectively. The numbers of requests in italic means that N V (p,b) > N T (p,b), the number of requests in bold means that N V (p,b) < N T (p,b), and for the others, N V (p,b) = N T (p,b).

Table 1 Number of requests per bit plane and per band

It results that, for correcting the temporal SI, the encoder would send R T  = 0.164 bits per pixel (bpp). For correcting the inter-view SI, the encoder would send R V  = 0.185 bpp. Since a subset of the parity bits is in common for the two SIs, once, for example, the temporal parity bits have been sent, we need only R estimation = 0.216 bpp for correcting the inter-view SI. In conclusion, with this rate, the decoder is able to know R T and R V . The overhead for this operation is given by the difference between R estimation and the smallest among R T , R V , and R F . In this case, using the occlusion detection (OD) fusion, we would have obtained R F  = 0.135 bpp, so the estimation overhead is R estimation - R F  = 0.081 bpp.

The key point is to ask for this additional rate only for a subset of frames and to use the information about the parity rates R T and R V to decide, for the following WZFs, which side information to use. In other words, if we request the two sets of parity bits for all the WZFs, we incur in a large rate overhead and end up with a very inefficient system, which, in the best case, matches the performance of the worst side information. On the contrary, instead of using R T  + R V  - R 0 or max(R T ,R V ) bits, we would like to use a bit budget of R D bits, where D is the optimal decision, i.e.,

D=arg min d { T , V , F } R d

We estimate the optimal decision by observing the additional parity bit rate only for n over N WZFs, and for the following N - n, we use this decision. More precisely, for n frames out of N, we ask the channel coder the correction bits for both I ̂ T and I ̂ V . We call δ R  = R T  - R V the difference between the two rates. We expect the value of δ R to be quite correlated to the optimal decision: if this parameter is large in absolute value, one estimate is much better than the other, so it should be used; if it is small, the two interpolations have close quality, so the fusion should be used. This threshold-based approach is actually equivalent to a maximum a posteriori (MAP) estimation of the optimal decision, as shown below.

Let us model the optimal decision D as a discrete random variable with three possible values {T,V,F}. We want to estimate the best decision (according to Equation 7) given the parameter δ R modeled as a continuous random variable. We use the MAP value obtained by applying the Bayes’ law:

D ̂ = arg max d P ( D = d | δ R ) = arg max d p ( δ R | D = d ) P ( D = d ) p ( δ R ) = arg max d p ( δ R | D = d ) P ( D = d )

where P(·) is the probability mass function (marginal or conditional) for D and p(·) is the probability density function for δ R (marginal or conditional). In order to solve this problem, we have to find the three functions f d (δ R ) for d {T,V,F}, defined as

f d ( δ R ) = p ( δ R | d ) P ( d ) .

These functions are estimated off-line on a training set.

A couple of further issues has to be addressed, as illustrated in Section 5.3 in order to be able to use this method. First, as it was observed in some preliminary tests, the f d functions vary with the QP, so different thresholds must be determined as its function. Second, we have to decide how frequently to update the decision D ̂ during the coding process. We compute δ R for n WZFs out of N frames. For these n frames, we pay an overhead rate cost since we ask twice the channel bits to the encoder. Therefore, we cannot update δ R too often; otherwise, we lose all the advantage of having a better side information. On the other hand, a too small ratio n/N could affect the reliability of the decision. This dilemma is solved experimentally, but we can anticipate that small values of the n/N ratio (in the order of 1/100) work very well if scene changes do not occur.

5 Experimental results

In this section, we present the experimental results of the proposed SI generation and fusion methods. We consider a DISCOVER DVC system and modify the side information generation and fusion steps, using the methods proposed in the previous sections. The other tools remain the same: turbo codes are used for the WZFs; the correlation error is modeled as a Laplacian random variable. The statistical properties are evaluated on the residual error between the forward and the backward motion-compensated frames (or the left and the right disparity-compensated frames for inter-view correlation). The KFs are coded with the INTRA codec of H.264/AVC with four QPs: 31, 34, 37, and 40.

In order to evaluate the contributions of our techniques, we consider several configurations.

First, we consider the inter-view interpolation method IPLD: at this end, we use an asymmetric scheme (Figure2a), such that there is no temporal interpolation or fusion whose performance could affect the evaluation. Only the performance on the WZ views is considered.

Then, we consider the global fusion scheme, made up of the temporal interpolation, the inter-view interpolation, and the fusion algorithm; in order to isolate the contributions of the occlusion detection fusion and of the adaptive validation, we consider separately the cases when the latter is turned off or on. For these experiments, we need a scheme where both temporal and inter-view interpolations are available: we can consider the even views in a hybrid 1/2 scheme (Figure2b) or all but the side views in a symmetric 1/2 scheme (Figure2c).

Therefore, we now evaluate separately the contributions of IPLD, occlusion detection, and AV; for each of the three methods, we also report the experiments performed in order to tune their parameters. Then, we validate their effectiveness on the end-to-end rate-distortion performance for the encoded view(s). The latter is measured using the Bjontegaard metrics[58] with respect to the reference DISCOVER system: both the peak signal-to-noise ratio (PSNR) variation (ΔPSNR) and the percent rate variation (Δ R) are reported. For the IPLD technique, we also report the SI quality (measured in terms of PSNR with respect to the original WZF).

These experiments have been performed on a test set composed of nine popular multiview sequences, listed in Table2. They are characterized by different baseline distances, spatial resolutions, disparity ranges, and amount of motion in order to represent a sufficiently wide range of experimental conditions.

Table 2 The multiview sequences test set

5.1 Inter-view interpolation results

For all the experiments in this subsection, we consider an asymmetric scheme (Figure2a).

In the first set of experiments, we have to set the value of parameter γ of Equation 2, which defines the penalization for small disparity (i.e., background) blocks. In order to set the value of γ, we test our algorithm on five rectified multiview video sequences, with 20 frames per sequence. The neighboring views are encoded with H.264/AVC INTRA. In our experiments, the value maximizing the PSNR of the SI with respect to the WZF is γ = 0.6. By experiments, we have found that the optimal value of γ is negligibly affected by the QP. Therefore, we keep this value.

In the second experiment, we compute the SI for all the test sequences with 100 frames per sequence at various QP for the reference KFs. The resulting PSNR (with respect to the original WZF) is compared with the one achieved by DISCOVER, and the difference is reported in Table3. We observe that the proposed method improves almost always the DISCOVER quality, sometimes with very significant gains (up to 3.34 dB). This is because, as expected, our method succeeds at well reconstructing objects belonging to the foreground. We observe that for sequences such as 'lovebird’, 'pantomime’, 'outdoor’, the proposed method does not improve with respect to DISCOVER since they have a small range of disparity values; in such cases, using disparity to discriminate foreground and background is less successful. On the contrary, when the disparity range is larger, like in 'newspaper’ and 'balloons’, the improvement is remarkable.

Table 3 SI improvement (dB): IPLD versus DISCOVER

In the last experiment, we compare the end-to-end rate-distortion performance of a complete DVC system using IPLD for inter-view interpolation with another using DISCOVER, i.e., the temporal interpolation described in Section 2 is applied along the view domain. The results, shown in Table4, are related to the Wyner-Ziv camera only since the encoding of the KFs does not change in the two cases. We observe that the large improvement in SI quality shown in Table3 is actually translated into a significant bit rate reduction for the end-to-end system, in particular for sequences such as 'newspaper’ and 'balloons’, where background and foreground are hard to be distinguished. For these sequences, we obtain a bit rate reduction up to more than 20% and a gain in PSNR of around 1 dB. As for the previous experiments, for the sequences with small disparity range, we observed some small losses. However, on average, the IPLD allows a bit rate reduction of more than 7.7% achieved with a very small complexity increase.

Table 4 RD performance of IPLD vs. DISCOVER

5.2 Results for fusion with occlusion detection

Fusion techniques can be used when both temporal and inter-view interpolations are available; therefore, we can consider the hybrid cameras in a hybrid 1/2 scheme or all the views (except for the first and the last) in a symmetric 1/2 scheme. The side information produced in such a scheme depends on the two interpolation techniques and on the fusion. In order to isolate the contribution of each element to the global performance, we consider the following cases:

  •  DT. SI generated with DISCOVER along the temporal axis;

  •  HT. SI generated with HOMI along the temporal axis;

  •  DV. SI generated with DISCOVER along the view axis;

  •  IV. SI generated with IPLD along the view axis;

  •  FD. SI obtained by linear fusion of DV and DT as in[44];

  •  FHI. SI obtained by linear fusion of HT and IV;

  •  OD. FHI with the occlusion detection;

  •  AV. OD with adaptive validation of fusion.

Comparing all these methods allows us to have a deeper insight on the contribution of all the proposed tools to the final performance. The FD method is considered the reference since its performance is better than much of the state of the art and only slightly worse than (but very close to) more recent fusion techniques, as shown in[49].

The first experiments are conducted to set the disparity tolerance τ in Equation 4. To this end, we evaluate the PSNR of the SI information generated with the OD method for 10 frames of seven test sequences for four different QPs. We repeat the test while varying the value of τ, and the resulting SI PSNR is reported in Figure5. We find that the best value is τ = 2, and this is independent from the QP.

Figure 5

PNSR of the fused SI. PNSR of the fused SI versus the value of τ for different QPs.

Now, in order to assess the impact of OD, we compute the RD performance of OD and FHI with respect to the reference FD[44]. The results are shown in Table5. We observe that just using better temporal and inter-view interpolations than DISCOVER (FHI column) allows a non-negligible gain in the average (1.6% bit rate reduction). These results are fairly improved when the occlusion detection technique is used. On the average, we gain almost another 5%, achieving a remarkable 6.57% rate reduction. However, for some sequences, the OD approach actually worsens the performance due to false-positive errors on occlusion detection.

Table 5 RD performance of the fusion methods vs. state-of-the-art technique [[44]]

Some insights about the impact of different methods are shown in Figure6. In the first row, we show a detail of the 'balloons’ sequence, for which three views are available (left, center, and right). We notice that there are several occluded areas between the two external views, such as the musical notes on the background. In this case, the block-based IPLD (bottom-left figure) does not provide a good estimate of the central view: the blocks near the balloon contours are badly estimated. Using fusion (in particular FHI), we achieve a significant improvement, as shown in the bottom-center figure. The pixel-wise adaptivity of the fusion allows the use temporal interpolation where the inter-view is judged as not reliable; yet, there are some residual artifacts since not all the occlusions are correctly detected. When we apply the occlusion detection technique, we can then exclude the inter-view contribution from the fusion for the affected pixels, and this further improves the quality of the SI (bottom-right figure).

Figure 6

Detail of the 'balloons’ sequence. In the first row, a detail of the left, the central, and the right view of sequence, are shown. The central one has to be estimated. In the second row, the SI obtained by IPLD, FHI, OD are shown.

5.3 Adaptive validation results

Now we consider the AV technique. The first experiments are devoted to the estimation of the distribution of δ R , the rate difference between the corrections of I ̂ T and I ̂ V , given the optimal decision D. This process is performed off-line. We consider a training set of 100 frames from seven multi-view sequences. For each frame, we compute the temporal interpolation I ̂ T , the inter-view interpolation I ̂ V , and the fusion with the OD algorithm, I ~ . Then, we evaluate the number of parity bits needed to correct the three images. This information gives us both the optimal decision (the one associated to the minimum rate) and the δ R parameter. Repeating this operation for all the frames and for the four QP values, we obtain a large set of samples from the distribution of δ R |D, along with the samples from the distribution of D. The relative frequencies of the latter constitute the estimate of the PMF of D. The samples of δ R |D can be used to estimate the corresponding PDF, using the Parzen window method[59]. We use thus a non-parametric estimate of these PDFs: this is a reasonable approach since our a priori knowledge of the problem hardly suggests a mathematical model for these distributions.

We are now able to compute the f d functions: in Figure7, we show them for QP = 37. It is obvious that selecting the function with the maximum value for a given δ R amounts to compare this parameter with a couple of thresholds, given by the intersections of the f d curves. Our experiments allow us to find these thresholds for different QP values. These values, reported in Table6, have been estimated off-line and will be used for all the sequences. We also observe that the data used for these estimates and the data used for the validation form two disjoint sets.

Figure 7

Functions f d R ) for QP = 37. If δ R  < T 1, temporal interpolation is chosen; if T 1 ≤ δ R  < T 2, we decide for fusion; and if δ R  ≥ T 2, the inter-view interpolation is used.

Table 6 Optimal values ofT 1 andT 2 estimated off-line

For the running example given in Section 4.2, we would have obtained δ R  = -0.021 bpp, assuming n = 1, i.e., only one image is used to estimate the rates. Since the QP is equal to 31, Table6 makes us conclude that the best decision is to use the SI generated by fusion. This decision will be kept for a set of N frames, and therefore, the overhead of 15,925 bits is to be shared among all these images.

We also have to find the optimal update frequency for δ R , i.e., the optimal values for parameters n and N defined in Section 4.2. In order to do this, we simply run the adaptive fusion algorithm with different values for these parameters as shown in Table7. We find that the global RD performance (measured by the Bjontegaard metrics with respect to our reference FD) improves when N increases and when n decreases. In particular we observe that they roughly depend on the ratio n/N rather than separately on these two parameters. In conclusion, this experiment allows us to conclude that, even if we have a large N and we use as few as n = 1 image to take the decision, we have very good performance, with a very small overhead since n/N is small. This means that the optimal decision for fusion stays almost unchanged within a given sequence, and then the rate overhead due to a too frequent update is not worth. Therefore, in the following we use n = 1 and N = 100, and so we have a (small) rate overhead only for one over one hundred frames. Nevertheless, this allows us to take the best decision for the next WZFs.

Table 7 RD Performance of AV w.r.t. FD[44] for different values of n andN

For the running example, the rate overhead needed to perform the adaptive validation amounts to 15,925 per N = 100 images, i.e., to 0.00081 bits per pixel.

These results seem to point out that the update frequency could be as low as the scene-change frequency since, within a single sequence, the best decision seems not to change. As a consequence, another approach could be envisaged: we could update δ R only when a scene change is detected.

These considerations are validated by introducing the concept of oracle, which is a decisor that a priori knows which the best side information is among temporal, inter-view, and fusion. Consequently, it uses directly the best SI without paying the cost of sending the redundant parity bits. The oracle is thus equivalent to the AV algorithm with n = 1 and N = 1 but with zero overhead. Of course, on one hand, the oracle is not realizable in practice; on the other hand, none of the proposed techniques can outdo it. Therefore, it is rather useful to understand whether the proposed estimator works well or not. In Table8, for each sequence, we list the decision provided by the AV algorithm and compare them to the decisions taken frame-by-frame by the oracle. We observe that for seven sequences out of nine, the oracle chooses almost always the same SI, which is the one decided by the AV method. This is quite a good news: for these sequences, the proposed method is able to pick the best SI even though it observes δ R only for one image over 100. For the other sequences, even if the AV does not choose the best SI, this does not affect the performance a lot since this happens only when all the estimates have comparable qualities. To support this interpretation, we report in Table9 the RD performance of HT, IV, OD, and AV with respect to the oracle. We observe that AV may not be the best method on one given sequence because of the rate overhead for δ R estimation. For example, for the 'lovebird’ sequence, the oracle would choose the HT method all the time; so does the AV method, but it needs a small bit rate overhead to take the decision.

Table 8 Decision for each sequence
Table 9 RD performance for HT, IV, OD and AV vs. the oracle

However, the AV method is the only one that can adapt to the different characteristics of the sequences, and for this reason, it has the best average performance, and it never has catastrophic increases with respect to the oracle as it happens for all the other methods: the temporal interpolation can cost up to 21% more than the oracle, the inter-view interpolation can cost up to 70%, and even the OD fusion can have a loss of about 11%. The AV fusion needs only 6.8% additional bit rate with respect to the oracle in the worst case and about 2.5% in average. The consistency of the RD performance is one of the most valuable properties of the proposed technique with respect to the state of the art.

Finally, in Table10, we give the RD results for AV versus FD[44] and H.264/INTRA. In H.264/AVC INTRA, all the frames are INTRA coded (high profile, CABAC), i.e., without exploiting temporal or inter-view correlations. We observe that our method can achieve an average bit rate reduction of 9.11% w.r.t.[44] and of 22.38% w.r.t. H.264/AVC INTRA.

Table 10 RD performance of adaptive validation w.r.t. vs. [[44]] and H.264/AVC INTRA

6 Conclusions

In this paper, we propose several methods to improve the quality of the estimate of the Wyner-Ziv frames in multiview DVC. We use the HOMI algorithm for temporal interpolation, taking advantage of its possibility of detecting non-linear trajectories. For the inter-view interpolation, we introduce a method that allows a better discrimination between background and foreground. Finally, as far as the fusion method is concerned, we introduce two innovations: an occlusion detection-based fusion and an adaptive fusion, capable of restoring the inter-view or temporal interpolations instead of the fusion when needed. The originality of the latter method resides in the fact that we use a MAP optimal decision based on the observation of the bit rate requested by the channel decoder. The resulting performance is significantly better than state-of-the-art SI generation and fusion techniques: we obtain up to 7.7% of bit rate reduction for inter-view interpolation. The occlusion detection-based fusion allows a rate reduction of 6.6%, while using adaptive validation, we achieve up to more than 20% rate reduction and obtain an average of more than 9%. As future work, we want to explore multiview video-plus-depth scenarios: depth information can improve the quality of the SI. Moreover, camera parameters and DIBR algorithm (using the depth maps) can help us in the generation on the SI along the view domain without communication between the cameras.


a For simplicity, in the following, we always consider the case where every second frame is a KF, i.e., the GOP size is 2. All the algorithms can be easily extended to larger GOP sizes.

b The 'disparity vector’ has of course a null vertical component. This vectorial notation allows a simple problem formulation.


  1. 1.

    Slepian D, Wolf JK: Noiseless coding of correlated information sources. IEEE Trans. Inform. Theory 1973, 19: 471-480. 10.1109/TIT.1973.1055037

    MathSciNet  Article  MATH  Google Scholar 

  2. 2.

    Wyner A, Ziv J: The rate-distortion function for source coding with side information at the receiver. IEEE Trans. Inform. Theory 1976, 22: 1-11. 10.1109/TIT.1976.1055508

    MathSciNet  Article  MATH  Google Scholar 

  3. 3.

    Aaron A, Zhang R, Girod B: Wyner-Ziv coding of motion video. In Conference Record of the Thirty-Sixth Asilomar Conference on Signals, Systems and Computers, Pacific Grove, 3–6 Nov 2002. IEEE, Piscataway; 2002:240-244.

    Chapter  Google Scholar 

  4. 4.

    Girod B, Aaron A, Rane S, D Rebollo-Monedero D: Distributed video coding. Proc. IEEE 2005, 93: 71-83.

    Article  Google Scholar 

  5. 5.

    Puri R, Majumdar A, Ramchandran K: PRISM: a video coding paradigm with motion estimation at the decoder. IEEE Trans. Image Process 2007, 16(10):2436-2448.

    MathSciNet  Article  Google Scholar 

  6. 6.

    Cheung G, Ortega A, Cheung N: Interactive streaming of stored multiview video using redundant framestructures. IEEE Trans. Image Process 2011, 20(3):744-761.

    MathSciNet  Article  Google Scholar 

  7. 7.

    Petrazzuoli G, Cagnazzo M, Dufaux F, Pesquet-Popescu B: Using distributed source coding and depth image based rendering to improve interactive multiview video access. In 18th IEEE International Conference on Image Processing (ICIP), Brussels, 11–14, Sept 2011. IEEE, Piscataway; 2011:605-608.

    Google Scholar 

  8. 8.

    Gelman A, Dragotti P, Velisavljevic V: Interactive multiview image coding. In 18th IEEE International Conference on Image Processing (ICIP), Brussels, 11–14, Sept 2011. IEEE, Piscataway; 2011:601-604.

    Chapter  Google Scholar 

  9. 9.

    Maugey T, Gauthier J, Cagnazzo M, Pesquet-Popescu B: Evaluation of side information effectiveness in distributed video coding. In IEEE Transactions on Circuits and Systems for Video Technology. IEEE, Piscataway, 2013 in press;

    Google Scholar 

  10. 10.

    Artigas X, Ascenso J, Dalai M, Klomp S, Kubasov D, Ouaret M: The DISCOVER codec: architecture, techniques and evaluation. In Picture Coding Symposium (PCS’07). Lisbon; November 2007.

    Google Scholar 

  11. 11.

    Ascenso J, Brites C, Dufaux F, Fernando A, Ebrahimi T, Pereira F, Tubaro S: The VISNET II DVC codec: architecture, tools and performance. In European Signal Processing Conference (EUSIPCO 2010). Aalborg; 23–27 August 2010.

    Google Scholar 

  12. 12.

    Huang X, Forchhammer S: Improved side information generation for distributed video coding. IEEE 10th Workshop on Multimedia Signal Processing, Cairns, 8–10 Oct 2008. Piscataway: IEEE; 2008:223-228.

    Google Scholar 

  13. 13.

    Andre T, Pesquet-Popescu B, Gastaud M, Antonini M, Barlaud M: Motion estimation using chrominance for wavelet-based video coding. In Proceedings of IEEE Picture Coding Symposium. San Francisco; December 2004.

    Google Scholar 

  14. 14.

    Macchiavello B, Brandi F, Peixoto E, de Queiroz R, Mukherjee D: Side-information generation for temporally and spatially scalable Wyner-Ziv codecs. J. Image Video Process 2009, 2009: 14.

    Google Scholar 

  15. 15.

    Ascenso J, Pereira F: Advanced side information creation techniques and framework for Wyner-Ziv video coding. J. Vis. Commun. Image Representation 2008, 19(8):600-613. 10.1016/j.jvcir.2008.06.001

    Article  Google Scholar 

  16. 16.

    Kubasov D, Guillemot C: Mesh-based motion-compensated interpolation for side information extraction in distributed video coding. IEEE International Conference on Image Processing, Atlanta, 8–11 Oct 2006. Piscataway: IEEE; 2006:261-264.

    Google Scholar 

  17. 17.

    Mys S, Slowack J, Skorupa J, Lambert P, de Walle RV: Introducing skip mode in distributed video coding. Signal Process Image Commun 2009, 24(3):200-213. 10.1016/j.image.2008.12.004

    Article  Google Scholar 

  18. 18.

    Fowler JE: An implementation of PRISM using QccPack. Technical report, Mississippi State University 2005.

    Google Scholar 

  19. 19.

    Cagnazzo M, Maugey T, Pesquet-Popescu B: A differential motion estimation method for image interpolation in distributed video coding. IEEE Intern. Conf. Acoust., Speech and Sign. Proc 2009, 1: 1861-1864.

    Google Scholar 

  20. 20.

    Cagnazzo M, Miled W, Maugey T, Pesquet-Popescu B: Image interpolation with edge-preserving differential motion refinement. 16th IEEE International Conference on Image Processing (ICIP), Cairo, 7–10 Nov 2009. Piscataway: IEEE; 2009:361-364.

    Google Scholar 

  21. 21.

    Ye S, Ouaret M, Dufaux F, Ebrahimi T: Improved side information generation with iterative decoding and frame interpolation for distributed video coding. 15th IEEE International Conference on Image Processing (ICIP 2008), San Diego, 12–15 Oct 2008. Piscataway: IEEE; 2008:2228-2231.

    Google Scholar 

  22. 22.

    Ye S, Ouaret M, Dufaux F, Ebrahimi T: Improved side information generation for distributed video coding by exploiting spatial and temporal correlations. J. Image Video Process 2009, 2009: 4.

    Google Scholar 

  23. 23.

    Abou-Elailah A, Dufaux F, Farah J, Cagnazzo M, Pesquet-Popescu B: Fusion of global and local motion estimation for distributed video coding. Circuits Syst. Video Technol, IEEE Trans 2012, 22: 1-12.

    Article  Google Scholar 

  24. 24.

    Abou-Elailah A, Dufaux F, Farah J, Cagnazzo M: Fusion of global and local side information using Support Vector Machine in transform-domain DVC. Proceedings of the 20th European Signal Processing Conference (EUSIPCO), Bucharest, 27–31 Aug 2012. Piscataway: IEEE; 2012:574-578.

    Google Scholar 

  25. 25.

    Martins R, Brites C, Ascenso J, Pereira F: Refining side information for improved transform domain Wyner-Ziv video coding, Circuits Syst. Video Technol. IEEE Trans 2009, 19(9):1327-1341.

    Google Scholar 

  26. 26.

    Abou-Elailah A, Dufaux F, Cagnazzo M, Pesquet-Popescu B, Farah J: Successive refinement of side information using adaptive search area for long duration GOPs in distributed video coding. 19th International Conference on Telecommunications (ICT), Jounieh, 23–25 April 2012. Piscataway: IEEE; 2012:1-6.

    Google Scholar 

  27. 27.

    Tagliasacchi M, Tubaro S: Hash-based motion modeling in Wyner-Ziv video coding. Proceed. of IEEE. Intern. Conf. Acoust., Speech and Sign. Proc 2007, 1: 509-512.

    Google Scholar 

  28. 28.

    Park S, Lee YY, Kim CS, Lee SU: Efficient side information generation using assistant pixels for distributed video coding. Picture Coding Symposium (PCS), Krakow, 7–9 May 2012. Piscataway: IEEE; 2012:161-164.

    Google Scholar 

  29. 29.

    Verbist F, Deligiannis N, Jacobs M, Barbarien J, Munteanu A, Schelkens P, Cornelis J: Maximum likelihood estimation for the generation of side information in distributed video coding. 18th International Conference on Systems, Signals and Image Processing (IWSSIP), Sarajevo, 16–18 June 2011. Piscataway: IEEE; 2011:1-4.

    Google Scholar 

  30. 30.

    Petrazzuoli G, Cagnazzo M, Pesquet-Popescu B: High order motion interpolation for side information improvement in DVC. IEEE International Conference on Acoustics Speech and Signal Processing (ICASSP), Dallas, 14–19 March 2010. Piscataway: IEEE; 2010:2342-2345.

    Google Scholar 

  31. 31.

    Petrazzuoli G, Cagnazzo M, Pesquet-Popescu B: Fast and efficient side information generation in distributed video coding by using dense motion rep. 18th European Signal Processing Conference (EUSIPCO), Aalborg, 23–27 August 2010

    Google Scholar 

  32. 32.

    Petrazzuoli G, Maugey T, Cagnazzo M, Pesquet-Popescu B: Side information refinement for long duration GOPs in DVC. IEEE International Workshop on Multimedia Signal Processing (MMSP), Saint Malo, 4–6 Oct 2010. Piscataway: IEEE; 2010:309-314.

    Google Scholar 

  33. 33.

    Brites C, Ascenso J, Pereira F: Side information creation for efficient Wyner–Ziv video coding: classifying and reviewing. Signal Process, Image Commun 2013, 28(7):689-726. 10.1016/j.image.2013.05.002

    Article  Google Scholar 

  34. 34.

    Maugey T, Pesquet-Popescu B: Side information estimation new symmetric schemes for multi-view distributed video coding. J. Vis, Commun. Image Representation 2008, 19(8):589-599. 10.1016/j.jvcir.2008.09.002

    Article  Google Scholar 

  35. 35.

    Shum H, Kang S: A review of image-based rendering techniques. Proceed. Intern. Symp. Visual Comm and Proc. 2000. doi: 10.1117/12.386541

    Google Scholar 

  36. 36.

    Guillemot C, Pereira F, Torres L, Ebrahimi T, Leonardi R, Ostermann J: Distributed monoview and multiview video coding: basics, problems and recent advances. IEEE Signal Process. Mag 2007, 24(5):67-76.

    Article  Google Scholar 

  37. 37.

    Guo X, Lu Y, Wu F, Gao W: Distributed multi-view video coding. Proceed. of Intern Symp. Visual Comm. and Image Proc 2006, 6077: 290-297.

    Google Scholar 

  38. 38.

    Artigas X, Angeli E, Torres L: Side information generation for multiview distributed video coding using a fusion approach. Proceedings of the 7th Nordic Signal Processing Symposium (NORSIG), Rejkjavik, 7–9 June 2006. Piscataway: IEEE; 2006:250-253.

    Google Scholar 

  39. 39.

    Ouaret M, Dufaux F, Ebrahimi T: Fusion-based multiview distributed video coding. Proceedings of the 4th ACM International Workshop on Video Surveillance and Sensor Networks, Santa Barbara, 23–27 Oct 2006. New York: ACM; 2006:139-144.

    Google Scholar 

  40. 40.

    Areia J, Ascenso J, Brites C, Pereira F: Wyner-Ziv stereo video coding using a side information fusion approach. IEEE 9th Workshop on Multimedia Signal Processing (MMSP), Crete, 1–3 Oct 2007. Piscataway: IEEE; 2007:453-456.

    Google Scholar 

  41. 41.

    Ouaret M, Dufaux F, Ebrahimi T: Multiview distributed video coding with encoder driven fusion. in European Signal Processing Conference (EUSIPCO 2007), Poznan, 03–07 Sept 2007. Cary: InfoScience; 2007:3-7.

    Google Scholar 

  42. 42.

    Ferre P, Agrafiotis D, Bull D: Fusion methods for side information generation in multi-view distributed video coding systems. Proceed. of IEEE Intern, Conf. Image Proc 2007, 6: 409-412.

    Google Scholar 

  43. 43.

    Tagliasacchi M, Prandi G, Tubaro S: Symmetric distributed coding of stereo video sequences. Proceed. of IEEE Intern. Conf. Image Proc 2007, 2: 29-32.

    Google Scholar 

  44. 44.

    Maugey T, Miled W, Cagnazzo M, Pesquet-Popescu B: Fusion schemes for multiview distributed video coding. European Signal Processing Conference (EUSIPCO), Glasgow, August 2009

    Google Scholar 

  45. 45.

    Li Y, Liu H, Liu X, Ma S, Zhao D, Gao W: Multi-hypothesis based multi-view distributed video coding. In Picture Coding Symposium (PCS), Chicago, 6–8 May 2009. IEEE Piscataway; 2009:1-4.

    Chapter  Google Scholar 

  46. 46.

    Kubasov D, Nayak J, Guillemot C: Optimal reconstruction in Wyner-Ziv video coding with multiple side information. IEEE 9th Workshop on Multimedia Signal Processing (MMSP), Crete, 1–3 Oct. 2007. Piscataway: IEEE; 2007:183-186.

    Google Scholar 

  47. 47.

    Salmistraro M, Zamarin M, Rakêt LL, Forchhammer S, Fotonik D, Plads Ø: Distributed multi-hypothesis coding of depth maps using texture motion information and optical flow. In IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2013). Vancouver: Vancouver Convention Exhibition Centre; 26–31 May 2013.

    Google Scholar 

  48. 48.

    Huang X, Raket L, Luong M, Nielsen HV, Lauze F, Forchhammer S: Multi-hypothesis transform domain Wyner-Ziv video coding including optical flow. IEEE 13th International Workshop on Multimedia Signal Processing (MMSP), Hangzhou, 17–19 Oct 2011. Piscataway: IEEE; 2011:1-6.

    Google Scholar 

  49. 49.

    Dufaux F: Support vector machine based fusion for multi-view distributed video coding. 17th International Conference on Digital Signal Processing (DSP), Corfu, 6–8 July 2011. Piscataway: IEEE; 2011:1-7.

    Chapter  Google Scholar 

  50. 50.

    Daribo I, Pesquet-Popescu B: Depth-aided image inpainting for novel view synthesis. IEEE International Workshop on Multimedia Signal Processing (MMSP), Saint Malo, 4–6 Oct 2010. Piscataway: IEEE; 2010:167-170.

    Chapter  Google Scholar 

  51. 51.

    Miled W, Pesquet J: Disparity map estimation using a total variation bound. The 3rd Canadian Conference on Computer and Robot Vision, Quebec, 7–9 June 2006. Piscataway: IEEE; 2006:48-48.

    Google Scholar 

  52. 52.

    Fua P: A parallel stereo algorithm that produces dense depth maps and preserves image features. Mach. Vis. Appl 1993, 6: 35-49. doi:10.1007/BF01212430 10.1007/BF01212430

    Article  Google Scholar 

  53. 53.

    Egnal G, Wildes R: Detecting binocular half-occlusions: empirical comparisons of five approaches. IEEE Trans. Pattern Anal Mach. Intell 2002, 24(8):1127-1133. 10.1109/TPAMI.2002.1023808

    Article  Google Scholar 

  54. 54.

    Scharstein D: R Szeliski, High-accuracy stereo depth maps using structured light. in IEEE Conference on Computer Vision and Pattern Recognition. 2003, 1: 195-202.

    Google Scholar 

  55. 55.

    Hirschmüller H: Improvements in real-time correlation-based stereo vision. Proceedings of the IEEEWorkshop on Stereo and Multi-Baseline Vision (SMBV 2001), Kauai, 9–10 Dec 2001. Piscataway: IEEE; 2001:141-148.

    Google Scholar 

  56. 56.

    Yoon KJ, Kweon IS: Adaptive support-weight approach for correspondence search. IEEE Trans. Pattern Anal, Mach. Intell 2006, 28(4):650-656.

    Article  Google Scholar 

  57. 57.

    Rowitch DN, Milstein LB: On the performance of hybrid FEC/ARQ systems using rate compatible punctured turbo (RCPT) codes. Commun. IEEE Trans 2000, 48(6):948-959. 10.1109/26.848555

    Article  Google Scholar 

  58. 58.

    Bjontegaard G: Calculation of average PSNR differences between RD-curves. In VCEG Meeting. Austin; 2–4 April 2001.

    Google Scholar 

  59. 59.

    Parzen E: On estimation of a probability density function and mode. Ann. Math. Stat 1962, 33(3):1065-1076. 10.1214/aoms/1177704472

    MathSciNet  Article  MATH  Google Scholar 

Download references

Author information



Corresponding author

Correspondence to Marco Cagnazzo.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors’ original submitted files for images

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 2.0 International License ( ), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and Permissions

About this article

Cite this article

Petrazzuoli, G., Cagnazzo, M. & Pesquet-Popescu, B. Novel solutions for side information generation and fusion in multiview DVC. EURASIP J. Adv. Signal Process. 2013, 154 (2013).

Download citation


  • Distributed video coding
  • Side information
  • Occlusion detection
  • Inter-view interpolation
  • Fusion