Skip to main content

A deblocking algorithm based on color psychology for display quality enhancement

Abstract

This article proposes a post-processing deblocking filter to reduce blocking effects. The proposed algorithm detects blocking effects by fusing the results of Sobel edge detector and wavelet-based edge detector. The filtering stage provides four filter modes to eliminate blocking effects at different color regions according to human color vision and color psychology analysis. Experimental results show that the proposed algorithm has better subjective and objective qualities for H.264/AVC reconstructed videos when compared to several existing methods.

Introduction

Many video coding techniques have been developed for various kinds of applications, including multimedia communication, videophone, videoconferencing, video streaming, and high-definition television [14]. The goal of video coding techniques is to reduce transmission bitrates and storage data while maintaining a certain quality. The Joint Video Team (JVT) of ISO/IEC MPEG and ITU-T VCEG (Video Coding Experts Group) has developed a new standard for video coding. In December 2001, VCEG and MPEG (Moving Picture Experts Group) formed a JVT to finalize the draft of new coding standard for formal approval submission as H.264/AVC [5] in March 2003. H.264/AVC provides several enhancements to improve coding efficiency.

The discrete cosine transform (DCT) and motion estimation efficiently reduce spatial and temporal redundancies in most video coding standard. H.264/AVC divides an image into 4 × 4 or 8 × 8 block of pixels, and transforming each block from the spatial domain to the frequency domain by DCT. At low bit-rate coding, there may be only one DC and few AC coefficients to represent an image block. Hence, coarse quantization of transform coefficients cause noticeable discontinuities at block boundaries called blocking artifacts. In other words, blocking effects are the grid noise along block boundaries in a relatively homogeneous region.

Several previous studies [610] have been proposed for the design of deblocking filters, which can largely be categorized into three groups. The first group uses a low pass filter around block boundaries to remove blocking artifacts [6, 7]. The second group applies maximum a posterior probability technique to eliminate blocking artifacts [8]. The third group uses the set-theoretic reconstruction method, which defines constraint sets by using observed data or a prior knowledge of the solution [9, 10]. The intersection of all constraint sets gives the feasible set. All samples outside the feasible set are illegal and the projection onto convex sets [11] technique is used to project them to be legal. The success of set-theoretic method depends on the constraint sets highly. In [12], the classification algorithm classifies a block into three types: PLANE, EDGE, and TEXTURE, according to statistical characteristic of an image in the pixel domain. Each of these types has its corresponding filter to remove blocking artifacts. Similar to that in [12], the study in [13] classifies block type in the transform domain and proposes a wavelet-based block analysis. In [14, 15], some characteristics in human visual system are utilized to detect blocking artifacts; the appropriate filter is used to remove these artifacts.

In order to eliminate blocking effects, many deblocking filter algorithms have been proposed to provide nice solutions in still images [1618]. The low-pass filter is also employed to eliminate blocking artifacts in video compression standards [19]. Tai et al. [20] proposed a deblocking algorithm based on five filtering modes by considering the activity across block boundaries. These five filtering modes include three frequency-related modes (the smooth mode, the intermediate mode, and the complex mode), one special mode, and a refined mode. Tai et al. [21] proposed a new method to remove blocking artifacts in the low bit-rate block-based video coding in H.264/AVC and MPEG-4. However, it is necessary to modify the filtered block size in order to suit different codecs. Then, Chen and her coauthors [22] proposed a general method which is suitable for various video compression standards. Nevertheless, when edges cross block boundaries, Chen’s method may blur real edge.

H.264/AVC in-loop filter performs simple operations to detect and analyze artifacts on coded blocking boundaries and attenuates those by applying a selected filter [23]. The H.264/AVC loop filter is an adaptive filter. The amount of filtering performed each position of the block edge depends on some factors. The filter can be divided into three levels: slice level, block level, and pixel level. The filter strength, i.e., the amount of filtering is computed based on the help of parameter boundary strength (BS), depending on the current quantizer, the macroblock type, the motion vector, gradient of the image samples across the blocking boundary, and other parameters. However, the filtering algorithm is highly complex and is account for 33% of the total decode time in the study of the baseline profile decoder [24].

This article proposes a deblocking filtering algorithm that eliminates the blocking artifacts from H.264/AVC encoded videos. The preliminary results are published in [25]. The theoretic analysis and experiments of this article are more solid. In the detection process, the proposed “Sobel-Wavelet” method can clearly reveal the positions of artifacts, and the proposed filter eliminates blocking effects in RGB domain, instead of YCbCr domain, considering the effects of human color vision and color psychology. In objective evaluation, the PSNR of the proposed method is approximated to the H.264/AVC loop filter. The proposed method has better performance in GBIM than H.264/AVC loop filter and other existing methods. In subjective evaluation, our proposed method has significant improvement when compared to other existing methods. Our method is also close to the H.264/AVC loop filter. In conclusion, the proposed method has the advantage of the post-processing-based method in terms of simplify and flexibility, but the performance is very close to H.264/AVC.

The rest of this article is organized as follows. Section “Blocking artifacts detection” describes the proposed blocking artifacts detection. Then, the proposed deblocking algorithm is explained in Section “Deblocking artifact algorithm”. Experimental results are shown in Section “Experimental results” in order to evaluate the performance of the proposed scheme. Finally, concluding remarks and some future work are given.

Blocking artifacts detection

Because most blocking artifacts occur at 4 × 4 or 8 × 8 block boundaries in H.264/AVC encoded video, the filtering should focus on pixels around these block boundaries. If filtering modifies all block boundaries, it will blur images. Hence, it is important to detect blocking boundaries and exclude real edges to avoid over-smoothing image textures and keep objective quality. The proposed blocking artifacts detection is based on Sobel edge detector and wavelet-based edge detector. According to our observation and analysis, Sobel edge detector detects all strong and coherent edges including real edges and blocking boundaries. Under noise condition, Sobel detector may fail to detect the real edges of an image. Wavelet-based edge detection has better detection especially in noise conditions, while Sobel equally detects both blocking artifacts and real edges. Combining the results of Sobel and Wavelet detectors together, the proposed “Sobel–Wavelet” method can clearly reveal artifacts among real edges avoid over-smoothing problem. Figure 1 illustrates the flowchart of our proposed detection scheme.

Figure 1
figure 1

Flowchart of blocking artifacts detection.

Sobel edge detector

Sobel edge detector is a well-known tool of edge detection in image processing. 3 × 3 Sobel operator acts locally on an image and only detects edges at small scales. The Sobel edge detector is sensitive to high frequency; hence, Sobel can detect textures in an image. In general, Sobel edge detector uses two 3 × 3 kernels, H and V, to detect horizontal and vertical edges shown in Figure 2.

Figure 2
figure 2

Sobel edge detector.

Two directional edge maps, horizontal and vertical edge information, can be determined by Equations (1) and (2), respectively.

S h = H * F , S h : horizontal edge map,
(1)
S v = V * F , S v : vertical edge map,
(2)

where F is an image with the size M × N, and * means convolution operation. When four values S h 1 , S h 2 , S h 3 , and S h 4 are all greater than zero, these pixels belong to edges. And the same situation applies to S v 1 , S v 2 , S v 3 , and S v 4 . Otherwise, they just belong to textures in an image. Figure 3 illustrates the positions of four continuous pixels on the vertical and horizontal block boundaries. The decision function is given as

s h 4 m + i , 4 n = { 1 , S h 1 & S h 2 & S h 3 & S h 4 > 0 0 , otherwise , 0 i < 4 ,
(3)
s v 4 m , 4 n + j = { 1 , S v 1 & S v 2 & S v 3 & S v 4 > 0 0 , otherwise , 0 j < 4 ,
(4)

for 0 m M / 4 1 and 0 n N / 4 1 , where S h = 1 and S v = 1 represent that the corresponding pixels belong to edge and smooth regions, respectively. In Sobel edge detector, both real edges and blocking artifacts are detected and shown in the edge map. In Figure 4, the black lines indicate all strong and coherent edges detected by Sobel edge detector. In Figure 5, we can see that the black line indicates the original horizontal direction edge of the image.

Figure 3
figure 3

Edge map across 4 × 4 block boundaries.

Figure 4
figure 4

Results of edge detection by Sobel operator for the first frame of TableTennis by H.264/AVC at 140 kbps. (a) Vertical direction. (b) Horizontal direction.

Figure 5
figure 5

Details of the blue region. (a) TableTennis video sequence by H.264/AVC at 140 kbps and (b) horizontal direction edge detection.

Proposed wavelet-based edge detector

The proposed algorithm uses the well-known wavelet basis, Haar wavelet [26]. In the one-dimensional case, the Haar wavelet’s mother wavelet function ψ t can be represented as

ψ t = { 1 , 1 , 0 , 0 t 1 / 2 1 / 2 t 1 otherwise ,
(5)

and its scaling function φ t can be defined as

φ t { 1 , 0 t < 1 0 , otherwise .
(6)

Two-scale wavelet decomposition with mother wavelet function ψ t and scaling functions φ t generates four coefficients, W 1 H , W 2 H , W 3 H , and W 4 H . The positions of these four coefficients are shown in Figure 6. According to Equations (7) and (8), four wavelet coefficients are used to generate Equations (9) and (10). The blocking effects in an image have horizontal and vertical directional discontinuities. By multi-resolution analysis, cross-scale wavelet coefficients, which refers to the wavelet coefficients that have different scales but have the same orientation information such as W 1 H and W 2 H , are employed to generate two directional maps as

E H = W 1 H × W 2 H ,
(7)
E V = W 1 V × W 2 V ,
(8)
Figure 6
figure 6

Illustration of wavelet coefficient positions of W 1 H , W 2 H , W 1 V , and W 2 V .

Next, two thresholds, T h 1 and T h 2 , the average edge strengths in horizontal and vertical directions of one frame are used for generating edge maps in Equations (11) and (12). If E H m , n is equal to or greater than T h 1 , e H m , n belongs to an edge pixel; otherwise it belongs to blocking artifact one. In this part, the edge maps, e H and e V , show real edges and suppress the appearance of blocking artifacts.

e H m , n = { 1 , i f E H m , n T h 1 0 , otherwise ,
(9)
e V m , n = { 1 , i f E V m , n T h 2 0 , otherwise ,
(10)
T h 1 = 4 M N m = 0 M 4 1 n = 0 N 1 F 4 m , n F 4 m 1 , n ,
(11)
T h 2 = 4 M N m = 0 M 1 n = 0 N 4 1 F m , 4 n F m , 4 n 1 .
(12)

Sobel-wavelet detector

The proposed “Sobel–Wavelet” scheme combines the detected results of Sobel edge detector and Wavelet-based edge detector through the following equations:

L h m , n = S h m , n e H m , n ,
(13)
L v m , n = S v m , n e V m , n .
(14)

If the pixel belongs both to the edge in the Sobel set and the true edge in the wavelet set, this pixel is determined as true edge. L h m , n = 0 and L v m , n = 0 mean that F m , n belongs to true edge in horizontal and vertical directions, respectively. Figure 7a,b shows edge maps S h and S v . The black lines in Figure 7c,d shows blocking effects. These two figures show blocking artifacts obtained by jointly considering the detected results of Wavelet-based edge detector that has high probability to exclude true edges. Figure 8 is the details of the red region of Figure 7.

Figure 7
figure 7

Illustration of edge maps (the first reconstructed frame of Foreman by H.264/AVC at 105 kbps). (a) S h , (b) S v , (c) L h , (d) L v .

Figure 8
figure 8

Details of the red region of Figure7. (a) S h , (b) S v , (c) L h , (d) L v .

Deblocking artifact algorithm

Our proposed blocking artifacts detection has high probability to find blocking boundaries and exclude real edges compared to the conventional schemes. In this section, a new deblocking filter is proposed to eliminate blocking artifacts detected by the proposed “Sobel–Wavelet” scheme. Blocking artifacts exists both in luminance and chrominance components; however, most of the conventional schemes only focus on luminance for deblocking process. Not only in luminance, the study in [19] removes the blocking artifacts both in luminance and chrominance components and obtains better results. The proposed algorithm removes blocking artifacts in the RGB domain instead of YCbCr; therefore, it can simultaneously eliminate blocking artifacts from two components. This is because three components of filtering RGB have effect on both luma and chrominance components of the YCbCr domain. Figure 9 presents the detailed flowchart of the proposed deblocking filter based on the above concept.

Figure 9
figure 9

Flowchart of the proposed deblocking algorithm.

Some features from human vision [27] and color psychology [28] are used to design the proper filter to eliminate blocking effects. Four modes including complex mode, R mode, G mode, and B mode are used in our filtering stage. If the pixels belong to inverse colors across blocking boundary, this case is the complex mode. Note that inverse color is those on the opposite sides of the color wheel shown in Figure 10[29]. Otherwise, the pixel’s major color, which has the largest amplitude among R, G, and B components, is used to determine which mode it belongs to. For example, if the pixel’s major color is R component, we deem this case as R mode.

Figure 10
figure 10

Illustration of color wheel [[26]].

Complex mode

The complex mode is shown in Equation (15). According to color psychology analysis, the pixels which across blocking boundary have different fundamental colors get more attention [28]. Hence, the proposed algorithm determines this case is the complex mode. Figure 11 shows the locations of the pixels and “c” and “d” across blocking boundary. In the complex mode, the proposed filter gives different filters depending on human vision system (Equation 15). Because human vision system has more attention in the complex mode, the proposed filter only update two pixels, “c” and “d” across the blocking boundary in the complex mode.

Figure 11
figure 11

Locations of pixels “a”–“f” in the blocking boundary.

Since human beings have more sensibility on inverse colors than close colors based on color psychology, the number of updated pixels is less than other modes. In the complex mode, the deblocking algorithm is defined as

r c r d g c g d b c b d = 1 1 32 1 32 0 0 0 0 1 + 1 32 1 32 0 0 0 0 0 0 1 1 16 1 16 0 0 0 0 1 + 1 16 1 16 0 0 0 0 0 0 1 1 8 1 8 0 0 0 0 1 + 1 8 1 8 R c R d G c G d B c B d ,
(15)

where r c , r d , g c , g d , b c , and b d are updated pixels, and the suffixes, c and d, are the location of pixels. R, G, and B represent the pixel’s R, G and B values, respectively. For most parts, blocking artifacts are caused by block-based transform coding. Coding errors in block boundaries are larger than that in the interior of the block [23]. This property is a good indicator to design the filter. Together with the BS, different filters are applied to each pixel position for objective and subjective quality improvement through a heuristic rule.

R, G, and B modes

Except the complex mode, we design other three modes, R, G, and B modes, according to color analysis. In the R mode, the position of the pixels is depicted in Figure 12. At the same time, we also need considering that pixel’s value in the R, G, and B components. Because this area belongs to red color, the R component needs to be filtered with more pixels. If R c and R d are not both less than 128, the proposed algorithm will determine that is a light color and the filter of the R component is shown Equation (16).

r b r c r d r e = 1 1 4 1 4 0 0 1 2 1 2 0 0 1 2 1 2 0 0 1 4 1 4 1 R b R c R d R e ,
(16)
Figure 12
figure 12

Position of the pixels in R mode.

If R c and R d are both less than 128, the updated pixels are not only four pixels but also include the adjacent two pixels as

r a r f = 1 1 8 1 8 0 0 1 8 1 8 1 R a R c R d R f .
(17)

In the G component, the filter will update less number of pixels than the R component. If G c and G d are not both less than 128, the filter is defined as

g c g d = 1 2 1 2 1 2 1 2 G c G d .
(18)

If G c and G d are both less than 128, the updated pixels not only two pixels but also include the adjacent two pixels as

g b g e = 1 1 4 1 4 0 0 1 4 1 4 1 G b G c G d G e .
(19)

The number of updated pixels in the B component and that in the G component are the same. If B c and B d are not both less than 128, the filter is defined as

b c b d = 1 2 1 2 1 2 1 2 B c B d .
(20)

If B c and B d are both less than 128, the updated pixels not only two pixels include the adjacent two pixels as

b b b e = 1 1 4 1 4 0 0 1 4 1 4 1 B b B c B d B e .
(21)

The deblocking process of the G mode and B mode are very similar to R mode shown in Figures 13 and 14. In summary, the number of updated pixels of the certain component is more than other two components. Moreover, the proposed algorithm further designs the luminance filter to avoid over-smooth the image texture for enhancing subjective quality. Here, Y is the luminance component. The designed filter is shown in the following equation:

Y = y block + y 2 ,
(22)

where y block represents the luminance component of the image without deblocking filter, and y means that the luminance component of the image processed by the proposed deblocking filter. When the difference between y block and y is greater than the predefined threshold, the filtered values are recalculated.

Figure 13
figure 13

Position of the pixels in G mode.

Figure 14
figure 14

Position of the pixels in B mode.

Experimental results

Several H.264/AVC-coded sequences, including Foreman, HallMonitor, Mother&Daughter, and TableTennis with different bitrates, are used in our experiments. The parameters setting in the reference software JM 12.3 high profile are shown in Table 1.

Table 1 Parameters setting in JM 12.3

Subjective quality comparison

We compare the subjective quality of Tai’s [21] algorithm, Chen’s algorithm [22], and H.264/AVC loop filter [23] with that of the proposed algorithm. Tai’s algorithm is designed for blocking effects occurred at 8 × 8 bock boundaries, but blocking effects may occur at 4 × 4 block boundaries in H.264/AVC. Hence, this algorithm should be modified to be able to process 4 × 4 block boundaries. The mode decision step and the filter step on each mode of the algorithm are not modified in the simulation. Figures in this section show the original frame (a), non-filtered frame (b), filtered by Tai’s [21] 4 × 4 filter (c), filtered by Chen’s [22] filter (d), filtered by H.264/AVC loop filter [23] (e), and filtered by proposed filter (f). The filtered results are shown in Figures 15, 16, 17, and 18. Furthermore, the details of the blue region of Figures 15, 16, 17, and 18 are shown in Figures 19, 20, 21, and 22. As can be seen in Figure 19c,d Tai’s algorithm does not filter the chrominance and Chen’s method filters the chrominance, but the blocking effects still occur. In Foreman’s eyes, the proposed deblocking filter can filter blocking boundaries based on human vision system and color psychology. Our method determines that Foreman’s eyes are a complex area; therefore, this region is not over-smoothed. In Figure 20c,d Tai’s and Chen’s methods cannot differentiate between blocking artifacts and true edges in an image. Hence, they may blur the true edges and decrease the subjective quality. The proposed method has the better subjective quality than Tai’s and Chen’s methods. Similar results can be seen in Figures 17, 18, 21, and 22.

Figure 15
figure 15

Deblocking results of Foreman encoded by H.264 (CIF, Frame #1, target bit-rate: 105 kbps). (a) The original frame, (b) non-filtered frame, (c) filtered by Tai’s [18] 4 × 4 filter, (d) filtered by Chen’s [19] filter, (e) filtered by H.264/AVC loop filter [20], and (f) filtered by the proposed algorithm.

Figure 16
figure 16

Deblocking results of HallMonitor sequence encoded by H.264 (CIF, Frame #1, target bit-rate: 58 kbps). (a) The original frame, (b) non-filtered frame, (c) filtered by Tai’s [18] 4 × 4 filter, (d) filtered by Chen’s [19] filter, (e) filtered by H.264/AVC loop filter [20], and (f) filtered by the proposed algorithm.

Figure 17
figure 17

Deblocking results of Mother&Daughter encoded by H.264 (CIF, Frame #1, target bit-rate: 36 kbps). (a) The original frame, (b) non-filtered frame, (c) filtered by Tai’s [18] 4 × 4 filter, (d) filtered by Chen’s [19] filter, (e) filtered by H.264/AVC loop filter [20], and (f) filtered by the proposed algorithm.

Figure 18
figure 18

Deblocking results of TableTennis encoded by H.264 (CIF, Frame #1, target bit-rate: 140 kbps). (a) The original frame, (b) non-filtered frame, (c) filtered by Tai’s [18] 4 × 4 filter, (d) filtered by Chen’s [19] filter, (e) filtered by H.264/AVC loop filter [20], and (f) filtered by the proposed algorithm.

Figure 19
figure 19

Details of the blue region in Foreman (CIF, Frame #1, target bit-rate: 105 kbps). (a) The original frame, (b) non-filtered frame, (c) filtered by Tai’s [18] 4 × 4 filter, (d) filtered by Chen’s [19] filter, (e) filtered by H.264/AVC loop filter [20], and (f) filtered by the proposed algorithm.

Figure 20
figure 20

Details of the blue region in HallMonitor (CIF, Frame #1, target bit-rate: 58 kbps). (a) The original frame, (b) non-filtered frame, (c) filtered by Tai’s [18] 4 × 4 filter, (d) filtered by Chen’s [19] filter, (e) filtered by H.264/AVC loop filter [20], and (f) filtered by the proposed algorithm.

Figure 21
figure 21

Details of the blue region in Mother&Daughter (CIF, Frame #1, target bit-rate: 36 kbps). (a) The original frame, (b) non-filtered frame, (c) filtered by Tai’s [18] 4 × 4 filter, (d) filtered by Chen’s [19] filter, (e) filtered by H.264/AVC loop filter [20], and (f) filtered by the proposed algorithm.

Figure 22
figure 22

Details of the blue region in TableTennis (CIF, Frame #1, target bit-rate: 140 kbps). (a) The original frame, (b) non-filtered frame, (c) filtered by Tai’s [18] 4 × 4 filter, (d) filtered by Chen’s [19] filter, (e) filtered by H.264/AVC loop filter [20], and (f) filtered by the proposed algorithm.

Objective quality comparison

Four measurements, PSNR Total , PSNR Y , the generalized block-edge impairment metric (GBIM) [30], and structural similarity index (SSIM) are used to evaluate objective quality. Here, PSNR Total is also used in objective quality comparison due to the design of the chrominance filter is considered in Chen’s method, loop filter [23], and the proposed algorithm. The function of PSNR Total is shown as

PSNR Total = 4 × PSNR Y + PSNR Cb + PSNR Cr 6 ,
(23)

In this equation, the weight of PSNR Y is 4, because the sampling pattern of our tested sequences is 4:2:0. A GBIM measures the horizontal and vertical differences, M h_GBIM and M v_GBIM , between the columns and rows at all block boundaries of the reconstructed frame. GBIM is done by averaging M h_GBIM and M v_GBIM ,

GBIM = M hG BIM + M vG BIM 2 .
(24)

The major advantage of GBIM is that it calculates the blockiness of a frame without the original frame. The purpose of this measurement is to detail the state of frame blockiness. The higher the GBIM value is, the greater the severity of the blocking effects become. We also use SSIM as one of objective quality assessment metrics. SSIM is based on the study by Wang et al. [31]. The main difference of SSIM compared to PSNR is that it does not estimate perceived errors to quantify image degradations but considers them as perceived changes in structural information variation. To evaluate the deblocking results, considering all PSNR, GBIM and SSIM seems better than only one of them.

Four different sequences are used to observe their objective qualities, including Foreman, HallMonitor, Mother&Daughter, and TableTennis with 100 frames in different bitrates: 105 k and 84 k, 58 k and 45 k, 36 k and 27 k, 140 k and 108 k, respectively. The results of PSNR Total and GBIM are shown in Tables 2, 3, and 4. The larger PSNR Total value means the better image quality, but the GBIM value is as smaller as better. In Tables 2, 3, 4, and 5, the PSNR Total of the proposed algorithm is higher than Tai’s algorithm by 0.03 to 0.3 dB. Because Tai’s algorithm does not filter the chrominance component, there are still some blocking artifacts in the Cb and Cr components. The PSNR Total of the proposed algorithm is higher than Chen’s algorithm by 0.01 to 0.03 dB. Because Chen’s algorithm may filter true edges at block boundaries, which blurs true edges and decreases the visual quality. The proposed algorithm’s PSNR Total is lower than that of H.264/AVC loop filter. The reason is that the H.264/AVC loop filter uses the system information of the codec, so it can find out blocking effects more accurately than post-processing-based method. For GBIM results, the proposed algorithm is better than other algorithms. Especially, the GBIM of the proposed scheme is much smaller than that of H.264/AVC loop filter, Chen’s filter and Tai’s filter. This is because the proposed algorithm filtered with luminance and chrominance components, and we also refine Y component to avoid over-smoothing original texture. In summary, the proposed method has better performance in terms of subjective and objective qualities when compared to other well-known methods, Chen’s filter and Tai’s filter. The proposed method has very close objective quality to H.264/AVC loop filter but the better GBIM. For SSIM and PSNR Y comparisons, the H.264/AVC loop filter still has the best performance among all algorithms because it is in-loop filter and has complete coding information. In SSIM, other post-processing-based algorithms’ performances are very similar. In PSNR Y , all the outcomes of our proposed method are better than that of non-filter while some outcomes of Tai’s and Chen’s are worse than that of non-filter.

Table 2 Objective quality comparison for “ Foreman
Table 3 Objective quality comparison for “ HallMonitor
Table 4 Objective quality comparison for “ Mother&Daughter
Table 5 Objective quality comparison for “ TableTennis

Evaluation of the proposed "Sobel-Wavelet" edge detector

Table 6 shows the efficiency of the proposed Sobel–Wavelet detector. The test frame is the first frame of Foreman sequence with the CIF resolution at 105 k bitrate. The ground truth of real edge and artifact for the test frame is determined by human directly. The detection results of wavelet-based edge detector for real edge and blocking artifact are 97.18 and 2.82%, respectively. Wavelet-based edge detector is robust for blocking artifacts to reveal real edge. Sobel edge detector equally detects real edge and blocking artifact. The proposed Sobel–Wavelet detector can clearly reveal blocking artifacts and detection accuracy reaches up to 99.1%.

Table 6 Efficiency of proposed Sobel–Wavelet detector

Computational complexity comparison

The comparisons of the computational complexity and the processing time are listed in Tables 7 and 8. The computational operators are addition, shift, comparison, and division. Table 8 shows the processing time comparisons of H.264/AVC encoded CIF Foreman sequence at 105 k. Experimental results show that our proposed algorithm has the similar complexity to the Chen’s filter [22], but higher than the Tai’s 4 × 4 filter [21]. The proposed algorithm filters in the RGB domain and Chen’s method filters in both Y and CbCr components, but the Tai’s 4 × 4 filter [21] just aims at Y component. For this reason, the computational complexity is a little higher than the Tai’s 4 × 4 filter [21], similar to Chen’s filter [22], and lower than the H.264/AVC loop filter [23].

Table 7 Computational complexity comparison
Table 8 Processing time comparison

Conclusions

This article proposes a deblocking algorithm to remove blocking artifacts for enhancement of the subjective and objective qualities of H.264/AVC encoded videos. The proposed detection algorithm differentiates between true edges and blocking artifacts to avoid over-smoothing the original texture in an image. Furthermore, the proposed deblocking filter determines the filtering stage according to color psychology analysis and gives different filters based on human vision system. As for objective quality, the proposed algorithm has higher PSNR than Tai’s and Chen’s methods; though the PSNR it is a little bit lower H.264/AVC loop filter, it has better GBIM. In subjective quality, the proposed algorithm is better than Tai’s and Chen’s methods and is close to H.264/AVC loop filter. Experimental results demonstrate the effectiveness of our proposal method. To make the proposed method more feasible, the computational complexity reduction is the major direction of our future work.

References

  1. Ngan K-N, Yap C-W, Tan K-T: Video Coding for Wireless Communications. New Jersey: Prentice Hall; 2002.

    Google Scholar 

  2. Tekalp A-M: Digital Video Processing. New Jersey: Prentice Hall PTR; 1995.

    Google Scholar 

  3. Wang Y, Ostermann J, Zhang Y-Q: Video Processing and Communications. New Jersey: Prentice Hall; 2002.

    Google Scholar 

  4. Sun M-T, Reibman A-R: Compressed Video over Networks. New Work: Marcel Dekker; 2001.

    Google Scholar 

  5. ITU-T Recommendation H.264/ISO/IEC 11496–10, Advance Video Coding, Final Committee Draft, Document JVT-F100. 2002.

  6. Jarske T, Haacisto P, Defee I: Post-filtering method for reducing blocking effects from coded images. IEEE Trans Consum Electron 1994, 40(8):521-526.

    Article  Google Scholar 

  7. Hsu Y-F, Chen Y-C: A new adaptive separable median filter for removing blocking effects. IEEE Trans Consum Electron 1993, 39(3):510-513. 10.1109/30.234628

    Article  Google Scholar 

  8. Jeong Y, Kim I, Lee S: On the POCS-based postprocessing technique to reduce blocking artifacts in transform coded images. IEEE Trans. Circuits Syst. Video Technol. 2000, 10(6):617-623.

    Article  Google Scholar 

  9. Zakhor A: Iterative procedures for reduction of blocking effects in transform image coding. IEEE Trans. Circuits Syst. Video Technol. 1992, 2(3):91-95.

    Article  Google Scholar 

  10. Luo J, Chen C-W, Parker K-J, Huang T-S: Artifact reduction in low bit rate DCT-based image compression. IEEE Trans Image Process 1996, 5(9):1363-1368. 10.1109/83.535848

    Article  Google Scholar 

  11. Trussell H-J, Civanlar M-R: Signal deconvolution by projection onto convex sets. In Proceedings of IEEE International Conference on Acoustics, Speech, and Signal Processing, vol. 9, part 1. California, USA: San Diego; 1984:496-499.

    Google Scholar 

  12. Zhang Y, Salari E: A novel post-processing method for reducing blocking artifacts in block-coded images. In Proceedings of IEEE International Conference on Electro/information Technology. MI, USA: East Lansing; 2006:548-551.

    Google Scholar 

  13. Kim J: Adaptive blocking artifact reduction using wavelet-based block analysis. IEEE Trans Consum Electron 2009, 55(2):933-940.

    Article  Google Scholar 

  14. Zhao W-B, Zhou Z-H: Fuzzy blocking artifacts reduction algorithm based on human visual system. In Proceedings of International Conference on Machine Learning and Cybernetics, vol. 3. Hong Kong; 2007:1626-1630.

    Google Scholar 

  15. Shi M, Yi Q, Gong J: Blocking effect reduction based on human visual system for highly compressed images. In Proceedings of Canadian Conference on Electrical and Computer Engineering. Ont, Canada: Ottawa; 2006:1948-1951.

    Google Scholar 

  16. Xiong Z, Orchard M-T, Zhang Y-Q: A deblocking algorithm for JPEG compressed images using overcomplete wavelet representations. IEEE Trans. Circuits Syst. Video Technol. 1997, 7(2):433-437. 10.1109/76.564123

    Article  Google Scholar 

  17. Triantafyllidis GA, Tzovaras D, Sampson D, Strintzis MG: Combined frequency and spatial domain algorithm for the removal of blocking artifacts. EURASIP J. Appl. Signal Process. 2002, 2002(6):601-612. 10.1155/S1110865702203133

    Article  MATH  Google Scholar 

  18. Yan WC, Chen YY: DCT-based image compression using wavelet-based algorithm with efficient deblocking filter. In Proceedings of 14th European Signal Processing Conference. Italy: Florence; 2006:489-494.

    Google Scholar 

  19. Smirnov S, Gotchev A, Egiazarian K: Methods for depth-map filtering in view-plus-depth 3D video representation. EURASIP J. Appl. Signal Process. 2012, 2012(25):1-21.

    Google Scholar 

  20. Tai S-C, Chen Y-Y, Shen S-F: Deblocking filter for low bit rate MPEG-4 video. IEEE Trans. Circuits Syst. Video Technol. 2005, 13(2):733-741.

    Google Scholar 

  21. Tai S-C, Chen Y-R, Chen C-Y, Chen Y-H: Low complexity deblocking method for DCT coded video signals. IEE Proc. Vis. Image Signal Process. 2006, 153(3):46-55.

    Article  Google Scholar 

  22. Yeh C-H, Ku T-F, Fan Jiang S-J, Chen M-J, Jhu J-A: Post-processing deblocking filter algorithm for various video decoders. IET Image Process (in press)

  23. List P, Joch A, Lainema J, Bjontegaard G, Karczewicz M: Adaptive deblocking filter. IEEE Trans. Circuits Syst. Video Technol. 2003, 13(7):614-619.

    Article  Google Scholar 

  24. Horowitz M, Joch A, Kossentini F, Hallapuro A: H.264/AVC baseline profile decoder complexity analysis. IEEE Trans. Circuits Syst. Video Technol. 2003, 13(7):715-727.

    Article  Google Scholar 

  25. Yeh C-H, Huang K-L, Chern S-J: Deblocking filter by color psychology analysis for H.264/AVC video coders. In Proceedings of IEEE International Conference on Intelligent Information Hiding and Multimedia Signal Processing, vol. 153. Heilongjiang, China: Harbin; 2008:802-805.

    Google Scholar 

  26. Haar A: Zur Theorie der orthogonalen Funktionensysteme. Math Ann 1911, 71(1):331-371.

    Article  MathSciNet  MATH  Google Scholar 

  27. Fairchid MD: Color Appearance Models. New York: John Wiley & Sons; 2004.

    Google Scholar 

  28. Hardin C-L, Maffi L: Color Categories in Thought and Language. CambridgeP: Cambridge University Press; 1997:163-193.

    Book  Google Scholar 

  29. Desktop Publishing. http://desktoppub.about.com/od/glossary/g/contrastingcolors.htm

  30. Wu H-R, Yuen M: A generalized block-edge impairment metric for video coding. IEEE Trans. Signal Process. Lett. 1997, 4(11):317-320.

    Article  Google Scholar 

  31. Wang Z, Bovik A-C, Sheikh H-R, Simonocelli E-P: Image quality assessment: from error visibility to structural similarity. IEEE Trans Image Process 2004, 13(4):600-612. 10.1109/TIP.2003.819861

    Article  Google Scholar 

Download references

Acknowledgements

This work was supported by the National Science Council, R.O.C, under the Grant NSC 99-2628-E-110-008-MY3.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Chia-Hung Yeh.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors’ original submitted files for images

Below are the links to the authors’ original submitted files for images.

Authors’ original file for figure 1

Authors’ original file for figure 2

Authors’ original file for figure 3

Authors’ original file for figure 4

Authors’ original file for figure 5

Authors’ original file for figure 6

Authors’ original file for figure 7

Authors’ original file for figure 8

Authors’ original file for figure 9

Authors’ original file for figure 10

Authors’ original file for figure 11

Authors’ original file for figure 12

Authors’ original file for figure 13

Authors’ original file for figure 14

Authors’ original file for figure 15

Authors’ original file for figure 16

Authors’ original file for figure 17

Authors’ original file for figure 18

Authors’ original file for figure 19

Authors’ original file for figure 20

Authors’ original file for figure 21

Authors’ original file for figure 22

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 2.0 International License (https://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and permissions

About this article

Cite this article

Yeh, CH., Tseng, WY. & Huang, KL. A deblocking algorithm based on color psychology for display quality enhancement. EURASIP J. Adv. Signal Process. 2012, 128 (2012). https://doi.org/10.1186/1687-6180-2012-128

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/1687-6180-2012-128

Keywords