Skip to main content
Computational Intelligence and Neuroscience logoLink to Computational Intelligence and Neuroscience
. 2021 Jul 7;2021:9912839. doi: 10.1155/2021/9912839

VVC In-Loop Filtering Based on Deep Convolutional Neural Network

Soulef Bouaafia 1,, Seifeddine Messaoud 1, Randa Khemiri 1,2, Fatma Elzahra Sayadi 3
PMCID: PMC8282386  PMID: 34306057

Abstract

With the rapid advancement in many multimedia applications, such as video gaming, computer vision applications, and video streaming and surveillance, video quality remains an open challenge. Despite the existence of the standardized video quality as well as high definition (HD) and ultrahigh definition (UHD), enhancing the quality for the video compression standard will improve the video streaming resolution and satisfy end user's quality of service (QoS). Versatile video coding (VVC) is the latest video coding standard that achieves significant coding efficiency. VVC will help spread high-quality video services and emerging applications, such as high dynamic range (HDR), high frame rate (HFR), and omnidirectional 360-degree multimedia compared to its predecessor high efficiency video coding (HEVC). Given its valuable results, the emerging field of deep learning is attracting the attention of scientists and prompts them to solve many contributions. In this study, we investigate the deep learning efficiency to the new VVC standard in order to improve video quality. However, in this work, we propose a wide-activated squeeze-and-excitation deep convolutional neural network (WSE-DCNN) technique-based video quality enhancement for VVC. Thus, the VVC conventional in-loop filtering will be replaced by the suggested WSE-DCNN technique that is expected to eliminate the compression artifacts in order to improve visual quality. Numerical results demonstrate the efficacy of the proposed model achieving approximately −2.85%, −8.89%, and −10.05% BD-rate reduction of the luma (Y) and both chroma (U, V) components, respectively, under random access profile.

1. Introduction

With emerging technologies that have rapidly evolved, multimedia services and video applications have significantly increased. Therefore, higher resolution (4K and 8K), especially for video games, e-learning, video conferencing, and surveillance tasks, is required to meet end-users viewing quality specifications. A next generation video encoding, established by the Joint Video Experts Team (JVET) in July 2020 [1], was the successor of high efficiency video coding (HEVC) [2]; it is the versatile video coding (VVC), which was also called H.266. VVC achieves a BD-rate savings up to 30% at the same quality as HEVC, which is the best standard adopted to offer an appropriate level of performance for new multimedia services. Although VVC aims to keep high-quality compressed video with additional encoding features, it still inevitably suffers from compression artifacts, which can lead to a decrease in the video quality. Therefore, VVC's quality compressed video and images need to be improved. In this case, loop filters play a crucial role in video and image quality optimization before they are used for interprediction as reference images.

In the same way, as for HEVC, in order to remove video compression artifacts and improve reconstructed video quality, VVC standard adopts the loop filtering technique, including the deblocking filter (DBF), sample adaptive offset (SAO), and adaptive loop filter (ALF). The DBF is designed to eliminate artifacts along block borders using discontinuity-based smoothing filters [3, 4]. Then, SAO is the second filter applied after DBF in HEVC and VVC [5], for compensating the reconstructed samples with different offset values in order to remove ringing effects.

ALF is a modern VVC function that removes distortions between restored and original images that are the most current loop filters [6]. Although traditional in-loop filters can alleviate those artifacts, the dynamic distortion produced by video compression is hard to resolve. Deep learning progress is known to be a strong technology to overcome this task, by using the convolutional neural network (CNN) as the most versatile and effective computational method for images and videos detection and analysis [7].

In order to increase the video quality, many CNN filtering methods have been suggested for HEVC and VVC standards [812]. These existing methods are proposed to minimize visual artifacts and to achieve great efficiency through CNN-based in-loop filtering and postprocessing. For example, Jia et al. in [8] proposed a HEVC postprocessing residue-guided loop filter. A deep network based on progressive rethinking and collaborative learning mechanisms was developed by Wang et al. in [9] to enhance the quality of the reconstructed frame for intra and interprediction. Inspired by emerging technology challenges, as well as high speed rate and high video and image resolution quality, the original in-loop filtering has become inadequate to satisfy the services demanded by the end users. In this study, we propose a powerful deep CNN-based filtering technique, called the wide-activated squeeze-and-excitation deep convolutional neural network (WSE-DCNN). The proposed technique provides powerful new loop filtering using typical VVC standards (DBF, SAO, and ALF). The goal is to effectively eliminate compression artifacts and improve the reconstructed video quality and then meet the end-users services. The purpose of this article is to propose a WSE-DCNN technique-based quality enhancement and then to implement the scheme proposed in the VVC standard, which provides coding gains accordingly for the random access configuration.

The remainder of this study is organized as follows: Section 2 presents the related work overview. The proposed deep CNN-based in-loop filtering for VVC standard is defined in Section 3. Then, in Section 4, the proposed method is evaluated. Finally, Section 5 concludes the study.

2. Related Work Overview

In recent years, artificial intelligence has seen tremendous progress in computer vision topics, in particular in image and video compression [1315]. Deep learning networks have been applied to enhance coding tools for HEVC and VVC standards, including intra and interprediction, transformation, quantization, and loop filtering [16, 17]. With regards to the HEVC, Bouaafia et al. in [14] proposed a reduction of HEVC complexity based on machine learning in the process of interprediction, which saves a good performance in terms of RD cost and computational complexity. Furthermore, a fast CNN-based algorithm is developed by Yeh et al. in [18] to improve the efficiency of HEVC intracoding. Pan et al. in [19] suggested an improved ED-CNN-based in-loop filtering to replace HEVC DBF and SAO in order to remove artifacts. The results prove that the proposed algorithm achieves BD-rate savings of 6.45% and PSNR gains of 0.238 dB. A novel technique for DBF and SAO in HEVC intracoding was proposed based on the Variable-filter-size Residue learning convolutional neural network (VRCNN) [20]. The obtained results show that the suggested technique achieves 4.6% BD-rate savings.

In order to enhance loop filtering and postprocessing, Ma et al. in [10] have developed a new CNN model, known as MFRNet for the VVC standard. The proposed model was implemented into the VVC test model to alleviate visual errors and increase video quality. In addition, a dense residual convolutional neural network (DRN) for the VVC filtering method proposed was applied after DBF and before SAO and ALF [12]. The H.265/VVC fast-intra-CU coding technique is based on the improved DAG-SVM classifier to minimize CU partition complexity [21]. Achieved results reveal that the proposed method achieves a 54.74% time saving. Moreover, Park et al. in [22] proposed to use a lightweight neural network (LNN) for the fast decision algorithm to remove redundant VVC block partitioning.

The suggested model provides a compromise between the compression and encoding complexity. In this study, we propose a wide-activated squeeze-and-excitation deep CNN- (WSE-DCNN-) based in-loop filtering approach for VVC video quality enhancement and achieve coding gains.

3. Proposed Method

3.1. Proposed WSE-DCNN-Based In-Loop Filtering for VVC

The VVC standard [1] still employs the block-based hybrid video coding architecture used in all video compression standards, since H. 261. It includes intraframe prediction, interframe prediction, transformation, quantization, loop filtering (DBF, SAO, and ALF), and entropy coding. Figure 1 depicts the block diagram of a hybrid video encoder. The VVC architecture is made up of two processes, such as encoder and decoder processing. Each picture is split into block-shaped regions, with the exact block partitioning, called coding tree unit (CTU), which is the basic block partition of the HEVC and VVC standards. The first picture of a video sequence is coded using only intrapicture prediction. For all remaining pictures of a sequence or between random access points, interpicture temporally predictive coding modes are typically used for most blocks. The encoding process for interpicture prediction consists of choosing motion data comprising, the selected reference picture, and motion vector to be applied for predicting the samples of each block. The residual signal of the intra or interpicture prediction, which is the difference between the original block and its prediction, is transformed by a linear spatial transform. The transform coefficients are then scaled, quantized, entropy-coded, and transmitted together with the prediction information. The encoder duplicates the decoder processing loop, such that both will generate identical predictions for subsequent data. Therefore, the quantized transform coefficients are constructed by inverse scaling and are then inverse transformed to duplicate the decoded approximation of the residual signal. The residual is then added to the prediction, and the result of that addition may then be fed into the loop filters (including, DBF, SAO, and ALF) to smooth out artifacts induced by block-wise processing and quantization. The final picture representation (the output of the decoder) is stored in a decoded picture buffer to be used for the prediction of subsequent pictures.

Figure 1.

Figure 1

Proposed WSE-DCNN-based in-loop filtering in VVC standard.

In our study, the proposed WSE-CNN model replaces the original VVC loop filtering module (including, DBF, SAO, and ALF), as shown in Figure 1. The principal goal of this strategy is to improve the visual quality of the reconstructed frame while maintaining coding gains. The rate distortion optimization (RDO) technique is used to determine whether to apply to each coding unit (CU) the proposed WSE-DCNN in-loop filter. Equation (1) is given for the RDO metric.

J=D+λR, (1)

where the distortion between the original and the reconstructed frame is denoted by D, the coding bits needed represents by R and the Lagrange multiplier controlling the trade-off between D and R is λ. The coding tree unit (CTU) level on/off control is adopted to avoid a reduction in RDO performance. The frame-level filtering would be shut off to prevent oversignal, if the enhancement quality is not worth to cost the signaled bits. Specifically, the control flags at the CTU-level and frame-level are designed as follows. For each CTU, if the RD performance of the filtered CTU achieves better quality, the corresponding CTU control flag is enabled; otherwise, the flag is disabled. After all the CTUs in one frame are determined, the frame-level RD cost before and after filtering are calculated in equation (1) indicated by J1 and J2, respectively. If J1 > J2, the frame-level flag will be enabled. Hence, the corresponding frame-level flag can be encoded in the slice header and CTU-level control flags can be signaled into each corresponding CTU syntax. Otherwise, the frame-level flag is disabled and CTU-level flags will not be encoded for transmission anymore.

3.2. WSE-DCNN Architecture

Figure 2 shows the proposed framework. The suggested technique, divided into two chromas (U and V) and luma (Y), would filter out the three components simultaneously. The WSE-DCNN model proposed consists of six inputs; three are YUV reconstructed and the other three include the QP quantization parameter and the luma and chroma coding unit. These inputs are first normalized to provide better convergence in the training process and then fed to the proposed model. Hence, the three (Y/U/V) reconstructions are normalized to [0, 1] based on the highest bit depth value. This means that the normalized values (P′(x, y)) are achieved by the following equation.

Px,y=Px,y1B1,x=1,,W,y=1,,H, (2)

where B denotes the bit depth, P(x, y) is the normalized value in normalized Y/U/V at (x,y), and W and H are the width and the height of the reconstructed frame, respectively.

Figure 2.

Figure 2

WSE-DCNN structure.

Various quantization parameters (QPs) contribute to a variety of reconstructed video quality. This makes it easier to use a single set of parameters to fit reconstructions with different qualities. QP should be normalized to QPmap (3).

QPmapx,y=QP63,x=1,,W,y=1,,H. (3)

The CU partition of the luma (Y) and chroma (U,  V) components also represents the inputs. Since the blocking artifacts are mainly caused by CU block partition, the division information of CU is converted into coding unit maps (CUmaps) and normalized. For example, for each CU in each frame, the boundary position is filled with two and the other positions are filled with one. However, the normalization factor is two, and two CUmaps can be obtained, one as Y − CUmap and the other denoted by UV − CUmap.

The WSE-DCNN process has three levels, as shown in Figure 2. The three Y,  U,  V components are processed via WSE blocks at the first level, and each component is fused with its own CUmap. Moreover, before it is concatenated to feature maps, CUmap would be multiplied by its own channel. Since U' and V' size is just the half of Y, the above needs to be used for size alignment. In the second level, the feature maps of different channels are connected together and then processed by several WSE blocks. At this level, the QPmap is also concatenated. At the last level, in order to produce the output residual image, the three channels are processed separately again.The WSE is the principal module for the proposed WSE-DCNN-based in-loop filtering technique, as shown in Figure 3. Furthermore, the wide-activated convolution [23] and the squeeze-and-excitation (SE) operation [24] compose this simple module. The wide-activated convolution performs very well in super-resolution and noise reduction tasks. It composed of 3 × 3 wide convolution followed by the rectified linear unit (ReLU) [25] activation function and a convolution layer with kernel size 1 × 1. Next comes the SE operation, the most used operation to weigh each convolutional layer. It can use the complex relationship between different channels and generate a weighting factor for each channel.

Figure 3.

Figure 3

WSE module.

The WSE module includes the following steps as shown in Figure 3, given a feature map X with shape H × W × C, where C means the channel amounts. First, given Y1 and Y2 are the outputs of the wide-activated convolution, as shown in the following equations.

Y1=ReLUW1X+b1, (4)
Y2=W2Y1+b2. (5)

In the second step, each channel obtains a value according to the squeeze operation using global average pooling (GAP) Y3(k).

Y3k=1H×Wi=1Hj=1WY2i,j,k. (6)

The excitation operation is described by two fully connected layers followed by ReLU and sigmoid (σ) activation functions, respectively. Y4 is the first fully connected layer followed by ReLU, which is refined by a certain ratio r. Then, the second fully connected layer followed by the sigmoid activation function is denoted by Y5, and it gives each channel a smoothing gating ratio in the range of [0, 1].

Y4=ReLUW4Y3+b4,Y5=σW5Y4+b5. (7)

According to the WSE function, each Y2 channel is multiplied by the gating ratio.

Y6i,j,k=Y2i,j,k×Y5ki1,,H,j1,,W,k1,,C. (8)

Finally, when the number of input equals to the output channels C, a skip connection will be added directly from input to output to learn the residue. Otherwise, there is no skipped connection.

4. Results and Discussion

The efficiency of the proposed WSE-DCNN-based in-loop filtering scheme under VVC standards is assessed in this section. Then, a comparative performance with the existing approaches is introduced.

4.1. Training Settings

In this contribution, the public video dataset (BVI-DVC) is exploited to train the deep video compression techniques [26]. The BVI-DVC dataset contains 800 video sequences with different resolutions between 270p and 2160p. In this case, we choose 80% video sequences for the training process and 20% for the validation phase. These sequences are compressed under random access scenario by the VVC VTM-4.0 test model [27] with QP values (22, 27, 32, and 37). For each QP, the reconstruction video images, including luma and chroma components, and its corresponding ground truth are divided into 64×64 patches, which were selected in a random order.

The proposed deep learning framework is trained offline in a supervised learning manner. The deep framework used during the training phase is the TensorFlow-GPU [28]. In the experiments, the training parameters used are denoted by the following: the batch size is set to 128, the training epochs to 200, the learning rate is set to 0.001, and weight decay of 0.1 for every 50 epochs. To train the proposed deep model, we applied an optimizer, such as the Adam algorithm [29]. Intel®core TM i7-3770 @3.4 GHz CPU with 16 GB RAM and an NVIDIA GeForce RTX 2070 GPU are used as the training platforms.

To train the proposed WSE-DCNN model, we assume that the mean square error (MSE) [30] is applied as the loss function between the reconstructed and the ground truth image. The MSE loss function is defined in the following equation.

Lθ=1Ni=1NFYi,θXi22, (9)

Let Xi is the ground truth of the proposed model, where i ∈ {1,…, N}. F(.) is the output of the WSE-DCNN model, where Yi represents the compressed images, i ∈ {1,…, N}, and θ is the parameter set of the proposed framework.

The loss function evaluation is the way to judge whether the model is well trained or not. It indicates, as shown in Figure 4, that the model converged reasonably quickly by tending to zero the loss function. In addition, the loss (defined in equation (9)) value remains the same from epoch 100 onwards, which means that no training problem arose during the training process. This proves that model's weight is well tuned.

Figure 4.

Figure 4

Training analysis “Loss curve.”

The proposed WSE-DCNN technique is implemented in the VVC standard in order to replace the conventionally applied filtering system during the testing process. All experiments are evaluated using a random access configuration at four QP values (22, 27, 32, and 37) under the VVC JVET common test conditions (CTC) [31]. The RD performance analysis is performed based on Bjøntegaard-delta bitrate (BD-rate) [32]. The BD-rate represents the average bitrate saving calculated between two RD curves for the same video quality, where negative BD-rate values indicate actual bitrate saving and positive values indicate how much the bitrate is increased.

4.2. RD Performance Evaluation

Compared to the original VVC standard, Table 1 provides the RD performance results of the proposed technique. Columns Y, U, and V in the table show the BD-rate of Y, U, and V components, respectively.

Table 1.

VVC performance evaluation of the proposed model under random access profile.

Class Sequences BD-rate (%)
Y U V
Class A1 Tango2 −2.89 −10.02 −11.35
Campfire −1.22 −2.75 −10.28

Class A2 CatRobot1 −1.89 −10.76 −8.03
DaylightRoad2 −1.47 −12.36 −2.55

Class B Kimono2 −0.51 −8.13 −20.63
ParkScene −4.18 −9.25 −12.94
Cactus −2.36 −12.27 −9.70
BasketballDrive −2.53 −4.83 −7.82
BQTerrace 0.11 −2.88 0.63

Class C BasketballDrill −3.84 −7.01 −9.97
BQMall −3.89 −11.48 −10.92
PartyScene −4.65 −9.69 −9.63
RaceHorses −1.35 −10.70 −13.66

Class D BasketballPass −3.40 −8.21 −7.79
BQSquare −5.27 −4.39 −11.44
BlowingBubbles −4.15 −8.52 −5.19
RaceHorses −5.08 −18.04 −19.74

Overall −2.85 −8.89 −10.05

The proposed technique achieves better mean coding gains when integrated into VVC standard. It can achieve 2.85% BD-rate savings for luma Y component and 8.89% and 10.05% for both chroma U and V components under random access profile, as given in Table 1. The proposed system provides substantial efficiency of RD compression primarily for all test sequences in U and V chrominance. It is also apparent that, for some sequences, the compression performance varies widely, such that video sequence content affects the proposed model. In addition, the suggested model performs better in terms of RD performance for high motion or rich texture video sequences, such as Campfire, CatRobot1, Kimono2, RaceHorses, and BQSquare. Consequently, the suggested CNN-based loop filtering outperforms VVC with the conventional loop filtering algorithm in terms of RD performance.

PSNR is also used as a quality metric to test the performance of our proposed filtering technique integrated into the VVC standard, which is defined by the following equation [33].

PSNRYUV=6×PSNRY+PSNRU+PSNRV8. (10)

In order to show the subjective visual quality and to further verify the effectiveness of the suggested model, the RaceHorses video sequence for class D was encoded by QP 22 under random access profile. Figure 5 shows the visual quality comparison. It is obvious that frame details are blurry when compressed by the original VVC standard, but become clearer after being filtered by the proposed technique. In contrast to the regular VVC with/without conventional in-loop filtering, the proposed technique effectively removes all blocking artifact, such as ringing and blurring artifacts, which enhances video quality.

Figure 5.

Figure 5

Visual quality comparison (the 26th frame of RaceHorses with QP = 22: (a) Original; (b) VVC without in-loop filtering (PSNR=39.84 dB); (c) VVC (PSNR=39.96 dB); (d) VVC-based proposed model (PSNR=40.15 dB).

A comparative performance of the proposed approach was made with other CNN-based filtering methods, as given in Table 2. Based on VVC CTC, Table 2 provides the comparison of the encoding performance in terms of reducing RD performance with other approaches [12, 33]. In this work [12], Chen et al. proposed to improve reconstructed video quality through the in-loop filter of a dense residual convolutional neural network (DRN). This network is placed after DF and before SAO and ALF into VVC VTM-4.0 reference software, in which the DIV2K dataset [34] is used in the training phase. In addition, for both inter and intraimages, the CNN in-loop filter algorithm is proposed [33], which is implemented in VVC VTM-4.0 before ALFs with DBF and SAO are disabled.

Table 2.

Comparative RD performance with other approaches.

Class Approach [12] Approach [33] Proposed approach
BD-rate (%) Y U V Y U V Y U V
Class A1 −1.27 −3.38 −5.10 0.87 0.12 0.22 −2.05 −6.38 −10.81
Class A2 −2.21 −5.74 −2.88 −1.12 −0.52 −2.11 −1.68 −11.56 −5.29
Class B −1.13 −4.73 −4.55 −0.83 −0.47 −1.20 −1.89 −7.47 −10.09
Class C −1.39 −3.63 −4.36 −1.76 −3.64 −6.80 −3.43 −9.72 −11.05
Class D −1.39 −1.96 −3.08 −2.95 −3.27 −7.35 −4.47 −9.79 −11.04
Overall −1.47 −3.88 −3.99 −1.16 −1.56 −3.44 −2.70 −8.98 −9.65

Compared to other previous approaches, for all test sequences from class A1 to class D, the proposed WSE-DCNN system implemented in VVC better performed in terms of compression performance for both luma and chroma components, as given in Table 2. This means that in terms of objective and subjective visual quality, the model proposed works well. As results of the proposed technique, the effectiveness of the WSE-DCNN approach is shown in comparison to other approaches in almost all test sequences.

We presented a RD performance curves for the suggested model-based in-loop filtering, compared to other approaches with QPs values under random access scenario for class A1 to class D. The RD performance curves comparisons are given in Figure 6. Comparing the corresponding methods, we can see that the proposed filter model considerably enhances the VVC compression performance. The in-loop filtering suggested works well, in particular, in high-resolution video sequences, such as in class A1, class A2, and class B.

Figure 6.

Figure 6

Comparative RD performance curves. (a) Class A1 “Campfire @ 3840 × 2160.” (b) Class A2 “CatRobot @ 3840 × 2160.” (c) Class B “BasketballDrive @ 1920 × 1080.” (d) Class C “BasketballDrill @ 832 × 480.” (e) Class D “BasketballPass @ 416 × 240.”

5. Conclusion

In this article, we have introduced a deep learning technique to improve VVC video quality while enhancing the user's services. To alleviate the coding artifacts as well as ringing, blocking, and blurring, the proposed WSE-DCNN technique is integrated into VVC standard to replace the traditional in-loop filtering. Compared to original VVC filters, simulation results show that the proposed system offers best objective and subjective compression efficiency, with a BD-rate reduction of approximately −2.85%, −8.89%, and −10.05% for Y, U, and V components, respectively. The comparative results reveal that the proposed in-loop filtering framework proves its effectiveness in improving video quality. In future work, two deep learning algorithms will be developed, one to improve the VVC CU partition at interprediction in order to reduce VVC complexity reduction and the other to replace original filters to enhance visual quality.

Data Availability

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

References

  • 1.Bross B., Chen J., Liu S. Versatile video coding (Draft 4) JVET-M1001. Proceedings of the 13th Meeting of the Joint Video Exploration Team (JVET); October 2019; Marrakech, Morrocco. pp. 9–18. [Google Scholar]
  • 2.Khemiri R., Kibeya H., Sayadi F. E., Bahri N., Atri M., Masmoudi N. Optimisation of HEVC motion estimation exploiting SAD and SSD GPU-based implementation. IET Image Processing. 2017;12(2):243–253. [Google Scholar]
  • 3.Ichigaya A., Iwamura S., Nemoto S. Syntax and semantics changes of luma adaptive deblocking filter. Proceedings of the Joint Video Exploration Team (JVET); June 2018; Macao, China. ITU-T; [Google Scholar]
  • 4.Kotra Meher A., Esenlik S., Wang B., Gao H., Alshina E. Non-CE5: chroma QP derivation fix for deblocking filter. Proceedings of the Joint Video Exploration Team (JVET); October 2019; Geneva, Switzerland. [Google Scholar]
  • 5.Browne A., Sharman K., Keating S. SAO modification for 12-bit. Proceedings of the Joint Video Exploration Team (JVET); October 2020; Brussels, Belgium. [Google Scholar]
  • 6.Hu N., Seregin V., Karczewicz M. Non-CE5: spec fix for ALF filter and transpose index calculation. Proceedings of the Joint Video Exploration Team (JVET); October 2019; Geneva, Switzerland. [Google Scholar]
  • 7.Krizhevsky A., Sutskever I., Hinton G. E. ImageNet classification with deep convolutional neural networks. Communications of the ACM. 2017;60(6):84–90. doi: 10.1145/3065386. [DOI] [Google Scholar]
  • 8.Jia W., Li L., Li Z., Liu S. Residue guided loop filter for hevc post processing. 2019. http://arxiv.org/abs/1907.12681.
  • 9.Wang D., Xia S., Yang W., Liu J. Combining progressive rethinking and collaborative learning: a deep framework for in-loop filtering. 2020. http://arxiv.org/abs/2001.05651. [DOI] [PubMed]
  • 10.Ma D., Zhang F., Bull D. MFRNet: a new CNN architecture for post-processing and in-loop filtering. IEEE Journal of Selected Topics in Signal Processing. 2020;14 [Google Scholar]
  • 11.Liu D., Li Y., Lin J., Li H., Wu F. Deep learning-based video coding. ACM Computing Surveys. 2020;53(1):1–35. doi: 10.1145/3368405. [DOI] [Google Scholar]
  • 12.Chen S., Chen Z., Wang Y., Liu S. In-loop filter with dense residual convolutional neural network for VVC. Proceedings of the 2020 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR); August 2020; Shenzhen, China. IEEE; pp. 149–152. [Google Scholar]
  • 13.Bouaafia S., Khemiri R., Sayadi F. E., Atri M., Liouane N. A deep CNN-lstm framework for fast video coding. Proceedings of the International Conference on Image and Signal Processing; 2020; Marrakech, Morocco. Springer; pp. 205–212. [DOI] [Google Scholar]
  • 14.Bouaafia S., Khemiri R., Sayadi F. E., Atri M. Fast CU partition-based machine learning approach for reducing HEVC complexity. Journal of Real-Time Image Processing. 2020;17(1):185–196. doi: 10.1007/s11554-019-00936-0. [DOI] [Google Scholar]
  • 15.Bouaafia S., Khemiri R., Sayadi F. E., Atri M. SVM-based inter prediction mode decision for HEVC. Proceedings of the 2020 17th International Multi-Conference on Systems, Signals & Devices (SSD); July 2020; Monastir, Tunisia. IEEE; pp. 12–16. [Google Scholar]
  • 16.Amna M., Imen W., Ezahra S. F., Mohamed A. Fast intra-coding unit partition decision in H.266/FVC based on deep learning. Journal of Real-Time Image Processing. 2020;17(6):1971–1981. doi: 10.1007/s11554-020-00998-5. [DOI] [Google Scholar]
  • 17.Bouaafia S., Khemiri R., Maraoui A., Sayadi F. E. CNN-LSTM learning approach-based complexity reduction for high-efficiency video coding standard. Scientific Programming. 2021;2021:10. doi: 10.1155/2021/6628041.6628041 [DOI] [Google Scholar]
  • 18.Yeh C.-H., Zhang Z.-T., Chen M.-J., Lin C.-Y. HEVC intra frame coding based on convolutional neural network. IEEE Access. 2018;6:50087–50095. doi: 10.1109/access.2018.2867342. [DOI] [Google Scholar]
  • 19.Pan Z., Yi X., Zhang Y., Jeon B., Kwong S. Efficient in-loop filtering based on enhanced deep convolutional neural networks for HEVC. IEEE Transactions on Image Processing. 2020;29:5352–5366. doi: 10.1109/tip.2020.2982534. [DOI] [PubMed] [Google Scholar]
  • 20.Dai Y., Liu D., Wu F. A convolutional neural network approach for post-processing in HEVC intra coding. Proceedings of the International Conference on Multimedia Modeling; January 2017; Daejeon, South Korea. Springer; pp. 28–39. [DOI] [Google Scholar]
  • 21.Zhang Q., Wang Y., Huang L., Jiang B., Wang X. Fast CU partition decision for H.266/VVC based on the improved DAG-SVM classifier model. Multimedia Systems. 2020;27(1) doi: 10.1007/s00530-020-00688-z. [DOI] [Google Scholar]
  • 22.Park S.-h., Kang J. Fast multi-type tree partitioning for versatile video coding using a lightweight neural network. IEEE Transactions on Multimedia. 2020;175 [Google Scholar]
  • 23.Yu J., Fan Y., Yang J., et al. Wide activation for efficient and accurate image super-resolution. 2018. http://arxiv.org/abs/1808.08718.
  • 24.Hu J., Shen L., Sun G. Squeeze-and-excitation networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; June 2018; Seattle, WA, USA. pp. 7132–7141. [Google Scholar]
  • 25. V. Nair and G. E. Hinton, “Rectified linear units improve restricted Boltzmann machines,” In Icml, Ed., 2010.
  • 26.Ma D., Zhang F., Bull D. R. BVI-DVC: a training database for deep video compression. 2020. http://arxiv.org/abs/2003.13552.
  • 27. VTM 4.0 Software, Available at: https://vcgit.hhi.fraunhofer.de/jvet/VVCSoftware_VTM/-/tree/VTM-4.0.
  • 28.Abadi M., Agarwal A., Barham P., et al. Large-scale machine learning on heterogeneous distributed systems. 2016. http://arxiv.org/abs/1603.04467.
  • 29.Kingma D. P., Ba J. Adam: a method for stochastic optimization. 2016. p. p. 12. http://arxiv.org/abs/014126980.
  • 30.Messaoud S., Bradai A., Ahmed O. B., Quang P., Atri M., Hossain M. S. Deep federated Q-learning-based network slicing for industrial IoT. IEEE Transactions on Industrial Informatics. 2020;17 [Google Scholar]
  • 31.Bssen F., Boyce J., Li X., Seregin V., Sühring K. JVET common test conditions and software reference configurations for SDR video. Proceedings of the 13th JVET Meeting; June 2019; Marrakesh, Morrocco. pp. 9–18. [Google Scholar]
  • 32. G. Bjontegaard, “Calculation of average PSNR differences between RD-curves,” VCEG-M33, 2001.
  • 33.Kawamura K., Kidani Y., Naito S. CE13-2.6/CE13-2.7: evaluation results of cnn based in-loop filtering. Proceedings of the 13th JVET Meeting; October 2019; Geneva, Switzerland. pp. 19–27. [Google Scholar]
  • 34. DIV2K, https://data.vision.ee.ethz.ch/cvl/DIV2K/

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The data used to support the findings of this study are available from the corresponding author upon request.


Articles from Computational Intelligence and Neuroscience are provided here courtesy of Wiley

RESOURCES