|
[1] Y.H. Ho, C.C. Chan, W.H. Peng, H.M. Hang, and M. Domański, “Anfic: Image compression using augmented normalizing flows,” IEEE Open Journal of Circuits and Systems, vol. 2, pp. 613–626, 2021. [2] Y.H. Ho, C.P. Chang, P.Y. Chen, A. Gnutti, and W.H. Peng, “Canfvc: Conditional augmented normalizing flows for video compression,” in European Conference on Computer Vision, 2022. [3] X. Wang, K. Yu, C. Dong, and C. C. Loy, “Recovering realistic texture in image superresolution by deep spatial feature transform,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 606–615. [4] “”5th challenge on learned image compression”,” URL http://compression.cc, 2022. [5] G. Lu, W. Ouyang, D. Xu, X. Zhang, C. Cai, and Z. Gao, “Dvc: An endtoend deep video compression framework,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 11 006–11 015. [6] G. Lu, X. Zhang, W. Ouyang, L. Chen, Z. Gao, and D. Xu, “An endtoend learning framework for video compression,” IEEE transactions on Pattern Analysis and Machine Intelligence, 2020. [7] J. Lin, D. Liu, H. Li, and F. Wu, “Mlvc: multiple frames prediction for learned video compression,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 3546–3554. 30 [8] O. Rippel, A. G. Anderson, K. Tatwawadi, S. Nair, C. Lytle, and L. Bourdev, “Elfvc: Efficient learned flexiblerate video coding,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), October 2021, pp. 14 479–14 488. [9] E. Agustsson, D. Minnen, N. Johnston, J. Balle, S. J. Hwang, and G. Toderici, “Scalespace flow for endtoend optimized video compression,” in Proceedings of the IEEE/ CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 8503–8512. [10] Z. Cheng, H. Sun, M. Takeuchi, and J. Katto, “Learning image and video compression through spatialtemporal energy compaction,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 10 071–10 080. [11] O. Rippel, S. Nair, C. Lew, S. Branson, A. G. Anderson, and L. Bourdev, “Learned video compression,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 3454–3463. [12] R. Yang, F. Mentzer, L. Van Gool, and R. Timofte, “Learning for video compression with recurrent autoencoder and recurrent probability model,” IEEE Journal of Selected Topics in Signal Processing, vol. 15, no. 2, pp. 388–401, 2020. [13] R. Yang, F. Mentzer, L. V. Gool, and R. Timofte, “Learning for video compression with hierarchical quality and recurrent enhancement,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 6628–6637. [14] A. Golinski, R. Pourreza, Y. Yang, G. Sautiere, and T. S. Cohen, “Feedback recurrent autoencoder for video compression,” in Proceedings of the Asian Conference on Computer Vision, 2020. [15] Z. Hu, Z. Chen, D. Xu, G. Lu, W. Ouyang, and S. Gu, “Improving deep video compression by resolutionadaptive flow coding,” in European Conference on Computer Vision. Springer, 2020, pp. 193–209. [16] H. Liu, M. Lu, Z. Ma, F. Wang, Z. Xie, X. Cao, and Y. Wang, “Neural video coding using multiscale motion compensation and spatiotemporal context model,” IEEE Transactions on Circuits and Systems for Video Technology, 2020. 31 [17] Z. Hu, G. Lu, and D. Xu, “Fvc: A new framework towards deep video compression in feature space,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 1502–1511. [18] T. Ladune, P. Philippe, W. Hamidouche, L. Zhang, and O. Déforges, “Optical flow and mode selection for learningbased video coding,” in 2020 IEEE 22nd International Workshop on Multimedia Signal Processing (MMSP). IEEE, 2020, pp. 1–6. [19] T. Ladune, P. Philippe, W. Hamidouche, L. Zhang, and O. Déforges, “Conditional coding for flexible learned video compression,” in Neural Compression: From Information Theory to Applications–Workshop@ ICLR 2021, 2021. [20] J. Li, B. Li, and Y. Lu, “Deep contextual video compression,” Advances in Neural Information Processing Systems, 2021. [21] G. Lu, C. Cai, X. Zhang, L. Chen, W. Ouyang, D. Xu, and Z. Gao, “Content adaptive and error propagation aware deep video compression,” in European Conference on Computer Vision. Springer, 2020, pp. 456–472. [22] J. Campos, S. Meierhans, A. Djelouah, and C. Schroers, “Content adaptive optimization for neural image compression,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2019. [23] N. Zou, H. Zhang, F. Cricri, H. R. Tavakoli, J. Lainema, M. Hannuksela, E. Aksu, and E. Rahtu, “L 2 c–learning to learn to compress,” in 2020 IEEE 22nd International Workshop on Multimedia Signal Processing (MMSP). IEEE, 2020, pp. 1–6. [24] G. Lu, C. Cai, X. Zhang, L. Chen, W. Ouyang, D. Xu, and Z. Gao, “Content adaptive and error propagation aware deep video compression,” in European Conference on Computer Vision. Springer, 2020, pp. 456–472. [25] T. van Rozendaal, J. Brehmer, Y. Zhang, R. Pourreza, and T. S. Cohen, “Instanceadaptive video compression: Improving neural codecs by training on the test set,” arXiv preprint arXiv:2111.10302, 2021. 32 [26] H. Zhang, F. Cricri, H. R. Tavakoli, M. Santamaria, Y.H. Lam, and M. M. Hannuksela, “Learn to overfit better: finding the important parameters for learned image compression,” in 2021 International Conference on Visual Communications and Image Processing (VCIP). IEEE, 2021, pp. 1–5. [27] F. Brand, K. Fischer, and A. Kaup, “Ratedistortion optimized learningbased image compression using an adaptive hierachical autoencoder with conditional hyperprior,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2021, pp. 1885–1889. [28] M. Song, J. Choi, and B. Han, “Variablerate deep image compression through spatiallyadaptive feature transform,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 2380–2389. [29] Z. Hu, G. Lu, J. Guo, S. Liu, W. Jiang, and D. Xu, “Coarsetofine deep video coding with hyperpriorguided mode prediction,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 5921–5930. [30] R. Feng, Z. Guo, Z. Zhang, and Z. Chen, “Versatile learned video compression,” arXiv preprint arXiv:2111.03386, 2021. [31] X. Sheng, J. Li, B. Li, L. Li, D. Liu, and Y. Lu, “Temporal context mining for learned video compression,” arXiv preprint arXiv:2111.13850, 2021. [32] J. Liu, S. Wang, W.C. Ma, M. Shah, R. Hu, P. Dhawan, and R. Urtasun, “Conditional entropy coding for efficient video compression,” in European Conference on Computer Vision. Springer, 2020, pp. 453–468. [33] X. Sheng, J. Li, B. Li, L. Li, D. Liu, and Y. Lu, “Temporal context mining for learned video compression,” arXiv preprint arXiv:2111.13850, 2021. [34] J. Ballé, V. Laparra, and E. Simoncelli, “Endtoend optimized image compression,” International Conference for Learning Representations, 2017. 33 [35] J. Ballé, D. Minnen, S. Singh, S. J. Hwang, and N. Johnston, “Variational image compression with a scale hyperprior,” in International Conference on Learning Representations, 2018. [36] D. Minnen, J. Ballé, and G. D. Toderici, “Joint autoregressive and hierarchical priors for learned image compression,” Advances in Neural Information Processing Systems, vol. 31, pp. 10 771–10 780, 2018. [37] H. Ma, D. Liu, N. Yan, H. Li, and F. Wu, “Endtoend optimized versatile image compression with waveletlike transform,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2020. [38] H. Ma, D. Liu, R. Xiong, and F. Wu, “iwave: Cnnbased waveletlike transform for image compression,” IEEE Transactions on Multimedia, vol. 22, no. 7, pp. 1667–1679, 2019. [39] Y. Choi, M. ElKhamy, and J. Lee, “Variable rate deep image compression with a conditional autoencoder,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 3146–3154. [40] Z. Cheng, H. Sun, M. Takeuchi, and J. Katto, “Learned image compression with discretized gaussian mixture likelihoods and attention modules,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 7939–7948. [41] F. Mentzer, G. D. Toderici, M. Tschannen, and E. Agustsson, “Highfidelity generative image compression,” Advances in Neural Information Processing Systems, vol. 33, pp. 11 913–11 924, 2020. [42] F. Mentzer, E. Agustsson, M. Tschannen, R. Timofte, and L. Van Gool, “Conditional probability models for deep image compression,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 4394–4402. [43] T. Chen, H. Liu, Z. Ma, Q. Shen, X. Cao, and Y. Wang, “Endtoend learnt image compression via nonlocal attention optimization and improved context modeling,” IEEE Transactions on Image Processing, vol. 30, pp. 3179–3191, 2021. 34 [44] C. Ma, Z. Wang, R. Liao, and Y. Ye, “A cross channel context model for latents in deep image compression,” arXiv preprint arXiv:2103.02884, 2021. [45] J. Ballé, N. Johnston, and D. Minnen, “Integer networks for data compression with latentvariable models,” in International Conference on Learning Representations, 2018. [46] M. Li, W. Zuo, S. Gu, J. You, and D. Zhang, “Learning contentweighted deep image compression,” IEEE transactions on pattern analysis and machine intelligence, vol. 43, no. 10, pp. 3446–3461, 2020. [47] M. Li, K. Ma, J. You, D. Zhang, and W. Zuo, “Efficient and effective contextbased convolutional entropy modeling for image compression,” IEEE Transactions on Image Processing, vol. 29, pp. 5900–5911, 2020. [48] R. Yang, F. Mentzer, L. Van Gool, and R. Timofte, “Learning for video compression with recurrent autoencoder and recurrent probability model,” IEEE Journal of Selected Topics in Signal Processing, vol. 15, no. 2, pp. 388–401, 2020. [49] R. Feng, Z. Guo, Z. Zhang, and Z. Chen, “Versatile learned video compression,” arXiv preprint arXiv:2111.03386, 2021. [50] D. Sun, X. Yang, M.Y. Liu, and J. Kautz, “Pwcnet: Cnns for optical flow using pyramid, warping, and cost volume,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 8934–8943. [51] T. Xue, B. Chen, J. Wu, D. Wei, and W. T. Freeman, “Video enhancement with taskoriented flow,” International Journal of Computer Vision, vol. 127, no. 8, pp. 1106–1125, 2019. [52] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” International Conference for Learning Representations, 2015. [53] A. Mercat, M. Viitanen, and J. Vanne, “Uvg dataset: 50/120fps 4k sequences for video codec analysis and development,” in Proceedings of the 11th ACM Multimedia Systems Conference, 2020, pp. 297–302. 35 [54] G. J. Sullivan, J.R. Ohm, W.J. Han, and T. Wiegand, “Overview of the high efficiency video coding (hevc) standard,” IEEE Transactions on circuits and systems for video technology, vol. 22, no. 12, pp. 1649–1668, 2012.
|