跳到主要內容

臺灣博碩士論文加值系統

(44.201.97.138) 您好!臺灣時間:2024/09/20 16:30
字體大小: 字級放大   字級縮小   預設字形  
回查詢結果 :::

詳目顯示

: 
twitterline
研究生:LE VIET HUNG
研究生(外文):LE VIET HUNG
論文名稱:Deep Residual and Classified Neural Networks for Inverse Halftoning
論文名稱(外文):Deep Residual and Classified Neural Networks for Inverse Halftoning
指導教授:郭景明郭景明引用關係
指導教授(外文):Jing-Ming Guo
口試委員:鍾國亮楊傳凱謝君偉蘇順豐郭景明
口試委員(外文):Kuo-Liang ChungChuan-Kai YangJun-Wei HsiehShun-Feng SuJing-Ming Guo
口試日期:2019-07-31
學位類別:碩士
校院名稱:國立臺灣科技大學
系所名稱:電機工程系
學門:工程學門
學類:電資工程學類
論文種類:學術論文
論文出版年:2019
畢業學年度:107
語文別:英文
論文頁數:72
中文關鍵詞:Inverse halftoningHalftoningConvolutional neural networkResidual networkStatistical analysis
外文關鍵詞:Inverse halftoningHalftoningConvolutional neural networkResidual networkStatistical analysis
相關次數:
  • 被引用被引用:0
  • 點閱點閱:67
  • 評分評分:
  • 下載下載:0
  • 收藏至我的研究室書目清單書目收藏:0
In this thesis, an inverse halftoning method using Deep Residual and Classified Neural Networks (DRCNN) is proposed. The technique has deployed an effective deep learning method that leverages the Convolution Neural Network (CNN) as a non-linear mapping function to transform the halftone image into its continuous-tone version. The proposed DRCNN has sequent residual blocks and involves two main stages. Firstly, the CNN network extracts numerous intrinsic features of an image and approximately reconstructs the halftone patterns. In the next step, the residual learning technique is integrated to reconstruct the fine details and edge information with good accuracy.
From in-depth studies, it is been observed that the statistical properties of halftone patterns can significantly influence the training efficiency and performance. For instance, the model trained with the halftone pattern of homogenous distribution cannot perform ideally for the patterns with high structural information. Consequently, in the present work, the halftone patterns are initially classified in accordance with its variances and then used to train the corresponding neural networks. The proposed method comprises the integration of various deep neural networks which are trained over different statistical ranges and thus it can establish enhanced reconstruction rate with respect to the statistics of a halftone patch.
Comprehensive experimental results demonstrate that the proposed deep learning-based technique outperforms other inverse halftoning methods significantly and achieves a new state-of-the-art performance in terms of peak-signal-to-noise ratio (PSNR), Structural Similarity Index (SSIM) and perceptual quality.
In this thesis, an inverse halftoning method using Deep Residual and Classified Neural Networks (DRCNN) is proposed. The technique has deployed an effective deep learning method that leverages the Convolution Neural Network (CNN) as a non-linear mapping function to transform the halftone image into its continuous-tone version. The proposed DRCNN has sequent residual blocks and involves two main stages. Firstly, the CNN network extracts numerous intrinsic features of an image and approximately reconstructs the halftone patterns. In the next step, the residual learning technique is integrated to reconstruct the fine details and edge information with good accuracy.
From in-depth studies, it is been observed that the statistical properties of halftone patterns can significantly influence the training efficiency and performance. For instance, the model trained with the halftone pattern of homogenous distribution cannot perform ideally for the patterns with high structural information. Consequently, in the present work, the halftone patterns are initially classified in accordance with its variances and then used to train the corresponding neural networks. The proposed method comprises the integration of various deep neural networks which are trained over different statistical ranges and thus it can establish enhanced reconstruction rate with respect to the statistics of a halftone patch.
Comprehensive experimental results demonstrate that the proposed deep learning-based technique outperforms other inverse halftoning methods significantly and achieves a new state-of-the-art performance in terms of peak-signal-to-noise ratio (PSNR), Structural Similarity Index (SSIM) and perceptual quality.
ABSTRACT i
ACKNOWLEDGEMENTS ii
CONTENTS iii
List of Figures vi
List of Tables ix
List of Abbreviations x
CHAPTER 1 - INTRODUCTION 1
1.1 Motivation and Problem Statement 1
1.2 Proposed Solution 2
1.3 Organization of Thesis 2
CHAPTER 2 - Digital Halftoning and Inverse Halftoning 4
2.1 Halftoning 4
2.1.1 Ordered Dithering 6
2.1.2 Error Diffusion 7
2.1.3 Dot Diffusion 9
2.1.4 Direct Binary Search 11
2.1.5 Halftone Types Comparison 12
2.2 Inverse Halftoning 13
2.2.1 A Naïve Approach 14
2.2.2 Look-up Table Method 15
2.2.3 Wavelet-based Method (WInHD) 16
2.2.4 Multiscale gradient estimator (FastIT) 16
2.2.5 Deep learning-based Method 17
CHAPTER 3 - Neural Networks 19
3.1 Machine Learning 19
3.2 Supervised Learning 19
3.3 Artificial Neural Networks 20
3.4 Multi-layer networks 20
3.5 Convolutional Neural Networks 21
3.6 Convolutional layer 22
3.7 Pooling layer 23
3.8 Fully-connected layer 23
3.9 Residual Learning and Skip Connection 23
3.10 Generative Adversarial Network 24
3.11 Internal Covariate Shift 25
CHAPTER 4 - – Proposed DRCNN Method 26
4.1 Network Architecture 26
4.1.1 Generator Network 26
4.1.2 Residual Blocks 32
4.1.3 Depth of network 32
4.2 Loss Functions 33
4.3 Statistical Analysis 36
4.4 Image Quality Assessments 40
4.5 Multi-tones Color Images 41
CHAPTER 5 - Experiments 42
5.1 Experimental Setup 42
5.2 Datasets 42
5.3 Hyper-parameters 43
5.4 Investigation of loss functions 44
5.5 Investigation of perceptual loss at different convolutional layers 45
5.6 Performance on different variance models 46
5.7 Experimental Results 47
5.7 Future Works 52
CHAPTER 6 - Conclusion 55
BIBLIOGRAPHY 56
[1] R. Ulichney, Digital halftoning. MIT press, 1987.
[2] D. L. Lau and G. R. Arce, Modern digital halftoning. CRC Press, 2008.
[3] V. Ostromoukhov, “A simple and efficient error-diffusion algorithm,” in Proceedings of the 28th annual conference on Computer graphics and interactive techniques, 2001, pp. 567–572.
[4] D. E. Knuth, “Digital halftones by dot diffusion,” ACM Trans. Graph., vol. 6, no. 4, pp. 245–273, 1987.
[5] D. J. Lieberman and J. P. Allebach, “Efficient model based halftoning using direct binary search,” in Proceedings of International Conference on Image Processing, 1997, vol. 1, pp. 775–778.
[6] T. Silva, “An Intuitive Introduction to Generative Adversarial Networks - GAN.” .
[7] P.-C. Chang and C.-S. Yu, “Neural net classification and LMS reconstruction to halftone images,” in Visual Communications and Image Processing’98, 1998, vol. 3309, pp. 592–603.
[8] M. Mese and P. P. Vaidyanathan, “Look-up table (LUT) method for inverse halftoning,” IEEE Trans. Image Process., vol. 10, no. 10, pp. 1566–1578, 2001.
[9] K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” arXiv Prepr. arXiv1409.1556, 2014.
[10] J. M. Guo and S. Sankarasrinivasan, “Digital Halftone Database (DHD): A Comprehensive Analysis on Halftone Types,” in 2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), 2018, pp. 1091–1099.
[11] O. Gustavsson, “AM halftoning and FM halftoning.” .
[12] B. E. Bayer, “An optimum method for two-level rendition of continuous tone pictures,” in IEEE International Conference on Communications, June, 1973, 1973, vol. 26.
[13] R. W. Floyd and L. S. Steinberg, “An adaptive algorithm for spatial gray scale,” 1975.
[14] J. Jarvis, C. R.-I. T. on Communications, and undefined 1976, “A new technique for displaying continuous tone images on a bilevel display,” ieeexplore.ieee.org.
[15] P. Stucki, “A Multiple-error Correction Computation Algorithm for Bi-level Image Hardcopy Reproduction,” 1981.
[16] R. Neelamani, R. D. Nowak, and R. G. Baraniuk, “WInHD: Wavelet-based inverse halftoning via deconvolution,” IEEE Trans. Image Process., 2002.
[17] T. D. Kite, N. Damera-Venkata, B. L. Evans, and A. C. Bovik, “A fast, high-quality inverse halftoning algorithm for error diffused halftones,” IEEE Trans. Image Process., vol. 9, no. 9, pp. 1583–1592, 2000.
[18] M. Mese and P. P. Vaidyanathan, “Optimized halftoning using dot diffusion and methods for inverse halftoning,” IEEE Trans. Image Process., vol. 9, no. 4, pp. 691–709, Apr. 2000.
[19] Y. F. Liu, J. M. Guo, and J. Der Lee, “Inverse halftoning based on the bayesian theorem,” IEEE Trans. Image Process., vol. 20, no. 4, pp. 1077–1084, 2011.
[20] P. W. Wong, “Inverse halftoning and kernel estimation for error diffusion,” IEEE Trans. Image Process., vol. 4, no. 4, pp. 486–498, 1995.
[21] Y.-T. Kim, G. R. Arce, and N. Grabowski, “Inverse halftoning using binary permutation filters,” IEEE Trans. Image Process., vol. 4, no. 9, pp. 1296–1311, 1995.
[22] M. Mese and P. P. Vaidyanathan, “Recent advances in digital halftoning and inverse halftoning methods,” IEEE Trans. Circuits Syst. I Fundam. Theory Appl., vol. 49, no. 6, pp. 790–805, 2002.
[23] C.-H. Son and H. Choo, “Local learned dictionaries optimized to edge orientation for inverse halftoning,” IEEE Trans. Image Process., vol. 23, no. 6, pp. 2542–2556, 2014.
[24] J. Luo, R. De Queiroz, Z. F.-I. T. on Signal, and undefined 1998, “A robust technique for image descreening based on the wavelet transform,” ieeexplore.ieee.org.
[25] Zhang Xiaohua, Liu Fang, and Jiao LiCheng, “An effective image halftoning and inverse halftoning technique based on HVS,” in Proceedings Fifth International Conference on Computational Intelligence and Multimedia Applications. ICCIMA 2003, pp. 441–445.
[26] R. Girshick, J. Donahue, T. Darrell, and J. Malik, “Rich feature hierarchies for accurate object detection and semantic segmentation,” Nov. 2013.
[27] J. Wang, J. Yang, K. Yu, F. Lv, … T. H.-2010 I. computer, and undefined 2010, “Locality-constrained linear coding for image classification,” Citeseer.
[28] R. Haralick, K. S.-I. T. on systems, and undefined 1973, “Textural features for image classification,” ieeexplore.ieee.org.
[29] O. Stenroos, “Object detection from images using convolutional neural networks.” .
[30] P. Viola, M. J.-C. (1), and undefined 2001, “Rapid object detection using a boosted cascade of simple features,” researchgate.net.
[31] U. Schmidt and S. Roth, “Shrinkage Fields for Effective Image Restoration.”
[32] Y. Du, W. Wang, L. W. and pattern recognition, and undefined 2015, “Hierarchical recurrent neural network for skeleton based action recognition,” openaccess.thecvf.com.
[33] S. Ji, W. Xu, M. Yang, K. Y.-I. transactions on pattern, and undefined 2012, “3D convolutional neural networks for human action recognition,” ieeexplore.ieee.org.
[34] O. Ronneberger, P. Fischer, and T. Brox, “U-Net: Convolutional Networks for Biomedical Image Segmentation.”
[35] S. Xie and Z. Tu, “Holistically-nested edge detection,” in Proceedings of the IEEE international conference on computer vision, 2015, pp. 1395–1403.
[36] G. Larsson, M. Maire, and G. Shakhnarovich, “Learning representations for automatic colorization,” in European Conference on Computer Vision, 2016, pp. 577–593.
[37] S. Iizuka, E. Simo-Serra, and H. Ishikawa, “Let there be color!: joint end-to-end learning of global and local image priors for automatic image colorization with simultaneous classification,” ACM Trans. Graph., vol. 35, no. 4, p. 110, 2016.
[38] “Convolutional Neural Network.” .
[39] S. Y. Fei Fei Li Justin Johnson, “Pooling Layer - CS231n.” .
[40] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
[41] C. Ledig et al., “Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network.”
[42] S. Ioffe and C. Szegedy, “Batch Normalization: Accelerating Deep Network Training b y Reducing Internal Covariate Shift,” 2015.
[43] L. Zhang, Q. Wang, H. Lu, and Y. Zhao, “End-to-End Learning of Multi-scale Convolutional Neural Network for Stereo Matching,” arXiv Prepr. arXiv1906.10399, 2019.
[44] C. Szegedy, S. Ioffe, V. Vanhoucke, and A. A. Alemi, “Inception-v4, inception-resnet and the impact of residual connections on learning,” in Thirty-First AAAI Conference on Artificial Intelligence, 2017.
[45] M. Xia and T.-T. Wong, “Deep Inverse Halftoning via Progressively Residual Learning.”
[46] A. A. Rusu et al., “Progressive Neural Networks,” Jun. 2016.
[47] T. Gao, J. Du, L. Dai, C. L.- INTERSPEECH, and undefined 2016, “SNR-Based Progressive Learning of Deep Neural Network for Speech Enhancement.,” staff.ustc.edu.cn.
[48] B. Lim, S. Son, H. Kim, S. Nah, and K. M. Lee, “Enhanced Deep Residual Networks for Single Image Super-Resolution,” Jul. 2017.
[49] L. Gatys, A. S. Ecker, and M. Bethge, “Texture synthesis using convolutional neural networks,” in Advances in neural information processing systems, 2015, pp. 262–270.
[50] J. Bruna, P. Sprechmann, and Y. LeCun, “Super-resolution with deep convolutional sufficient statistics,” arXiv Prepr. arXiv1511.05666, 2015.
[51] J. Johnson, A. Alahi, and L. Fei-Fei, “Perceptual losses for real-time style transfer and super-resolution,” in European conference on computer vision, 2016, pp. 694–711.
[52] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” in Advances in neural information processing systems, 2012, pp. 1097–1105.
[53] H. Zhao, O. Gallo, I. Frosio, and J. Kautz, “Loss functions for image restoration with neural networks,” IEEE Trans. Comput. Imaging, vol. 3, no. 1, pp. 47–57, 2016.
[54] T.-Y. Lin et al., “Microsoft coco: Common objects in context,” in European conference on computer vision, 2014, pp. 740–755.
[55] M. E. et al., “Visual Object Classes Challenge 2012.” 2012.
[56] MIT, “Places365-Challenge.” .
[57] Z. Wang, A. C. Bovik, H. R. Sheikh, E. P. Simoncelli, and others, “Image quality assessment: from error visibility to structural similarity,” IEEE Trans. image Process., vol. 13, no. 4, pp. 600–612, 2004.
[58] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv Prepr. arXiv1412.6980, 2014.
[59] R. Neelamani, R. Nowak, and R. Baraniuk, “Model-based inverse halftoning with wavelet-vaguelette deconvolution,” in Proceedings 2000 International Conference on Image Processing (Cat. No. 00CH37101), 2000, vol. 3, pp. 973–976.
[60] Z. Xiong, M. T. Orchard, and K. Ramchandran, “Inverse halftoning using wavelets,” IEEE Trans. image Process., vol. 8, no. 10, pp. 1479–1483, 1999.
[61] P.-C. Chang, C.-S. Yu, and T.-H. Lee, “Hybrid LMS-MMSE inverse halftoning technique,” IEEE Trans. Image Process., vol. 10, no. 1, pp. 95–103, 2001.
[62] J.-M. Guo, Y.-F. Liu, J.-H. Chen, and J.-D. Lee, “Inverse Halftoning With Context Driven Prediction,” IEEE Trans. Image Process., vol. 23, no. 4, pp. 1923–1924, 2014.
[63] I. Goodfellow et al., “Generative adversarial nets,” in Advances in neural information processing systems, 2014, pp. 2672–2680.
連結至畢業學校之論文網頁點我開啟連結
註: 此連結為研究生畢業學校所提供,不一定有電子全文可供下載,若連結有誤,請點選上方之〝勘誤回報〞功能,我們會盡快修正,謝謝!
QRCODE
 
 
 
 
 
                                                                                                                                                                                                                                                                                                                                                                                                               
第一頁 上一頁 下一頁 最後一頁 top
無相關期刊