(JBE Vol. 23, No. 5, September 2018) (Special Paper) 23 5, (JBE Vol. 23, No. 5, September 2018) ISSN

Similar documents
09권오설_ok.hwp

(JBE Vol. 21, No. 1, January 2016) (Regular Paper) 21 1, (JBE Vol. 21, No. 1, January 2016) ISSN 228

(JBE Vol. 20, No. 5, September 2015) (Special Paper) 20 5, (JBE Vol. 20, No. 5, September 2015) ISS

(JBE Vol. 22, No. 2, March 2017) (Regular Paper) 22 2, (JBE Vol. 22, No. 2, March 2017) ISSN

(JBE Vol. 23, No. 2, March 2018) (Special Paper) 23 2, (JBE Vol. 23, No. 2, March 2018) ISSN

THE JOURNAL OF KOREAN INSTITUTE OF ELECTROMAGNETIC ENGINEERING AND SCIENCE Nov.; 26(11),

19_9_767.hwp

THE JOURNAL OF KOREAN INSTITUTE OF ELECTROMAGNETIC ENGINEERING AND SCIENCE. vol. 29, no. 10, Oct ,,. 0.5 %.., cm mm FR4 (ε r =4.4)

(JBE Vol. 23, No. 5, September 2018) (Regular Paper) 23 5, (JBE Vol. 23, No. 5, September 2018) ISSN

2 : (JEM) QTBT (Yong-Uk Yoon et al.: A Fast Decision Method of Quadtree plus Binary Tree (QTBT) Depth in JEM) (Special Paper) 22 5, (JBE Vol. 2

2 : (Seungsoo Lee et al.: Generating a Reflectance Image from a Low-Light Image Using Convolutional Neural Network) (Regular Paper) 24 4, (JBE

(JBE Vol. 23, No. 2, March 2018) (Special Paper) 23 2, (JBE Vol. 23, No. 2, March 2018) ISSN

High Resolution Disparity Map Generation Using TOF Depth Camera In this paper, we propose a high-resolution disparity map generation method using a lo

2 : 3 (Myeongah Cho et al.: Three-Dimensional Rotation Angle Preprocessing and Weighted Blending for Fast Panoramic Image Method) (Special Paper) 23 2

1 : 360 VR (Da-yoon Nam et al.: Color and Illumination Compensation Algorithm for 360 VR Panorama Image) (Special Paper) 24 1, (JBE Vol. 24, No

THE JOURNAL OF KOREAN INSTITUTE OF ELECTROMAGNETIC ENGINEERING AND SCIENCE Dec.; 27(12),

2 : (Juhyeok Mun et al.: Visual Object Tracking by Using Multiple Random Walkers) (Special Paper) 21 6, (JBE Vol. 21, No. 6, November 2016) ht

<30312DC1A4BAB8C5EBBDC5C7E0C1A4B9D7C1A4C3A52DC1A4BFB5C3B62E687770>

THE JOURNAL OF KOREAN INSTITUTE OF ELECTROMAGNETIC ENGINEERING AND SCIENCE Jun.; 27(6),

THE JOURNAL OF KOREAN INSTITUTE OF ELECTROMAGNETIC ENGINEERING AND SCIENCE. vol. 29, no. 6, Jun Rate). STAP(Space-Time Adaptive Processing)., -

(JBE Vol. 24, No. 2, March 2019) (Special Paper) 24 2, (JBE Vol. 24, No. 2, March 2019) ISSN

(JBE Vol. 23, No. 6, November 2018) (Special Paper) 23 6, (JBE Vol. 23, No. 6, November 2018) ISSN 2

V28.

3 : 3D (Seunggi Kim et. al.: 3D Depth Estimation by a Single Camera) (Regular Paper) 24 2, (JBE Vol. 24, No. 2, March 2019)

(JBE Vol. 23, No. 1, January 2018) (Special Paper) 23 1, (JBE Vol. 23, No. 1, January 2018) ISSN 2287-

(JBE Vol. 24, No. 4, July 2019) (Special Paper) 24 4, (JBE Vol. 24, No. 4, July 2019) ISSN

(JBE Vol. 23, No. 1, January 2018). (VR),. IT (Facebook) (Oculus) VR Gear IT [1].,.,,,,..,,.. ( ) 3,,..,,. [2].,,,.,,. HMD,. HMD,,. TV.....,,,,, 3 3,,

(JBE Vol. 7, No. 4, July 0)., [].,,. [4,5,6] [7,8,9]., (bilateral filter, BF) [4,5]. BF., BF,. (joint bilateral filter, JBF) [7,8]. JBF,., BF., JBF,.

(JBE Vol. 24, No. 1, January 2019) (Regular Paper) 24 1, (JBE Vol. 24, No. 1, January 2019) ISSN 2287

THE JOURNAL OF KOREAN INSTITUTE OF ELECTROMAGNETIC ENGINEERING AND SCIENCE Feb.; 29(2), IS

08김현휘_ok.hwp

07.045~051(D04_신상욱).fm

<4D F736F F D20B1E2C8B9BDC3B8AEC1EE2DC0E5C7F5>

<372DBCF6C1A42E687770>

THE JOURNAL OF KOREAN INSTITUTE OF ELECTROMAGNETIC ENGINEERING AND SCIENCE Mar.; 28(3),

(JBE Vol. 23, No. 6, November 2018) (Special Paper) 23 6, (JBE Vol. 23, No. 6, November 2018) ISSN 2

THE JOURNAL OF KOREAN INSTITUTE OF ELECTROMAGNETIC ENGINEERING AND SCIENCE Jun.; 27(6),

À±½Â¿í Ãâ·Â

2 : (Jaeyoung Kim et al.: A Statistical Approach for Improving the Embedding Capacity of Block Matching based Image Steganography) (Regular Paper) 22

DBPIA-NURIMEDIA

°í¼®ÁÖ Ãâ·Â

03-16-김용일.indd

THE JOURNAL OF KOREAN INSTITUTE OF ELECTROMAGNETIC ENGINEERING AND SCIENCE Jul.; 27(7),

4 : (Hyo-Jin Cho et al.: Audio High-Band Coding based on Autoencoder with Side Information) (Special Paper) 24 3, (JBE Vol. 24, No. 3, May 2019

DBPIA-NURIMEDIA


<313120C0AFC0FCC0DA5FBECBB0EDB8AEC1F2C0BB5FC0CCBFEBC7D15FB1E8C0BAC5C25FBCF6C1A42E687770>

µðÇÃÇ¥Áö±¤°í´Ü¸é

1 : UHD (Heekwang Kim et al.: Segment Scheduling Scheme for Efficient Bandwidth Utilization of UHD Contents Streaming in Wireless Environment) (Specia

02손예진_ok.hwp

07변성우_ok.hwp

04 김영규.hwp

THE JOURNAL OF KOREAN INSTITUTE OF ELECTROMAGNETIC ENGINEERING AND SCIENCE Sep.; 30(9),

1 : (Sunmin Lee et al.: Design and Implementation of Indoor Location Recognition System based on Fingerprint and Random Forest)., [1][2]. GPS(Global P

DBPIA-NURIMEDIA

3 : (Won Jang et al.: Musical Instrument Conversion based Music Ensemble Application Development for Smartphone) (Special Paper) 22 2, (JBE Vol

김기남_ATDC2016_160620_[키노트].key

(JBE Vol. 23, No. 4, July 2018) (Special Paper) 23 4, (JBE Vol. 23, No. 4, July 2018) ISSN

(JBE Vol. 20, No. 6, November 2015) (Regular Paper) 20 6, (JBE Vol. 20, No. 6, November 2015) ISSN

THE JOURNAL OF KOREAN INSTITUTE OF ELECTROMAGNETIC ENGINEERING AND SCIENCE Mar.; 25(3),

지속가능경영보고서도큐_전체

08년요람001~016

Ⅱ. Embedded GPU 모바일 프로세서의 발전방향은 저전력 고성능 컴퓨팅이다. 이 러한 목표를 달성하기 위해서 모바일 프로세서 기술은 멀티코 어 형태로 발전해 가고 있다. 예를 들어 NVIDIA의 최신 응용프 로세서인 Tegra3의 경우 쿼드코어 ARM Corte

THE JOURNAL OF KOREAN INSTITUTE OF ELECTROMAGNETIC ENGINEERING AND SCIENCE Jul.; 27(7),

THE JOURNAL OF KOREAN INSTITUTE OF ELECTROMAGNETIC ENGINEERING AND SCIENCE Mar.; 30(3),

04 최진규.hwp

04_이근원_21~27.hwp

DBPIA-NURIMEDIA

DBPIA-NURIMEDIA

THE JOURNAL OF KOREAN INSTITUTE OF ELECTROMAGNETIC ENGINEERING AND SCIENCE Oct.; 27(10),

<30345F D F FC0CCB5BFC8F15FB5B5B7CEC5CDB3CEC0C720B0BBB1B8BACE20B0E6B0FCBCB3B0E8B0A120C5CDB3CE20B3BBBACEC1B6B8ED2E687770>

인문사회과학기술융합학회

(JBE Vol. 24, No. 2, March 2019) (Regular Paper) 24 2, (JBE Vol. 24, No. 2, March 2019) ISSN

<353420B1C7B9CCB6F52DC1F5B0ADC7F6BDC7C0BB20C0CCBFEBC7D120BEC6B5BFB1B3C0B0C7C1B7CEB1D7B7A52E687770>

디지털포렌식학회 논문양식

THE JOURNAL OF KOREAN INSTITUTE OF ELECTROMAGNETIC ENGINEERING AND SCIENCE Nov.; 28(11),

3 : ATSC 3.0 (Jeongchang Kim et al.: Study on Synchronization Using Bootstrap Signals for ATSC 3.0 Systems) (Special Paper) 21 6, (JBE Vol. 21

#유한표지F

(JBE Vol. 23, No. 5, September 2018) (Regular Paper) 23 5, (JBE Vol. 23, No. 5, September 2018) ISSN

(JBE Vol. 22, No. 5, September 2017) (Special Paper) 22 5, (JBE Vol. 22, No. 5, September 2017) ISSN

4 : WebRTC P2P DASH (Ju Ho Seo et al.: A transport-history-based peer selection algorithm for P2P-assisted DASH systems based on WebRTC) (Special Pape

<333820B1E8C8AFBFEB2D5A B8A620C0CCBFEBC7D120BDC7BFDC20C0A7C4A1C3DFC1A42E687770>

1 : HEVC Rough Mode Decision (Ji Hun Jang et al.: Down Sampling for Fast Rough Mode Decision for a Hardware-based HEVC Intra-frame encoder) (Special P

8-VSB (Vestigial Sideband Modulation)., (Carrier Phase Offset, CPO) (Timing Frequency Offset),. VSB, 8-PAM(pulse amplitude modulation,, ) DC 1.25V, [2

09김수현_ok.hwp

(JBE Vol. 23, No. 1, January 2018) (Regular Paper) 23 1, (JBE Vol. 23, No. 1, January 2018) ISSN 2287

04김호걸(39~50)ok

???? 1

3 : OpenCL Embedded GPU (Seung Heon Kang et al. : Parallelization of Feature Detection and Panorama Image Generation using OpenCL and Embedded GPU). e

14.531~539(08-037).fm

03이승호_ok.hwp

1 : (Su-Min Hong et al.: Depth Upsampling Method Using Total Generalized Variation) (Regular Paper) 21 6, (JBE Vol. 21, No. 6, November 2016)

,. 3D 2D 3D. 3D. 3D.. 3D 90. Ross. Ross [1]. T. Okino MTD(modified time difference) [2], Y. Matsumoto (motion parallax) [3]. [4], [5,6,7,8] D/3

(JBE Vol. 23, No. 5, September 2018) (Regular Paper) 23 5, (JBE Vol. 23, No. 5, September 2018) ISSN

THE JOURNAL OF KOREAN INSTITUTE OF ELECTROMAGNETIC ENGINEERING AND SCIENCE Jan.; 26(1),

1. 서 론

(JBE Vol. 23, No. 6, November 2018) (Regular Paper) 23 6, (JBE Vol. 23, No. 6, November 2018) ISSN 2

¿ì¾ç-ÃÖÁ¾

THE JOURNAL OF KOREAN INSTITUTE OF ELECTROMAGNETIC ENGINEERING AND SCIENCE Dec.; 26(12),

45-51 ¹Ú¼ø¸¸

Transcription:

(JBE Vol. 23, No. 5, September 2018) (Special Paper) 23 5, 2018 9 (JBE Vol. 23, No. 5, September 2018) https://doi.org/10.5909/jbe.2018.23.5.614 ISSN 2287-9137 (Online) ISSN 1226-7953 (Print) Generative Adversarial Network a), a), a) Depth Image Restoration Using Generative Adversarial Network John Junyeop Nah a), Chang Hun Sim a), and In Kyu Park a) generative adversarial network (GAN). 3D morphable model convolutional neural network (3DMM CNN) largescale CelebFaces Attribute (CelebA) FaceWarehouse deep convolutional GAN (DCGAN) (generator) Wasserstein distance (discriminator).. Abstract This paper proposes a method of restoring corrupted depth image captured by depth camera through unsupervised learning using generative adversarial network (GAN). The proposed method generates restored face depth images using 3D morphable model convolutional neural network (3DMM CNN) with large-scale CelebFaces Attribute (CelebA) and FaceWarehouse dataset for training deep convolutional generative adversarial network (DCGAN). The generator and discriminator equip with Wasserstein distance for loss function by utilizing minimax game. Then the DCGAN restore the loss of captured facial depth images by performing another learning procedure using trained generator and new loss function. Keyword : Deep learning, generative adversarial network, depth image, depth camera, restoration a) (Inha University, Department of Information and Communication Engineering) Corresponding Author : (In Kyu Park) E-mail: pik@inha.ac.kr Tel: +82-32-860-9190 ORCID: https://orcid.org/0000-0003-4774-7841 IPIU 2018. 2016 ( ) (NRF-2016R1A2B4014731). This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIP) (No. NRF-2016R1A2B4014731). Manuscript received May 8, 2018; Revised July 13, 2018; Accepted July 16, 2018. Copyright 2016 Korean Institute of Broadcast and Media Engineers. All rights reserved. This is an Open-Access article distributed under the terms of the Creative Commons BY-NC-ND (http://creativecommons.org/licenses/by-nc-nd/3.0) which permits unrestricted non-commercial use, distribution, and reproduction in any medium, provided the original work is properly cited and not altered.

2 : Generative Adversarial Network (John Junyeop Nah et al.: Depth Image Restoration Using Generative Adversarial Network). Time-of-flight (ToF).,.. [1][2]. RGB (inpainting) [3].. DCGAN [4] RGB [5]. DCGAN Jensen-Shannon divergence.. DCGAN [4]. DCGAN Jensen-Shannon divergence (weak) Wasserstein distance [6]. DCGAN GAN [7] convolutional neural network (CNN) [4]. DCGAN 3DMM CNN [8] CelebA FaceWarehouse [9] 23,500... RGB DCGAN Wasserstein distance,. 2. 3 DCGAN, 4. 5.. 1. 3DMM CNN CNN [8] (principal component analysis, PCA) 3 Basel face model (BFM) [10] 3D. CelebA 15,000 3D FaceWarehouse [9] 8,500. 23,500 3D 1 (-1, 1).

(JBE Vol. 23, No. 5, September 2018) (a) (b) 1.. (a) 3DMM CNN CelebA, (b) FaceWarehouse Fig. 1. Depth image used in minimax game. (a) Extracted depth image by using 3DMM CNN and CelebA dataset, (b) Depth image from FaceWarehouse dataset 2. 20cm 150 cm Intel Real- Sense SR300 [11]. 2.. 2. Fig. 2. Facial depth image captured by depth camera. DCGAN 1. -1 1, 23,500., [7]. DCGAN Jensen-Shannon divergence (1) Wasserstein distance [6] 4 (b). 3. Fig. 3. Training procedure using minimax game

2 : Generative Adversarial Network (John Junyeop Nah et al.: Depth Image Restoration Using Generative Adversarial Network) (a) (b) 4.. (a) DCGAN, (b) Wasserstein distance DCGAN (: 0.0002, : 10,000) Fig. 4. Comparison of generated facial depth image after performing minimax game using two different loss functions. (a) Original DCGAN, (b) DCGAN with Wasserstein distance (Learning rate: 0.0002, iteration count: 10,000) P P P, P (joint probability distribution). (lower bound). Wasserstein distance (metric) [6]. 2. 0 1. [5]. [3]. L L log (2)., -1 1,. (3) GAN [7].. (2) (3). L L L (4) 5. Fig. 5. Depth image restoration using trained generator

(JBE Vol. 23, No. 5, September 2018) (5). L (6).. TensorFlow NVIDIA GeForce GTX 1080 Ti GPU 5 DCGAN (hyper parameter) 1. 6 4 (b) 25,000.. 7 1. Table 1. Hyper parameters used in minimax game and restoration Hyper Parameters Values Iteration Number 25,000 Batch Size 64 Regularization Weight ( ) 0.1 Learning Rate Optimization Function 0.0001, 0.00012, 0.00015, 0.00017, 0.0002 Adam optimizer 7.,, Fig. 7. Eye, nose, mouth region of obtained facial depth image (a) (b) (c) (d) (e) 6.. (a) 0.0001, (b) 0.00012, (c) 0.00015, (d) 0.00017, (e) 0.0002 Fig. 6. Generated facial depth image by trained generator using five different learning rates. (a) 0.0001, (b) 0.00012, (c) 0.00015, (d) 0.00017, (e) 0.0002

나준엽 외 2인: Generative Adversarial Network를 이용한 손실된 깊이 영상 복원 (John Junyeop Nah et al.: Depth Image Restoration Using Generative Adversarial Network) 켜 합성 실험 영상을 얻기 위해 눈, 코, 입에 이진 마스크를 어떻게 적용했는 지를 나타낸다. 각 부위는 얼굴 영역의 맨 윗부분을 기준으로 눈은 30% ~ 45% (빨간 부분), 코는 35% ~ 55% (초록 부분), 입은 57% ~ 75% (파란 부분)으로 지정하였으며 각 부위에 대해 이진 마스크의 619 높이를 고정시 너비를 변화시켜가며 손실된 깊이 영상을 생성하였다. 그림 8은 위 그림 7에서 설명한 눈, 코, 입 부분에 각각 이진 마스크를 여러 비율에 따라 적용시켜 합성 실험 영상 키고 (a) (b) (c) (d) (e) (f) (g) (h) (i) (j) 그림 8. 실제 깊이 영상에 각 부위 별 여러 비율의 이진 마스크를 적용하여 영상을 손상시킨 후 다시 복원한 결과. (a) 원본 깊이 영상, (b), (c), (d) 눈 부위에 각각 너비의 40%, 50%, 60%를 이진 마스크 삭제한 뒤 복원한 결과, (e), (f), (g) 코 부위에 각각 너비의 20%, 30%, 40%를 이진 마스크로 삭제한 뒤 복원한 결과, (h), (i), (j) 입 부위에 각각 너비의 20%, 30%, 40%를 이진 마스크로 삭제한 뒤 복원한 결과 (미니맥스 게임에 사용된 학습률: 0.00017, 복원 과정에 사용된 학습률: 0.00015) Fig. 8. Results of depth image restoration after corrupting the original depth image with binary masks with different area. (a) Original depth image, (b), (c), (d) Restoration result after corrupting 40%, 50%, and 60% of eyes part width, (e), (f), (g) Restoration result after corrupting 20%, 30%, and 40% on nose part width, (h), (i), (j) Restoration result after corrupting 20%, 30%, and 40% on mouth part width (Learning rate used in minimax game: 0.00017, learning rate used in restoration: 0.00015)

(JBE Vol. 23, No. 5, September 2018). 40%, 50%, 60%, 20%, 30%, 40%. 0.00017, 0.00015., 8 PSNR 2. PSNR PSNR.. 9. 30cm 40cm 0.. DCGAN Wasserstein distance.. (References) [1] K. Xu, J. Zhou, and Z. Wang, "A method of hole-filling for the depth map generated by Kinect with moving objects detection," Proceeding of IEEE international Symposium on Broadband Multimedia Systems and Broadcasting, pp. 1-5, June 2012. 2. Table 2. Comparison of restoration results and original depth images for each part Eyes Nose Mouth Image Corruption Ratio (%) 40 50 60 20 30 40 20 30 40 PSNR (db) 29.34 30.06 30.11 31.83 31.7 30.39 29.03 29.15 29.6 (a) (b) (c) 9.. (a) RGB, (b), (c) Fig. 9. Restoration result of real corrupted facial depth image. (a) RGB image, (b) Obtained corrupted facial depth image, (c) Restoration result on loss part

나준엽 외 2인: Generative Adversarial Network를 이용한 손실된 깊이 영상 복원 (John Junyeop Nah et al.: Depth Image Restoration Using Generative Adversarial Network) [2] [3] [4] [5] [6] L. Feng, L.-M. Po, X. Xu, K.-H. Ng, C.-H. Cheung, and K.-w. Cheung, An adaptive background biased depth map hole-filling method for Kinect, Proceeding of IEEE Industrial Electronics Society, pp. 2366 2371, November 2013. S. Ikehata, J. Cho, and K. Aizawa, "Depth map inpainting and super-resolution based on internal statistics of geometry and appearance," Proceeding of IEEE International Conference on Image Processing, pp. 938-942, September 2013. A. Radford, L. Metz, and S. Chintala, Unsupervised representation learning with deep convolutional generative adversarial networks, Proceeding of International Conference on Learning Representations, May 2016. R. A. Yeh, C. Chen, T. Yian Lim, A. G. Schwing, M. HasegawaJohnson, and M. N. Do, Semantic image inpainting with deep generative models, Proceeding of IEEE Conference on Computer Vision and Pattern Recognition, July 2017. M. Arjovsky, S.Chintala, and L. Bottou, Wasserstein GAN, Proceeding of International Conference on Machine Learning, vol. 70, pp.214-223, August. 2017. 나준엽 [7] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D.Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, Generative adversarial networks, Proceeding of Advances in Neural Information Processing Systems, December 2014. [8] A. T. Tran, T. Hassner, I. Masi, and G. Medioni, Regressing robust and discriminative 3D morphable models with a very deep neural net-work, Proceeding of IEEE Conference on Computer Vision and Pattern Recognition, July 2017. [9] C. Cao, Y. Weng, S. Zhou, Y. Tong, and K.Zhou, FaceWarehouse: a 3D facial expression database for visual computing, IEEE Transaction on Visualization and Computer Graphics, vol. 20, no. 3, pp. 413-425, March 2014. [10] P. Paysan, R. Knothe, B. Amberg, S. Romdhani, and T. Vetter, A 3D face model for pose and illumination invariant face recognition, Proceeding of IEEE International Conference on Advanced Video and Signal Based Surveillance, October 2009. [11] Intel RealSense Camera SR300, https://software.intel.com/sites/ de-fault/files/managed/ 0c/ec/realsense-sr30 0-pro duct- datasheet-rev-1-0.pdf (accessed August 13, 2018). 저자소개 년 월 인하대학교 정보통신공학과 학사 년 월 현재 인하대학교 정보통신공학과 석사과정 주관심분야 컴퓨터비전 - 2018 2 : - 2018 3 ~ : - ORCID : http://orcid.org/0000-0002-2576-5240 :, deep learning, GPGPU 심창훈 년 월 인하대학교 정보통신공학과 학사 주관심분야 컴퓨터비전 - 2018 2 : - ORCID : http://orcid.org/0000-0002-1367-1896 :, deep learning, GPGPU 박인규 - 년 2월 : 서울대학교 제어계측공학과 학사 년 2월 : 서울대학교 제어계측공학과 석사 년 8월 : 서울대학교 전기컴퓨터공학부 박사 년 9월 ~ 2004년 3월 : 삼성종합기술원 멀티미디어랩 전문연구원 년 1월 ~ 2008년 2월 : Mitsubishi Electric Research Laboratories (MERL) 방문연구원 년 9월 ~ 2015년 8월 : MIT Media Lab 방문부교수 년 7월 ~ 2019년 2월 : University of California San Diego (UCSD) 방문연구원 년 3월 ~ 현재 : 인하대학교 정보통신공학과 교수 : http://orcid.org/0000-0003-4774-7841 주관심분야 : 컴퓨터 그래픽스 및 비젼 (영상기반 3차원 형상 복원, 증강현실, computational photography), GPGPU 1995 1997 2001 2001 2007 2014 2018 2004 ORCID 621