top of page
57.jpg

Original Image

57-output.jpg

 Corrected Image

Abstract

Gaze correction aims to redirect the person’s gaze into the camera by manipulating the eye region, and it can be considered as a specific image resynthesis problem. Gaze correction has a wide range of applications in real life, such as, the eye contract of remote users in video conference systems. In this paper, we propose a novel method which is based on the inpainting model to learn from the face image to filling-in eye regions with new contents representing corrected eye gaze. Moreover, our model does not require the training dataset labelled with the specific head pose and eye angle information; thus, the training data is easy to collect. To retain the identity information of the eye region in the original input, we propose a self-guided pretrained model to learn the angle-invariance feature. Experiments show our model achieves very compelling gaze-correction results in the wild dataset which is collected from the website and will be introduced in details.

TIM图片20190628215517.png
Paper
arxiv 1906.00805 , 2019
Citation

Jichao Zhang∗, Meng Sun∗, Jingjing Chen∗, Hao Tang, Yan Yan, Xueying Qin, Nicu Sebe

(* indicates equal contributions)Bibtex

Code: Python

to About
to Services
to Work

Network Architecture

20190226154735.jpg

More Details of the Network Architecture

Supplementary materials are shown below. The network architectures of GazeGAN are shown in Table 1 ,Table 2 and Table 3. The Self-Guided network, which is employed to preserve the identity information , takes as an input the local image rescaled to 128×128 pixels. Note that both local images share variables in Self-Guided network. For the completion network, we use an encoder-decoder architecture, which will add the angle-invariance feature learned by Self-Guided network. And these features are used in discriminator as additional information when determining if the generated image is real or fake.

Here are some notations should be noted. h: the height of input images. w: the width of input images. C:the number of output channels. K:the size of kernel. S:the size of stride. P: the padding method. IN:instance normalization. FC:Fully-Connected layers. SN:use spectral normalization

(%7N_JYK8VG~Y8G7W~_A4`R.png

Table 1: Guided architecture

VQ0IN{H_TMRM5S({TZ)LTUL.png

Table 2: Generator architecture

table3.png

Table 3: Discriminator architecture

Dataset Introduction——NewGaze Dateset

 To evaluate the proposed method and the overall framework, we have investigated the benchmark datasets. However, none of them meet our task for eye correction in the wild. Thus, we collected a new dataset called NewGaze dataset. NewGaze consists of  a set of unpaired data, 40000 images.

Note that the unpaired data is not labelled with the specific eye angle and head pose information, thus, is very easy to be collected.

The unpaired data, which is collected from the CelebA-ID and the website,consists of two domains.

Domain X: 35000 face images with eyes staring at the camera;

Domain Y : 5000 face images with eyes not staring at the camera.

We crop all images (256×256) with face detection algorithm and compute the eye mask region by using facial landmarks detection algorithm. As described above,we use all data of domain X to train our model, while all data in domain Y are just as the test dataset.

aaron-carter-6.jpg

Example of Domain X

alice-trewinnard-6.jpg

Example of Domain Y

More Results

Notes that, all of our results are without any postprocessing argorithm.

Head pitch rotations greater than large degrees
86_ 9_real.jpg
86_ 9_new_output.jpg
04_ 8_real.jpg
04_ 8_new_output.jpg
76_11_real.jpg
76_11_new_output.jpg
14_ 3_real.jpg
14_ 3_new_output.jpg
14_ 9_real.jpg
14_ 9_new_output.jpg
33_ 5_real.jpg
33_ 5_new_output.jpg
Head yaw rotations greater than large degrees
31_ 4_real.jpg
31_ 4_new_output.jpg
119_11_real.jpg
119_11_new_output.jpg
109_ 6_real.jpg
109_ 6_new_output.jpg
61_12_real.jpg
61_12_new_output.jpg
76_14_real.jpg
76_14_new_output.jpg
94_ 8_real.jpg
94_ 8_new_output.jpg
173_ 4_real.jpg
173_ 4_new_output.jpg
27_ 3_real.jpg
27_ 3_new_output.jpg
87_11_real.jpg
87_11_new_output.jpg
56_13_real.jpg
56_13_new_output.jpg
78_12_real.jpg
78_12_new_output.jpg
178_ 1_real.jpg
178_ 1_new_output.jpg
180_ 6_real.jpg
180_ 6_new_output.jpg
00_ 1_real.jpg
00_ 1_new_output.jpg
0.png
1.png
156_ 5_new_output.jpg
156_ 5_real.jpg
79_ 7_real.jpg
79_ 7_new_output.jpg
92_ 1_real.jpg
97_10_new_output.jpg
0.png
1.png
128_14_new_output.jpg
128_14_real.jpg
60_13_real.jpg
60_13_new_output.jpg
182_ 2_new_output.jpg
182_ 2_real.jpg
108_ 9_real.jpg
108_ 9_new_output.jpg
0.png
1.png

More Results in GIF

created_gif17.gif
created_gif36.gif
created_gif10.gif
created_gif18.gif
created_gif15.gif
created_gif13.gif
created_gif7.gif
created_gif7.gif
created_gif5.gif
created_gif16.gif
created_gif13.gif
created_gif5.gif
created_gif14.gif
created_gif14.gif
created_gif4.gif
created_gif3.gif
created_gif12.gif
created_gif52.gif
created_gif60.gif
created_gif62.gif
created_gif54.gif
created_gif9.gif
created_gif6.gif
created_gif8.gif
created_gif70.gif
created_gif24.gif
created_gif30.gif
created_gif1.gif
created_gif3.gif
created_gif14.gif
created_gif6.gif
created_gif69.gif
created_gif39.gif
created_gif2.gif
created_gif5.gif
created_gif6.gif
created_gif22.gif
created_gif15.gif
created_gif3.gif
created_gif0.gif
bottom of page