• Non ci sono risultati.

Ph.D Forum: Generating domain and pose variations between pair of cameras for person re-identification

N/A
N/A
Protected

Academic year: 2021

Condividi "Ph.D Forum: Generating domain and pose variations between pair of cameras for person re-identification"

Copied!
2
0
0

Testo completo

(1)

Ph.D Forum: Generating Domain and Pose Variations between

Pair of Cameras for Person Re-Identification

Asad Munir

University of Udine asad.munir@uniud.it

Gian Luca Foresti

University of Udine gianluca.foresti@uniud.it

Christian Micheloni

University of Udine christian.micheloni@uniud.it

ABSTRACT

Person re-identification (re-id) remains an important task that aims to retrieve a person’s images from an image dataset, given a probe image. The lack of cross-view (pose variations) training data and significant intra-class (domain) variations across different cameras make re-id more challenging. To solve these issues, this work pro-poses a Domain and Pose Invariant Generative Adversarial Network (DPI-GAN) to generate images for both domain and pose variations capture. It is based on a CycleGAN structure in which the gener-ator networks are conditioned on a new pose. Identity and pose discriminators networks are used to monitor the image genera-tion process. These generated images are used for learning domain and pose invariant features to improve the performance of person re-identification.

KEYWORDS

person re-identification, image generation, domain variations , pose variations, camera to camera translation

1

INTRODUCTION

One of the most important and challenging problem in the field of video surveillance is person re-identification (re-id) which aims to match person images with the same identity across non over-lapping camera views. In this task, person image encouters many changes independent of the person’s identity. These changes in-clude appearance, background (domain variations), viewpoint, pose variations, lightning and occlusions. Wide range of deep learning methods have been proposed to improve the performance of re-id. Generative adversarial network (GAN) [2] is gaining popularity in image generation to increase the re-id performance. Existing GAN based methods consider either domain variations [6] or pose variations [4] to generate new images.

In this work, we propose Domain and Pose Invariant Generative Adversarial Network (DPI-GAN) to generate images by changing both domain and pose in a pair of cameras. The proposed DPI-GAN uses CycleDPI-GAN [7] approach to translate images from one domain to another. The generators are conditioned on a new pose to generate image in new domain with given pose. Identity and pose discriminators are used with each generator to preserve the

Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from permissions@acm.org.

ICDSC 2019, September 9–11, 2019, Trento, Italy © 2019 Association for Computing Machinery. ACM ISBN 978-1-4503-7189-6/19/09. . . $15.00 https://doi.org/10.1145/3349801.3357135

identity and conversion to new pose in the generated images. For each of the two training cycles, the proposed framework trains the two generators and two discriminators. The images are generated from one camera’s domain to other camera’s domain with a new pose and returning back to original domain with a new pose.

The next section explains the proposed DPI-GAN framework. Section 3 describes experimental results and parameters. Acknowl-edgement and conclusion are included in last two sections.

2

DOMAIN AND POSE INVARIANT

GENERATIVE ADVERSARIAL NETWORK

Our proposed DPI-GAN aims to generate an image with new pose and domain from an input image and a skeleton pose image. Skele-ton pose images are calculated using human pose estimator [1]. Input image and skeleton pose image are concatenated and fed into the generator network to generate the image into given pose. We used CycleGAN [7] approach to train the network for capturing the domain changes between pair of cameras. First Cycle of our framework is shown in Figure 1 which is transferring an image from camera A to camera B and then reconstruct that image back to camera A with a different pose. Second Cycle is the same with start-ing from camera B to A and the reconstruct to Camera B. With this, we are training the two generators which are generating images from domain A to B and vice versa. Identity and pose discriminators are used to identify real and fake generated images.

2.1

Training

Assume we have

ami , xm m=1...Mi=1...Mi and

bnj,yn n=1...Ni=1...N jpersons images in domainA and B respectively, where M and N are the number of images in both domains.i and j are the pose indexes from the total posesMi andNj of a person. The skeleton pose images are denoted asPiandPjfor ith and jth pose respectively. xmandynare the persons identities and for every training sample xm = yn. The full loss function is denoted as:

L = arg min

G maxD λ1LGAN+ λ2Lcycle+ λ3Lidentity (1)

where

LGAN = Ea,ai∈ρ,pi∈ρplog(Dp(Pi, ai).Di(a, ai))

+ Ea ∈ρ,pi∈ρP, ˆai∈ρˆlog[(1 −Dp(Pi, ˆai)).(1 − Di(a, ˆai))]

(2)

Lcycle = || ˆai−ai||1+ || ˆbj−bj||1 (3)

(2)

ICDSC 2019, September 9–11, 2019, Trento, Italy Asad Munir, Gian Luca Foresti, and Christian Micheloni

Figure 1: Overview of our framework. Gen A-B and Gen B-A are the generators to transfer from domain A to B and from B to A respectively. Dp and Di are the pose and identity discriminators. C is symbol for concatenation

for real,fake and skeleton pose images. We use least square loss which is more stable [7]. InLcycle, ˆaiand ˆbjare the reconstructed images as shown in figure 1.ai andbj are the ground truth im-ages for skeleton posesPiandPj. Identity mapping loss is used to preserve the color composition between input and output [7].

(a) (b)

(c) (d)

Figure 2: Results generated by the two generators. (a) and (c) are the ground truths from camera 1 (A domain) and camera 6(B domain) respectively. (b) shows the output of generator B − A. The output of generator A − B is shown in (d).

3

EXPERIMENTAL RESULTS

We select camera 1 and camera 6 images from Market-1501 [5] dataset and all the images are resized into 256 x 128. Adam optimizer is used withβ1= 0.5 and β2= 0.999. Learning rates for generator

and discriminator are 0.0002 and 0.0001 respectively. Generator consists of encoder decoder network with 9 ResNet basic blocks and PatchGAN [3] structure is used for all discriminators

3.1

Quantitative Results

The qualitative results of the proposed method are shown in figure 2. Generated images between two camera domains and their ground truths are shown. The inputs are from opposite domain and having different poses in each case.

4

CONCLUSION

We have proposed an image generation method which captures both domain and pose changes for re-id. In contrast with the pre-vious approaches the proposed method merge both these varia-tions in a single network. Generated images with the proposed approach provide domain and pose invariant features for person re-identification. Experimental results prove the image generation with above mentioned variations.

ACKNOWLEDGMENTS

This work was supported by EU H2020 MSCA through Project ACHIEVE-ITN (Grant No 765866)

REFERENCES

[1] Zhe Cao, Tomas Simon, Shih-En Wei, and Yaser Sheikh. 2017. Realtime multi-person 2d pose estimation using part affinity fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 7291–7299.

[2] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. 2014. Generative adversarial nets. In Advances in neural information processing systems. 2672–2680. [3] Phillip Isola, Jun-Yan Zhu, Tinghui Zhou, and Alexei A Efros. 2017.

Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE conference on computer vision and pattern recognition. 1125–1134.

[4] Xuelin Qian, Yanwei Fu, Tao Xiang, Wenxuan Wang, Jie Qiu, Yang Wu, Yu-Gang Jiang, and Xiangyang Xue. 2018. Pose-normalized image generation for person re-identification. In Proceedings of the European Conference on Computer Vision (ECCV). 650–667.

[5] Liang Zheng, Liyue Shen, Lu Tian, Shengjin Wang, Jingdong Wang, and Qi Tian. 2015. Scalable person re-identification: A benchmark. In Proceedings of the IEEE international conference on computer vision. 1116–1124.

[6] Zhun Zhong, Liang Zheng, Zhedong Zheng, Shaozi Li, and Yi Yang. 2018. Camera style adaptation for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 5157–5166.

Riferimenti

Documenti correlati

Despite the observed differences between healthy controls and ALS patients in terms of MST dissimilarity and leaf fraction, and clinically relevant correlations between disability

To evaluate the protective ability of the vesicles against oxidative damage, the cells were seeded into 96-well plates, incubated for 24 h and exposed simultaneously to

Bridging: formazione di connessioni e reti locali flessibili, approfittando della presenza di una ‘comunità debole’. Promozioni delle reti internazionali: Tramite strategie di

Ecco, sono questi gli aspetti che consentono di misurare l’urgenza e la rilevanza dello studio delle falsificazioni di Max Aub e dell’importanza dei risultati sintetizzati in questo

% 2011 Evoluzione del contenuto di sostanza organica su un campione di 63 suoli vitati trentini: confronto 1980-2011 per zone... di Cembra Vallagarina Valdadige Zona

In this work, we have introduced the order-theoretic notion of uniform closure operator on complete lattices and shown that this is the right key property for generalizing the

how: one rig + one skinned mesh + few skeletal animations mesh skinning: single bone enough in this case. if very low poly mesh (few polys):

The technique consists in a model- based feature matching pose initialization process which employs single low- resolution images captured by a monocular vision system and