|
1.IntroductionFor the last decade, surveillance systems have become an active research topic of computer vision, since they have become ubiquitous in public places such as airports, railway stations, college campuses, and office buildings.1–2 There are a large number of cameras in surveillance systems and they provide huge amounts of video data. The analysis of the computer vision abstained in a surveillance system often requires the ability to track people across multiple cameras. Therefore, person re-identification (Re-ID) model is generating more and more interests.3–8 Re-ID has been widely treated as a recognition problem of matching different persons across disjoint cameras.9–11 In the past five years, a large number of models have been proposed for Re-ID models. The current work can be categorized generally into two types: (1) designing of discriminative, descriptive, and robust visual descriptors to characterize a person’s appearance12–15 and (2) learning suitable distance metrics that maximize the chance of a correct correspondence.16–20 In this paper, we focus on the second type, i.e., we learn the optimal distance measure to give correct matches in Re-ID. However, it is not easy to develop a deployable and efficient Re-ID model in a new scenario (e.g., from an indoor classroom to an outdoor square). First, due to different illumination environments, posture, and view angle, the robust features obtained in one scenario will not have good performance for another scenario. Second, in order to obtain a robust Re-ID model, one must collect a large number of labeled person images about the new scenario for training. However, the work is very expensive due to both monetary cost and labeling time. Some unsupervised methods are proposed to address this problem. For example, Ma et al.21 introduced a time shift dynamic time warping model for unsupervised person representation. Ye et al.22 proposed a dynamic graph matching method to mine the intermediate estimated labels across disjoint cameras, and then with the estimated labels, its remaining steps can be considered as a supervised learning method. However, compared to supervised Re-ID methods, the matching performance of unsupervised methods is less effective when a person recognizable is under severe appearance changes.23 Recently, transfer learning mechanism has been widely used in Re-ID. The principal goal of transfer learning is to help build a Re-ID model in a new scenario (target domain) by leveraging the data collected from the other scenarios (source domain).24 For example, in a crowded station, there may exist of a large number of data used for building some Re-ID models for their own respective scopes. In order to build a Re-ID model for a new scenario, we may use these existed data in the source domain without collecting a lot of labeled data in the target domain. In Ref. 25, it is demonstrated that certain discriminative information or common variations (such as pose and resolution) shared in different scenarios can lead to significant performance gains in a new scenario. Different from original multitask learning which aims to benefit all tasks both on the target domain and source domain, transfer learning for Re-ID mainly aims to benefit the target one. In this work, we first propose a maximum mean discrepancy based on class distribution called MMDCD to measure distribution difference across domains. MMDCD embeds the discriminative information of data taken from the source domain into the concept of the maximum mean discrepancy (MMD).26 Minimizing MMDCD leads to minimize the distribution difference across domains in a supervised way. Then we propose a discriminative deep transfer metric learning method called DDTML for cross-scenario transfer Re-ID. Figure 1 shows the basic idea of the proposed method. Using a deep neural network, DTDML learns a set of multilayers nonlinear transformations to transfer discriminative information from the source domain to the target domain; meanwhile, DTDML reduces the distribution divergence between the source data and the target data by minimizing MMDCD at the top layer of the network. The contribution of this work can be summarized in the following three aspects.
2.Related WorkAccording to the process of Re-ID, existing works can be generally divided into two categories, namely, seeking robust features methods and seeking the optimal distance learning methods. The goal of seeking robust features methods is to increase their representative capability. For example, Ma et al.27 proposed a BiCov descriptor based on Gabor filters and the covariance descriptor to track persons. Kviatkovsky et al.28 constructed an invariant intradistribution structure of color to adopt with a wide range of imaging conditions. Yang et al.29 developed a robust semantic salient color names-based color descriptor to calculate photometric variance. However, descriptors of visual appearance are so highly susceptible to cross-view variations and heavily rely on foreground segmentations that it is difficult for them to achieve a balance between discriminative power and robustness. As the popular similarity distance learning methods, the goal of metric learning methods is to find a distance or similarity function of extracted features from different persons’ images to make the most likely correct matching. For example, Pedagadi et al.30 applied a two-stage method, local Fisher discriminant analysis (LFDA), in a low-manifold learning framework using principal component analysis (PCA) and the LFDA. Kostinger et al.16 proposed a metric learning principle of keeping it simple and straightforward (KISSME) to learn a distance metric from equivalence constraints based on a statistical inference perspective. Hu et al.31 exploited the discriminative information to propose a discriminative deep metric learning (DDML), which is a major reference of this paper. Note that cross-scenarios transfer learning has been adopted for Re-ID methods in the hope that the target domain (new scenario) can exploit transferable discriminative information from the source domain (other scenarios) with labeled images. For example, Wang et al.25 proposed the constrained asymmetric multitask discriminative component analysis (cAMT-DCA) method to explore discriminative modeling in the shared latent space for cross-scenarios transfer learning. Cheng et al.32 proposed a transfer metric learning method OurTransD to learn both the commonalities and the personality of the data from different scenarios jointly. Zhang et al.33 proposed a two-stage transfer metric learning (TSTML) method, which transfers the generic knowledge information from the source set in the first stage and then transfers the distance metric for each probe-specific person in the second stage. In terms of similarity function, optimization method, whether a transfer learning and deep learning method, Table 1 summarizes seven Re-ID methods, i.e., LFDA, KISSME, DDML, TSTML, cAMT-DCA, OurTransD, and DDTML, which is proposed in this study. Different from the other three transfer learning methods, our proposed DDTML uses a deep learning network to learn a set of multilayer nonlinear projections for the cross-scenario transfer learning. In particular, an MMDCD is proposed to measure distribution difference across domains. Table 1LFDA, KISSME, DDML, TSTML, cAMT-DCA, and OurTransD versus DDTML.
3.Proposed Methods3.1.Discriminative Deep Metric LearningDDML method is originally proposed for face verification in the wild. DDML uses a deep neural network to learn the nonlinear mapping function of samples for projecting face samples into the feature space. Assume DDML constructs a deep neural network with layers, is the units in the ’th layer, where . For a given person image sample , is the original input of the network and is the output of the first layer, where and are the projection matrix and bias vector in the first layer, respectively. is a nonlinear activation function, which operates component wisely, such as widely used tanh or sigmoid functions. Then using as the input of the second layer, we can obtain the output of this layer , i.e., . In this case, we can obtain the output of topmost layer where is a parametric nonlinear function determined by the parameters and ().For two person images and , they will be finally represented as and at the topmost layer of the network. Then using the squared Euclidean distance, the distance between and at the top level can be measured as The optimization problem of DDML is designed as follows: where the function is the smoothed approximation for , is a sharpness parameter, is the Frobenius norm, is a regularization parameter, and is a threshold. The pairwise label denotes the similarity of the pairs : means and are matched image pairs, means and are mismatched image pairs. can be determined as follows:From the optimization problem shown in Eq. (3), it can be seen that without enough training data in a new scenario, we cannot directly use data collected from different scenarios to help build the Re-ID model in this new scenario. This is the key problem we aim to solve in this work. 3.2.Discriminative Deep Transfer Metric Learning methodBased on the projection scheme for deep neural network, we learn a set of multilayers nonlinear transformations to project the data in the source domain and target domain into the same transformed space. Therefore, it is needed to measure the distribution difference between the source domain and target domain in this transformed space. As a well-known criterion to estimate the distance between different distributions, MMD) is a nonparametric estimation criterion and it does not need an intermediate density estimate.26 Let and be the training set in the source domain and target domain, respectively, where both and are the samples of dimensionality , and are the labels of and , respectively, and are the numbers of training data in the source domain and target domain, respectively. The distance between distributions of two domains is equivalent to the distance between the mean of total-class data across domains, which can be written as follows:26 However, MMD measures the distribution difference between two domains in an unsupervised way. That is to say, MMD ignores the label information of samples. In addition, for a practical transfer Re-ID task, there often exist imbalance between matched (positive) image pairs and mismatched (negative) pairs. In order to carry out effective transfer learning, we propose an MMDCD. MMDCD embeds the discriminative information of data taken from the source domain into the concept of the MMD by the following equation: where and are the matched and mismatched image samples in the source domain, respectively. and () are the numbers of matched and mismatched image samples in the source domain, respectively. Following the deep network learning strategy in Ref. 29, the nonlinear representation can be computed using Eq. (1) at the topmost layer of the network. Obviously, in order to measure the distance between the mean of the data across domains, MMDCD not only utilizes the label information of data taken from the source domain, but also sets the different coefficients to represent the weight of matched and mismatched pairs according to their different sizes.As shown in Fig. 1, DDTML constructs a deep neural network to obtain the representations of data in the source domain and target domain through a multiple layers of nonlinear transformations. Considering minimizing MMDCD at the top layer of the network, the optimization problem of DDTML can be given as follows: where is the MMDCD at the ’th layer of deep neural network. and are the regularization parameters.To solve the optimization problem in Eq. (7), we use the stochastic subgradient descent scheme to obtain the parameters and , where . The gradient of the objective function with respect to the parameters and can be computed as follows: where and , and are the original inputs.For the ’th layer of our network, we can obtain the following updating equations: For the other layers of our network, we can obtain the following updating equations: where denotes the element-wise multiplication. and () are given as follows:Then and can be updated using the gradient descent algorithm until convergence as follows: where is the learning rate.Based on the analysis above, we summarize the entire construction procedure of DDTML in Algorithm 1. Algorithm 1DDTML
4.Experiments4.1.Datasets and Experimental SettingIn our experiments, four Re-ID datasets are adopted: 3DPeS,34 i-LIDS,35 CAVIAR,19 and VIPeR.36 The 3DPeS dataset is a collection of 1011 person images of 192 individuals from eight different surveillance cameras captured on an academic campus. The i-LIDS dataset is a collection of 119 person images captured in airport. Each person is with an average of four images. Therefore, i-LIDS consists of 476 images in total. The CAVIAR dataset is a collection of 1220 person images from 72 individuals with 10 to 20 images per person. The VIPeR dataset is a collection of 632 person images by two different camera views, so it consists of 1264 images. In order to construct the transfer learning Re-ID model, we choose one dataset as the target dataset and another dataset as the source dataset from the other three datasets following the same settings of.25 So there are in total 12 cross-scenario transfer learning tasks. In our experiments, all person images from the above four datasets are scaled to for feature extraction. Following the same settings of Ref. 25, three kinds of features descriptor: color, LBP, and HOG are generated for each image. After extracting the feature vector, we use PCA to compress them into 500-dimensional feature vectors. For comparison purposes, six state-of-the-art Re-ID methods are applied to compare against our proposed DDTML. The comparison methods can be grouped into two groups: (1) nontransfer learning methods: LFDA,30 KISSME,31 and DDML31 and (2) transfer learning methods: geometry preserving large margin nearest neighbor (GPLMNN),37 OurTransD,32 and cAMT-DCA.25 Furthermore, in order to better observe the behavior of MMDCD, we develop another transfer learning Re-ID method called DDTML-MMD through replacing MMDCD in DDTML with MMD criterion. We train a deep network with three layers for DDTML, and its neural nodes are given as: for all datasets. Based on our extensive experiments, the tanh function is used in function, and the parameters , , , and are set to be , 10, 3, and 0.3, respectively. In our experiments, we randomly split the target dataset into two equal partitions; one partition is used as target training set and the other partition is used as target testing set. For five transfer learning methods, all person images in the source dataset and target training set are used for training, and all images in the target testing set are used for testing. For three nontransfer learning methods, all images in source dataset are used for training. In particular, in order to observe the performance change of nontransfer learning methods on transfer datasets, LFDA and KISSME are trained in three cases. LFDA-S and KISSME-S only use the source dataset for training; LFDA-T and KISSME-T only use the target dataset for training, whereas LFDA-Mix and KISSME-Mix use both the source and target training datasets for training. Following Ref. 38, the performance of each method is evaluated in terms of the cumulative matching characteristic (CMC) in our experiments. The CMC represents the probability of finding the correct match over the top image ranking, with varying from 1 to 20. The CMC described above is usually used to measure the performance of closed-set Re-ID problem. It assumes the same person can be found both in the probe set and gallery set. But in many real-world scenarios, this assumption is often not satisfied, e.g., the scenarios with imposters. In order to simulate these open-set scenarios, only images of 40% of the gallery people are randomly removed. The receiving operating characteristic (ROC) curve on i-LIDS as target dataset is used as the evaluation metric to compare DDTML with other algorithms. In order to make our results fair, we repeat the aforementioned partition 10 times for each dataset, and both the CMC and ROC curves for 10 runs are recorded. 4.2.Results and AnalysisIn this section, we examine the effectiveness of the proposed method DDTML by comparing their performance with LFDA (LFDA-S, LFDA-T, and LFDA-Mix), KISSME (KISSME-S, KISSME-T, and KISSME-Mix), DDML, GPLMNN, DDTML-MMD, OurTransD, and cAMT-DCA on 12 cross-scenario transfer Re-ID datasets. The experimental results of CMC are shown in Tables 2Table 3Table 4–5, respectively. Best results are in boldface font. The ROC curves of eight methods (LFDA-Mix, KISSME-Mix, DDML, GPLMNN, DDTML-MMD, OurTransD, cAMT-DCA, and DDTML) on the i-LIDS dataset as target dataset are shown in Fig. 2. Because the performance of both LFDA-S and LFDA-T is weaker than LFDA-Mix and the performance of both KISSME-S and KISSME-T is weaker than KISSME-Mix, the ROC curves of these four methods are not demonstrated in Fig. 2. Table 2Matching rate (%) on the VIPeR dataset as target dataset.
Table 3Matching rate (%) on the i-LIDS dataset as target dataset.
Table 4Matching rate (%) on the CAVIAR as target dataset.
Table 5Matching rate (%) on the 3DPeS dataset as target dataset.
From Tables 2Table 3Table 4–5 and Fig. 2, we can have the following conclusions:
5.ConclusionIn this paper, by integrating DDML with transfer learning, we propose a DDTML method to learn a distance metric that measures the similarity between image pairs of Re-ID dataset. But DDTML is not a simple transfer learning version of DDML. Taking account of the discriminative information of data and inherent characteristics of Re-ID dataset, the developed method also utilizes an MMDCD to minimize the distribution divergence of source data and target data. Extensive experimental results on the 3DPeS, i-LIDS, CAVIAR, and VIPeR datasets have shown that our method outperforms the state-of-the-art methods on most of the cross-scenario transfer Re-ID tasks. Since the formula of MMDCD is uncomplicated, how to take full advantage of the Re-ID dataset is still an interesting direction of future work. DisclosuresThis paper has been listed in the proceedings of 2018 SPIE Commercial + Scientific Sensing and Imaging (SI18C), volume DL10670. AcknowledgmentsThis work was supported in part by the National Natural Science Foundation of China under Grant Nos. 61502058 and 61572085, and Jiangsu Joint Research Project of Industry, Education, and Research under Grant No. BY2016029-15. ReferencesS. Zhou et al.,
“Point to set similarity based deep feature learning for person re-identification,”
in Proc. of IEEE Conf. on Computer Vision and Pattern Recognition (CVPR),
(2017). https://doi.org/10.1109/CVPR.2017.534 Google Scholar
Z. Zhong et al.,
“Re-ranking person re-identification with k-reciprocal encoding,”
in Proc. of IEEE Conf. on Computer Vision and Pattern Recognition (CVPR),
(2017). https://doi.org/10.1109/CVPR.2017.389 Google Scholar
C. Su et al.,
“Attributes driven tracklet-to-tracklet person re-identification using latent prototypes space mapping,”
Pattern Recognit., 66 4
–15
(2017). https://doi.org/10.1016/j.patcog.2017.01.006 Google Scholar
X. Liu et al.,
“Person re-identification by multiple instance metric learning with impostor rejection,”
Pattern Recognit., 67 287
–298
(2017). https://doi.org/10.1016/j.patcog.2017.02.015 Google Scholar
L. Ren et al.,
“Multi-modal uniform deep learning for RGB-D person re-identification,”
Pattern Recognit., 72 446
–457
(2017). https://doi.org/10.1016/j.patcog.2017.06.037 Google Scholar
X. Ma et al.,
“Person re-identification by unsupervised video matching,”
Pattern Recognit., 65 197
–210
(2017). https://doi.org/10.1016/j.patcog.2016.11.018 Google Scholar
G. Watson and A. Bhalerao,
“Person reidentification using deep foreground appearance modeling,”
J. Electron. Imaging, 27
(5), 051215
(2018). https://doi.org/10.1117/1.JEI.27.5.051215 JEIME5 1017-9909 Google Scholar
L. Hou et al.,
“Normalized distance aggregation of discriminative features for person reidentification,”
J. Electron. Imaging, 27
(2), 023006
(2018). https://doi.org/10.1117/1.JEI.27.2.023006 JEIME5 1017-9909 Google Scholar
Z. Cao et al.,
“Face recognition with learning-based descriptor,”
in Proc. of IEEE Conf. on Computer Vision and Pattern Recognition (CVPR),
2707
–2714
(2010). https://doi.org/10.1109/CVPR.2010.5539992 Google Scholar
X. Glorot and Y. Bengio,
“Understanding the difficulty of training deep feedforward neural networks,”
in Proc. of the Int. Conf. on Artificial Intelligence and Statistics,
249
–256
(2010). Google Scholar
M. Guillaumin, J. Verbeek and C. Schmid,
“Is that you? Metric learning approaches for face identification,”
in Proc. of the IEEE 12th Int. Conf. on Computer Vision,
498
–505
(2009). https://doi.org/10.1109/ICCV.2009.5459197 Google Scholar
S. Liao et al.,
“Person re-identification by local maximal occurrence representation and metric learning,”
in Proc. of IEEE Conf. on Computer Vision and Pattern Recognition (CVPR),
2197
–2206
(2015). https://doi.org/10.1109/CVPR.2015.7298832 Google Scholar
L. Bazzani, M. Cristani and V. Murino,
“Symmetry-driven accumulation of local features for human characterization and re-identification,”
Comput. Vision Image Understanding, 117
(2), 130
–144
(2013). https://doi.org/10.1016/j.cviu.2012.10.008 Google Scholar
M. Farenzena et al.,
“Person re-identification by symmetry-driven accumulation of local features,”
in Proc. IEEE Computer Society Conf. on Computer Vision and Pattern Recognition,
2360
–2367
(2010). https://doi.org/10.1109/CVPR.2010.5539926 Google Scholar
L. An et al.,
“Person reidentification with reference descriptor,”
IEEE Trans. Circuits Syst. Video Technol., 26
(4), 776
–787
(2016). https://doi.org/10.1109/TCSVT.2015.2416561 Google Scholar
M. Köstinger et al.,
“Large scale metric learning from equivalence constraints,”
in Proc. of IEEE Computer Society Conf. on Computer Vision and Pattern Recognition,
2288
–2295
(2012). https://doi.org/10.1109/CVPR.2012.6247939 Google Scholar
P. M. Roth et al., Mahalanobis Distance Learning for Person Re-identification, 247
–267 Springer, Cambridge, UK
(2014). Google Scholar
C. Loy, C. Liu and S. Gong,
“Person re-identification by manifold ranking,”
in Proc. of the 20th IEEE Int. Conf. on Image Processing,
3567
–3571
(2013). https://doi.org/10.1109/ICIP.2013.6738736 Google Scholar
W. S. Zheng, S. Gong and T. Xiang,
“Re-identification by relative distance comparison,”
IEEE Trans. Pattern Anal. Mach. Intell., 35
(3), 653
–668
(2013). https://doi.org/10.1109/TPAMI.2012.138 ITPIDJ 0162-8828 Google Scholar
L. An, S. Yang and B. Bhanu,
“Person re-identification by robust canonical correlation analysis,”
IEEE Signal Process. Lett., 22
(8), 1103
–1107
(2015). https://doi.org/10.1109/LSP.2015.2390222 IESPEJ 1070-9908 Google Scholar
X. L. Ma et al.,
“Person Re-identification by unsupervised video matching,”
Pattern Recognit., 65 197
–210
(2017). https://doi.org/10.1016/j.patcog.2016.11.018 Google Scholar
M. Ye et al.,
“Dynamic label graph matching for unsupervised video re-identification,”
in Proc. of Int. Conf. on Computer Vision,
5152
–5160
(2017). https://doi.org/10.1109/ICCV.2017.550 Google Scholar
P. X. Peng et al.,
“Unsupervised cross-dataset transfer learning for person re-identification,”
in Proc. of IEEE Conf. on Computer Vision and Pattern Recognition (CVPR),
1306
–1315
(2016). https://doi.org/10.1109/CVPR.2016.146 Google Scholar
J. L. Hu et al.,
“Cross-scenario transfer metric learning for person re-identification,”
IEEE Trans. Image Process., 25
(12), 5576
–5588
(2016). https://doi.org/10.1109/TIP.2016.2612827 IIPRE4 1057-7149 Google Scholar
X. Wang et al.,
“Cross-scenario transfer person re-identification,”
IEEE Trans. Circuits Syst. Video Technol., 26
(8), 1447
–1460
(2016). https://doi.org/10.1109/TCSVT.2015.2450331 Google Scholar
S. J. Pan, J. T. Kwok and Q. Yang,
“Transfer learning via dimensionality reduction,”
in Proc. of the 23rd National Conf. on Artificial Intelligence (AAAI),
677
–682
(2008). Google Scholar
B. Ma, Y. Su and F. Jurie,
“BiCov: a novel image representation for person reidentification and face verification,”
in Proc. of the 2012 British Machine Vision Conf.,
1
–11
(2012). https://doi.org/10.5244/C.26.57 Google Scholar
I. Kviatkovsky, A. Adam and E. Rivlin,
“Color invariants for person re-identification,”
IEEE Trans. Pattern Anal. Mach. Intell., 35
(7), 1622
–1634
(2013). https://doi.org/10.1109/TPAMI.2012.246 ITPIDJ 0162-8828 Google Scholar
Y. Yang et al.,
“Salient color names for person re-identification,”
in Proc. of European Conf. on Computer Vision,
536
–551
(2014). https://doi.org/10.1007/978-3-319-10590-1_35 Google Scholar
S. Pedagadi et al.,
“Local Fisher discriminant analysis for pedestrian re-identification,”
in Proc. of IEEE Conf. on Computer Vision and Pattern Recognition,
3318
–3325
(2013). https://doi.org/10.1109/CVPR.2013.426 Google Scholar
J. Hu, J. Lu and Y. P. Tan,
“Discriminative deep metric learning for face verification in the wild,”
in Proc. of IEEE Conf. on Computer Vision and Pattern Recognition,
1875
–1882
(2014). https://doi.org/10.1109/CVPR.2014.242 Google Scholar
D. Cheng et al.,
“Cross-scenario transfer metric learning for person re-identification,”
Pattern Recognit. Lett.,
(2018). https://doi.org/10.1016/j.patrec.2018.04.023 PRLEDG 0167-8655 Google Scholar
G. Zhang et al.,
“People re-identification using two-stage transfer metric learning,”
in Proc. of 14th IAPR Int. Conf. on Machine Vision Applications (MVA),
588
–591
(2015). https://doi.org/10.1109/MVA.2015.7153260 Google Scholar
W. Li and X. Wang,
“Locally aligned feature transforms across views,”
in Proc. of IEEE Conf. on Computer Vision and Pattern Recognition,
3594
–3601
(2013). https://doi.org/10.1109/CVPR.2013.461 Google Scholar
M. Sugiyama,
“Dimensionality reduction of multimodal labeled data by local Fisher discriminant analysis,”
J. Mach. Learn. Res., 8
(8), 1027
–1061
(2007). Google Scholar
J. V. Davis et al.,
“Information-theoretic metric learning,”
in Proc. of the 24th Int. Conf. on Machine Learning,
209
–216
(2007). https://doi.org/10.1145/1273496.1273523 Google Scholar
P. Yang, K. Huang and C. L. Liu,
“Geometry preserving multi-task metric learning,”
Mach. Learn., 92
(1), 133
–175
(2013). https://doi.org/10.1007/s10994-013-5379-y MALEEZ 0885-6125 Google Scholar
W. Zheng et al.,
“Partial person re-identification,”
in Proc. of IEEE Int. Conf. on Computer Vision (ICCV),
4678
–4686
(2015). https://doi.org/10.1109/ICCV.2015.531 Google Scholar
BiographyTongguang Ni received his PhD from Jiangnan University in May 2015. He is a lecturer in the School of Information Science and Engineering, Changzhou University, Changzhou, China. His current research interests include pattern recognition, intelligent computation, and their application. Xiaoqing Gu received her PhD in light industry information technology and engineering from Jiangnan University, Wuxi, China, in 2017. She is a lecturer in the School of Information Science and Engineering, Changzhou University, Changzhou, China. She has published more than 10 papers in international/national journals, including the IEEE Transactions on Industrial Informatics and IEEE Transactions on Systems, Man and Cybernetics: Systems. Her current research interests include pattern recognition and machine learning. Hongyuan Wang received his PhD in computer science from Nanjing University of Science and Technology. He is currently a professor at Changzhou University. His general research interest is in pattern recognition and intelligence system. His current interest is in pedestrian trajectory discovery in intelligent video surveillance. |