Adaptive Human Tracking Across Non-overlapping Cameras in Depression Angles

2015-11-21 07:09ShaoQuan邵荃LiangBinbin梁斌斌ZhuYan朱燕ZhangHaijiao张海蛟ChenTao陈涛
关键词:陈涛

Shao Quan(邵荃)*,Liang Binbin(梁斌斌),Zhu Yan(朱燕),Zhang Haijiao(张海蛟),Chen Tao(陈涛)

1.College of Civil Aviation,Nanjing University of Aeronautics and Astronautics,Nanjing 211100,P.R.China;

2.Institute of Public Safety Research,Tsinghua University,Beijing 100084,P.R.China

Adaptive Human Tracking Across Non-overlapping Cameras in Depression Angles

Shao Quan(邵荃)1*,Liang Binbin(梁斌斌)1,Zhu Yan(朱燕)1,Zhang Haijiao(张海蛟)1,Chen Tao(陈涛)2

1.College of Civil Aviation,Nanjing University of Aeronautics and Astronautics,Nanjing 211100,P.R.China;

2.Institute of Public Safety Research,Tsinghua University,Beijing 100084,P.R.China

To track human across non-overlapping cameras in depression angles for applications such as multi-airplane visual human tracking and urban multi-camera surveillance,an adaptive human tracking method is proposed,focusing on both feature representation and human tracking mechanism.Eeature representation describes individual by using both improved local appearance descriptors and statistical geometric parameters.The improved feature descriptors can be extracted quickly and make the human feature more discriminative.Adaptive human tracking mechanism is based on feature representation and it arranges the human image blobs in field of view into matrix. Primary appearance models are created to include the maximum inter-camera appearanceinformation captured from different visual angles.The persons appeared in camera are first filtered by statistical geometric parameters.Then the one among the filtered persons who has the maximum matching scale with the primary models is determined to be the target person.Subsequently,the image blobs of the target person are used to update and generate new primary appearance models for the next camera,thus being robust to visual angle changes.Experimental results prove the excellence of the feature representation and show the good generalization capability of tracking mechanism as well as its robustness to condition variables.

adaptive human tracking;appearance features;geometric features;non-overlapping camera;depression angle

0 Introduction

Multi-camera visual surveillance systems have currently been widely distributed in many areas for applications of continuously tracking interesting objects,early-warning of abnormal events and so on.Particularly,in some cases of multi-airplane visual human tracking,aerial photography and urban multi-camera surveillance from tall buildings,there exists a multi-camera visual surveillance system in which all cameras are installed at high places with wide visual range and have large visual depression angles.A fundamental task for these particular multi-camera surveillance systems with visual depression angles is to associate people across different camera views at different locations and time.This is known as the human tracking problem in visual depression angles.

Human tracking in visual depression angles faces an issue of visually matching a target person across different cameras distributed over disjoint scenes of distance and time differences.In this case,classical human tracking algorithms will fail since cameras do not overlap.Hence,non-overlapping camera human tracking in this paper describes algorithms that deal with human tracking across non-overlapping camera views.Non-overlapping camera human tracking techniques build upon single camera human tracking techniques,for a person needs to be tracked within one camera field of view(EOV)before it can be tracked in that of another.Therefore,the problem of non-overlapping camera human tracking becomes how to match individual from one independent surveillance area to another.It has many challenges among which feature representation and tracking mechanism are the most difficult.The tracked people should be differentiated from numerous visually similar but different people in those views,which requires a sufficiently discriminative feature representation to distinguish the target person from those similar yet different candidates.Potentially,different views may be taken from various shooting angles,causing dissimilar backgrounds under diverse illumination conditions,or with other view variables.It thus requires a robust tracking mechanism that can resist inter-camera and intra-camera shooting angle changes,as well as illumination change.

Designing suitable feature representation for human tracking is a critical and challenging problem.Ideally,the features should be robust to illumination changes,visual angle altering,foreground errors,occlusion,and low image resolution.Contemporary approaches typically exploit low-level features such as appearance[1-3],spatial structure[4-5],or their combinations[6-12].This is because these features can be relatively easily and reliably measured.Moreover,they provide a reasonable level of inter-person discrimination and then can distinguish different people clearly.

Gianfranco et al.[1]and Hyun-Uk et al.[2]used appearance to re-identify people,and proved that appearance feature had good performance in identifying individuals.However,they could not deal with illumination change very well.Generally,in single visual angle individuals can be discriminated based on their appearances.However,appearance feature will alter with the change of visual angle,which occurs frequently across nonoverlapping cameras.In this case,appearance is quite limited to distinguish individuals.Related researches compensated the limitation by combining geometric features with appearance[6-10]. Madden et al.[6]focused on a framework based on robust shape and appearance features.However,his proposal solely employed height as shape feature without considering the limitation of height in distinguishing human beings due to the close resemblance of human height.A good way to amend the limitation of height is to combine gait feature with height to compose a shape feature. Takayuki et al.[8]proposed a method that tracked a walking human using gait features to calculate a robust motion signature.Despite the well-done performance of gait feature in his method,a high accuracy rate and low computational cost were still far from being achieved[9].Moreover,gait features are tough to adjust to the visual angle change.In short,it needs more explorations to obtain a better feature representation.

Once a suitable feature representation has been obtained,previous literatures typically used the nearest-neighbor[4]or model-based matching algorithms such as support-vector ranking[11]for human tracking.In each case,a distance metric must be chosen to measure the similarity between two samples.In single camera,both the modelbased matching approaches[12-13]and the nearest neighbor distance metrics[14-15]can be optimized to maximize tracking performance.However,despite their excellent performance in single camera,they are still limited in coping with those intractable challenges in non-overlapping cameras.The first challenge is to overcome the intercamera and intra-camera variations.These variations include the change of appearance feature,spatial structure,illumination condition and some other parameters,which makes non-overlapping camera human tracking tough to work well in various scenes from different visual angles.Eurthermore,such variations between non-overlapping cameras are in general complex and multimodal,and therefore complex for an algorithm to learn.The second challenge is how to achieve a good generalization capability.Previously,once trained for a specific pair of cameras,most models could not generalize well to other cameras from different visual angles[16]because there wasno connection between them.Therefore,it is necessary to establish a tracking mechanism with good generalization that models can be established once and then adaptively applied to different camera configurations.

To solve the problems mentioned above,improved appearance feature and geometric feature are explored,and an adaptive tracking mechanism is also designed.The improved appearance and geometric features make up a discriminative and robust human feature.The appearance feature includes color and texture information.Color is analyzed in hue-saturation-value(HSV)space.The HSV space is evenly partitioned and generates less color histogram bins than previous literatures,thus cutting the computational cost.Texture histograms are generated through an improved direction coded local binary pattern descriptor and they can describe local texture distribution better.As for the geometric feature,the mean value and standard deviation of height estimates of multi-shot blobs are calculated.Superior to single-shot geometric analysis,these two statistic parameters are computed from multi-shot blobs and can suppress the disturbance from noise blobs.Besides,these two geometric parameters can easily reflect height and gait movement simultaneously.To our knowledge,it is original to extract such geometric features in a statistical way of this paper.

In human tracking process,an adaptive tracking mechanism is designed.It aims to automatically match and track individuals based on both retrospective and on-the-fly information. The image blobs are divided into two groups,namely,the gallery group gathered by source images and the probe group gathered by target images.The gallery group trains the computation parameters of the target person and then tests the image blobs in the probe group.On one hand,in the gallery group the appearance feature from each visual angle is described by a primary appearance model.This paper creates primary appearance model to represent unique appearance feature seen from unique visual angle.On the other,the probe blobs are divided into groups based on state-of-the-art single camera human tracking techniques.Each group corresponds to a unique person and includes the appearance information of him/her.The geometric features of the groups are first compared with those of the gallery group.The geometrically similar groups are subsequently described and arrayed in an appearance model matrix,and then matched with the gallery primary appearance models.The group that has the maximum label scale with gallery primary appearance models is determined to be the target person.After being targeted,the blobs of the person are automatically collected into source blob sequences and updated for obtaining new gallery primary appearance models.The mechanism obtains a powerful generalization capability among different cameras.Superior to the existed methods,it has an increasingly better performance over time in term of generalization.

1 Motion Detection

Eig.1 shows a configuration of the proposed method.This paper employs ViBe to detect and segment moving objects[18].After segmenting an object,an external bounding box is used to contain it.The object segmentation actually obtains a foreground blob.Then the height to weight(H/W)ratio of foreground blob is computed.If its H/W ratio is between 5 and 10,this foreground blob is recognized as a human foreground blob.Otherwise,it will be deleted.

2 Appearance Feature Extraction

2.1 Partial body analysis

Instead of focusing on global appearance features,this paper analyzes the appearance features from human body parts.The areas around the chest,thigh and foot are chosen as the three parts which describes the most critical information of individual,and thereby the feature extraction becomes faster whilst not losing important appearance information.In this paper,the chest corresponds to the 15%—40%region of the external bounding box,while the thigh corresponds to50%—70%and the foot 85%—100%.The chest weights 60%in term of importance,the thigh 25%and the foot 15%.

Eig.1 Process flow diagram for the proposed method

2.2 Appearance model

Appearance model contains color and texture information.Color expresses chromatic information and texture gives spatial distribution information of pixels.The combination of color and texture enhances the discriminability to distinguish different people.

2.2.1 Color analysis

Color often varies with the illumination change.To cope with this,the HSV color space is employed.In the HSV color space,the effect from illumination change can be suppressed by decreasing Value characteristic.Eor an initial redgreen-blue(RGB)image,it is first converted into HSV space[19].After executing the convert,H has values from 0 to 360,S from 0 to 1,and V from 0 to 255.

In most previous appearance models,each characteristic in color space generated a histogram and thus produced massive bins which would be time-consuming.Eor saving time,this paper partitions HSV color space into rough segments. The Hue characteristic is partitioned evenly into QHsegments,Saturation into QSsegments and Value into QVsegments.The values of H,S,V components are then converted into segment level HC,SC,VC,respectively.

Then HC,SCand VCare integrated into a color descriptorγHSVas follows

whereγH,γS,γVdenote the weight coefficients of HC,SC,VC,round(·)represents a rounding function.γH,γS,γVare computed by

Since QS,QVand QHare all larger than 1,the weight of ValueγVwill be always less thanγHandγS.It suppresses the impact of brightness and strengthens the robustness to illumination change.To avoid losing too much Saturation information,the segment numbers should be sorted in the descending order of QH>QS>QV.Here,the values of QH,QSand QVare 250,50 and 5,respectively.

2.2.2 Texture analysis

Local binary pattern(LBP)descriptor is oneof the widely-used texture descriptors.Comparing with general LBP,the direction coded LBP(d LBP)descriptor presented in Ref.[20]considers the relations between center pixel and neighboring pixels,as well as the relations among border pixels along one direction,and thus can better describe the texture in interesting regions.Different from Ref.[20],we converts the eight-bit d LBP binary code into a decimal descriptor as

where NS stands for the neighbor size and NS= 2NS′.vpis the value of pixel regularly spaced on circle and vcthe value of center pixel.

The d LBP in Eq.(4)involves not only comparison information between border pixels and center pixel but also comparison information among border pixels themselves along one direction,and it can rank the three pixel values along one direction.Therefore,it can better discriminate person′s appearance.

2.2.3 Appearance modeling

Person′s appearance feature is modeled by the pixels from chest,thigh and foot regions. Each pixel has its color and d LBPP,Rvalue,and each region will thus generate a color histogram and a d LBP texture histogram.The appearance model is constructed by concatenating six histograms from left to right in such an order:the color histogram of chest,the color histogram of thigh,the color histogram of foot,the texture histogram of chest,the texture histogram of thigh,and the texture histogram of foot.

3 Geometric Feature Extraction

3.1 Height estimation

Before calculating height estimate,apparent height should be measured at first.Apparent height is computed as the length from the middle top to the middle bottom of the bounding box. This paper calculates height estimate in the same way as Ref.[21].

3.2 Statistical geometric features

Statistical geometric features are obtained by computing the statistical parameters of height estimates.To suppress the impact from noise height estimates,the heights of blobs are sorted in a descending order,at the same time,the tallest and the shortest 5%will be deleted.The remaining height estimates compute the height estimate parameters.This paper uses mean and standard deviation of the remaining height estimates as the statistical geometric parameters.

Imagining that there are N blobs in training sample,and the height estimate of blob i is denoted as hi.After deleting the tallest and shortest 5%blobs,the remaining blobs can be ranked as hm,hm+1,hm+2,…,hM-1,hM,…,hn-2,hn-1,hnin a descending order,where n-m+1=90%· N.The statistical geometric parameters are calculated by

whereμhmeasures the stature of a person andσhreflects the rhythm up and down displacement of the upper body.These two parameters connect the height feature with the gait feature in a simple way.

3.3 Geometric similarity

Once the statistical geometric parameters have been obtained,the similarity of geometric features of source sequences(μs,σs)and target sequences(μt,σt)is computed as

whereωμ=ωσ=0.5,representing the weight of mean value and standard deviation respectively.Simμand Simσdenote the similarity of(μs,μt)and(σs,σt)respectively.

whereμ0is a constant related to real situations.If Sims,tis greater than threshold Tgeo,it proves that source sequence and target sequence are similar.

4 Human Tracking

Erom different visual angles,the same person may appear strikingly different,especially when the colors and textures of the front,side and back clothing are totally diverse.But as long as one appears under the same visual angle,his or her appearance models will be pair-camera correlated.This correlation can help us identify the same person from frame to frame and track him/ her in different views.Accordingly,authors collect the maximum appearance models from different visual angles,and compares them in the coming camera view with the collected models,so as to immediately target the person as soon as he/ she appears under a collected visual angle.

When a person walks into a specified visual angle,he/she may produce a blob sequence,which mirrors the appearance feature in this specified visual angle.Ideally,each sequence corresponds to a visual angle,and the amount of sequences is equal to the amount of visual angles. However,since the noise blob sequences are unavoidable,thus leading to noise appearance models,primary appearance models are selected in an adaptive mechanism to prevent the noise appearance models.

4.1 Primary appearance models

Eor a person,each of his/her blobs has an appearance model and all the appearance models are constructed in the method described in Section 2.2.3.The appearance models are denoted as mi(i=1,2,3,…,n)and will be divided into classes. Those classes with over-threshold model population are selected as primary appearance model classes.

4.1.1 Appearance model classification

Appearance model classification is based on the pair-wise distance of appearance models.The appearance model mihas six histograms,including three color histogramsand three d LBP histograms(j=1,2,3 corresponding to chest,thigh and foot).The distance of miand mi′is computed from the correlations of the six histograms in the following

The Bhattacharyya distance[10]is used to measure the correlation of histograms H1and H2.Then Dis(mi,mi′)is compared with a set threshold Tdis.If Dis(mi,mi′)is less than Tdis,mi,mi′are collected into the same class;otherwise,classified into different classes.

4.1.2 Primary appearance model class

All the appearance models are classified into N classes,denoted as MCi.Eor MCi,the amount of its inner models is Si.Considering that noise blobs will not exceed 10%of total blobs in most cases,primary appearance model class is the one whose inner models are more than 10%of total models.

4.1.3 Primary appearance model

Eor each primary appearance model class,it has a primary appearance model,which represents the appearance feature of the model class. Primary appearance model comes from the arithmetic operations of all the inner models.As each appearance model has six histograms(three color histograms and three texture histograms),the primary appearance model is co-determined by the arithmetic operations of these six histograms. Each of the six histograms computes an arithmetic mean histogram,described quantitatively in Eq.(11).

4.2 Person re-identification

Relying on the state-of-the-art tracking techniques based on spatial-temporal relations[23],the blobs in single camera could be classified into different groups.Each group corresponds to a unique person.According to temporal relation,an individual′s blobs can be arranged as a sequence in time order.This sequence includes all the appearance information of an individual under all visual angles when one passes through a camera.Accordingly,each camera will have several blob sequences,and the amount of sequences is equal to the amount of individuals″seen″by the camera.Each blob sequence consists of some blobs and each blob has an appearance model.All the appearance models in camera C can be arrayed in a matrix MC

Row i lists the blob sequence of person i composed of liblobs.mCijstands for the appearance model of blob j in the blob sequence of person i.The column amount of MCis equal to the length of the largest blob sequence.Those smaller rows are filled with zero.

The statistical geometric parameters of target person are initially computed from the gallery blobs.The sequence in matrix MCwhich is similar to the sequence of target person will be geometrically eligible.

The geometrically eligible blob sequences are further tested by the primary appearance models denoted asin this paper,whererepresents the primary appearance model of class,andthe amount of its inner models.

In order to determine whetherin matrix MCcan be gathered into,sign function Dsignis designed as

The class label determination matrix reads how many class labels each blob has when matched with primary appearance model classes MCP.The accumulation of row i,denoted as LSiand calculated in Eq.(15),defines the label scale of blob sequence of person i when matched withthe primary appearance models.

where liis the length of blob sequence of person i.Sometimes,ifhas below-threshold distances with several different primary appearance models synchronously,it can be classified into several classes.Hence,the blob will have several different labels.In order to sign if a blob has at least one label,sign function lbsignis introduced.

A blob sequence with larger LBN means that more blobs in the sequence will be gathered into primary appearance model classes,but it does not mean it is more likely to be the target sequence. In fact,although those sequences containing more blobs tend to have larger LBNs,they are also likely to have more unlabeled blobs that do not belong to any primary appearance model class.In light of this,an appearance based blob sequence sign function ESsignis designed to determine if a sequence is an appearance based blob sequence.

where ESsign(LBNi)=1 indicates that the blob sequence of person i is an appearance eligible blob sequence.

The appearance eligible blob sequence of person t in matrix MC,which has the maximum label scale,is re-identified as the target sequence,and person t is re-identified as the target person, namely

4.3 Update of primary appearance models

The blobs of target person t captured from the newest camera have the newer information than those from foregoing cameras.Primary appearance models should be updated to ensure the accuracy of continuous disjoint tracking.

Eor a blob in sequence of target person t in new camera,if its minimum distance with old primary appearance models is shorter than the threshold,its appearance model will be collected into the closest primary appearance model classOtherwise,it will not be collected into any existed class but generate a new appearance model class itself.The new appearance model class is labeled asand its inner model number is

The appearance models in old primary model classes and new model classes renew the primary appearance models in the following sign functions

5 Experimental Results

5.1 Experimental setup

The experiments are successively conducted in both well-known benchmark VIPeR dataset and real complex scenarios.Eirstly,this paper utilizes the images in VIPeR to test the effectiveness of human re-identification based on the improved appearance feature descriptor.The images in the dataset are randomly splitted into two sets:Camera A and Camera B.It is the most challenging dataset currently available for single-shot pedestrian re-identification.Secondly,the effectiveness of human tracking is tested in real complex scenarios composed of seven non-overlapping views,which may witness many vast angle changes and illumination alters.The scenarios reflect closer to real-life and reveal the toughness of disjoint video surveillance.Each view in complex scenarios captures many continuous images and enables a multi-shot analysis.

5.2 Experiments in VIPeR dataset

In the experiments of this paper dataset is splitted into 10 random sets and the average of the 10 experimental results are compared with those of other two excellent benchmark methods,namely Gray′s ELE 200[22]and Chen′s Adaboost classifier based on multiple features[23].The 10 experimental results are concluded from 10 random sets of 200 pedestrians.The cumulative matching characteristic(CMC)curves are depicted in Eig.2.

Eig.2 CMC comparison of performance on VIPeR dataset

Eig.2 demonstrates the proposed feature representation has the stronger discriminative power than other two state-of-the-art methods.It declares an excellent performance of single-shot people re-identification based on the improved local uniformly-partitioned HSV color descriptor and the improved d LBP texture descriptor.

Eurther experiments conducted on VIPeR dataset test the computation time of the proposed appearance features.The result is compared with other two human tracking approaches,namely,Chen′s[23]and Hyun-Uk′s[2].All the normalized 632 images in Camera A are chosen to test the computation time of each approach.The comparison of computation time is illustrated in Eig.3.It infers that the proposed method attracts appearance features faster than the other two excellent algorithms.The real-time feature extraction underpins and ensures the multi-shot human tracking in term of computational speed.

Eig.3 Comparison of computation time for feature extraction

5.3 Experiments in complex scenarios

The complex scenario is chosen in the experiments.It aims to verify the continuous tracking across non-overlapping cameras.In the beginning of the track,the gallery group needs to be first initialized.

5.3.1 Gallery group initialization

The gallery group initialization is to choose initial source images to train the parameters of the target person.Here three experiments are conducted when initializing the gallery group. The first two use the images captured from single shooting angle,while the third one uses the images captured from multiple visual angles in the initialization.The three experiments intend to prove the importance of collecting feature information from multiple visual angles.

In the first two experiments,Camera 2 is the gallery camera and its first 168 images completethe initialization of the gallery group with only one visual angle.

The first experiment selects Camera 3 as the probe camera,which has a similar visual angle to the gallery Camera 2.Persons 1,2,3 appear in Camera 3,whose blob sequences′statistical geometric parameters are listed in Table 1.The similarity of sequence of Person 1 with the gallery group is 48.8%,less than the threshold 75%,indicating that Person 1 differs from the gallery target in term of geometric features and it will be deleted.The label amount of each frame in the sequences of remaining Person 2 and Person 3 is illustrated in Eig.4(a),where Person 2 has a larger label scale(the sum of the label amount of each frame)and its LBN is also manifestly larger than 30%of total sequence.Therefore,Person 2 is determined as the target person.

Table 1 Geometric parameters of blob sequences in Camera 3

Camera 6 is chosen as the probe camera in the second experiment,which monitors the same three persons as Camera 3 and has vast angle difference with the gallery Camera 2.Eig.4(b)shows that when Person 2 and Person 3 come into Camera 6,they almost have no label scale.No one can be truly tracked.It means tracking failure appears in Camera 6.

In the third experiment Camera 4 is regarded as the gallery camera.It initializes the gallery group in multiple shooting angles using the first 245 images.Eig.5 shows the EOV in Camera 4,displaying a wide view field.Eig.5 also lists four primary appearance models of the target Person 2.

Just like the second experiment,Camera 6 is also selected as the probe camera in the third experiment.In contrast,the label scales in the third experiment(Eig.6)dramatically outperform those of the second in Eig.4(b).

Eig.4 Label scales of Persons 2,3 in Cameras 3 and 6 when Camera 2 initializes the gallery group

Eig.5 EOV of Camera 4 with multiple visual angles and primary appearance models of Person 2

5.3.2 Continuous tracking

Another group of experiments is conducted in six non-overlapping cameras to validate continuous human tracking across non-overlapping cameras.Eig.7 lists six EOVs in these experimental cameras.Each camera shoots in a unique angle and has different illumination conditions.Eig.7 demonstrates the effectiveness of human tracking across six non-overlapping cameras.

Eig.6 Label scales of Persons 2,3 in Camera 6 when Camera 4 initializes the gallery group

Eig.7 Human tracking across six disjoint cameras

Eig.8 shows the label amount of each frame from Camera 2 to Camera 7.The frames in Eig.8 are arranged according to the time order.The label amount of each frame tends to increase from 0 to 4,which means more and more appearance features have been captured and stored in the gallery group.

Eig.9 shows the performance parameters from Camera 2 to Camera 7,where matching rate represents the ratio of matched frames to all the analyzed images,and erroneous matching rate means the ratio of incorrectly matched frames to all the matched frames.The accurate matching rate increases over time,which infers that human tracking can accurately track the target people when one appears under different visual angles as well as in different view conditions.The up-andup performance implys that the generalization capability generalizes across different cameras over time.The erroneous matching rate almost remains unchanged,and even gets worse over time. It is mainly due to the fact that incorrect blob is wrongly matched with primary appearance models and the increase of primary appearance models raises the risk of erroneously re-identifying a wrong blob.However,as the erroneous matching rate is just around 10%,it is still acceptable.

Eig.8 Label scale of each frame from Camera 2 to Camera 7

Eig.9 Performance parameters of human tracking from Camera 2 to Camera 7

The experimental results in complex scenarios prove the feasibility of this adaptive human tracking mechanism across non-overlapping cameras.In the mechanism,the gallery group initialization is important.When the gallery group is initialized at multiple angles,it is more likely to successfully track the person in the probe camera since more appearance information is captured under different visual angles.Eig.8 depicts the in-crease of label amount of each frame over time. Essentially,it proves that the re-identification ability improves when more primary appearance models are stored in the system.The update of gallery group renews the primary appearance models camera by camera,thus establishing an adaptive human tracking mechanism in non-overlapping cameras.

6 Conclusions

An adaptive human tracking approach based on primary appearance model and statistical geometric features is proposed to track human across disjoint cameras in depression angles.All the extracted features manage to keep robust to illumination variations,foreground errors,as well as visual angle changes.The local uniformly-partitioned HSV color features are extracted in realtime.The combination of appearance and statistical geometric features produces a discriminative and robust feature representation.The human tracking mechanism is the main contribution.It uses both retrospective and on-the-fly information to collect the maximum appearance information captured from different visual angles.The update of primary appearance models enables the human tracking mechanism to renew adaptively.In this adaptive mechanism,the later camera will gain a higher hit rate and its generalization capability will become better.The experiments conducted in benchmark dataset show the excellent accuracy and real-time extraction of feature representation. The experiments conducted in complex scenario prove the good generalization capability of the proposed mechanism and also show good performance in resisting inter-camera and intra-camera variations.

Acknowledgements

This work was funded by the Natural Science Eoundation of Jiangsu Province(No.BK2012389),the National Natural Science Eoundation of China(Nos.71303110,91024024),and the Eoundation of Graduate Innovation Center in NUAA(Nos.kfjj201471,kfjj201473).

[1] Doretto G,Sebastian T,Tu P,et al.Appearancebased person re-identification in camera networks: problem overview and current approaches[J].J Ambient Intelligence and Humanized Computing,2011(2):127-151.

[2] Chae Hyun-Uk,Jo Kang-Hyun.Appearance feature based human correspondence under non-overlapping views[C]∥Proceedings of 5th International Conference on Intelligent Computing,Emerging Intelligent Computing Technology and Applications.Ulsan: Springer,2009:635-644.

[3] Zeng Eanxiang,Liu Xuan,Huang Zhitong,et al. Robust and efficient visual tracking under illumination changes based on maximum color difference histogram and min-max-ratio metric[J].J Electron Imaging,2013,22(4):043022.

[4] Lee Seok-Han.Real-time camera tracking using a particle filter combined with unscented Kalman filters[J].J Electron Imaging,2014,23(1):013029.

[5] Wu Yiquan,Zhu Li,Hao Yabing,et al.Edge detection of river in SARimage based on contourlet modulus maxima and improved mathematical morphology[J].Transactions of Nanjing University of Aeronautics and Astronautics,2014,31(5):478-483.

[6] Madden C,Piccardi M.A framework for track matching across non-overlapping cameras using robust shape and appearance features[C]∥Proceedings of IEEE Conference on Advanced Video and Signal Based Surveillance.London:IEEE,2007:188-193.

[7] Zhang Chao,Wang Daobo,Earooq M.Real time tracking for fast moving object on complex background[J].Transactions of Nanjing University of Aeronautics and Astronautics,2010(4):321-325.

[8] Takayuki Hori,Jun Ohya,Jun Kurumisawa.Identifying a walking human by a tensor decomposition based approach and tracking the human across discontinuous fields of views of multiple cameras[C]∥Proceedings of Conference on Computational Imaging VIII.San Jose:SPIE,2010:75330.

[9] Lin Yu-Chih,Yang Bing-Shiang,Lin Yu-Tzu,et al. Human recognition based on kinematics and kinetics of gait[J].J Medical and Biological Engineering,2010(31):255-263.

[10]Trevor Montcalm,Bubaker Boufama.Object intercamera tracking with non-overlapping views:A new dynamic approach[C]∥Proceedings of 2010 Canadian Conference Computer and Robot Vision.Ottawa:IEEE,2010:354-361.

[11]Prosser B,Zheng W,Gong S,et al.Person re-identification by support vector ranking[C]∥Proceedings of British Machine Vision Conference.[S.l.]:BMVA Press,2010:21.1-21.11.

[12]Bazzani L,Cristani M,Perina A,et al.Multipleshot person re-identification by chromatic and epitomic analyses[J].Pattern Recognition Letters,2012(33):898-903.

[13]Zheng W,Gong S,Xiang T.Person re-identification by probabilistic relative distance comparison[C]∥Proceedings of IEEE Conference Computer Vision and Pattern Recognition.Colorado:IEEE,2011: 649-656.

[14]Avraham T,Gurvich I,Lindenbaum M,et al. Learning implicit transfer for person re-identification[C]∥Proceedings of European Conference on Computer Vision.Elorence:Springer,2012:381-390.

[15]Hirzer M,Beleznai C,Roth P,et al.Person re-identification by descriptive and discriminative classification[C]∥Proceedings of 17th Scandinavian Conference on Image Analysis.Ystad:Springer,2011:91-102.

[16]Zheng W,Gong S,Xiang T.Re-identification by relative distance comparison[J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2013(35):653-668.

[17]Layne R,Hospedales T M,Gong S.Domain transfer for person re-identification[C]∥Proceedings of ARTEMIS Workshop at ACM Multimedia.Barcelona:[s.n.],2013:25-32.

[18]Yang Chenhui,Kuang Weixiang.Robust foreground detection based on improved vibe in dynamic background[J].International Journal of Digital Content Technology and Its Applications(JDCTA),2013(7):754-763.

[19]Gonzalez R C,Woods R E.Digital image process[M].3rd Ed.Beijing:Publishing House of Electronics Industry,2010.

[20]JiríTrefny,JiríMatas.Extended set of local binary patterns for rapid object detection[C]∥Proceedings of the Computer Vision Winter Workshop.Nove Hrady:Czech Pattern Recognition Society,2010:37-43.

[21]Dai Xiaochen,Payandeh Shahram.Geometry-based object association and consistent labeling in multicamera surveillance[J].Emerging and Selected Topics in Circuits and Systems,IEEE Journal on,2013,3(2):175-184.

[22]Douglas G,Hai Tao.Viewpoint invariant pedestrian recognition with an ensemble of localized features[C]∥Proceedings of the 10th European Conference on Computer Vision.Marseille:Springer,2008:62-275.

[23]Chen Xiaotang,Huang Kaiqi,Tan Tieniu.Object tracking across non-overlapping cameras using adaptive models[C]∥Proceedings of ACCV 2012 International Workshops on Computer Vision.Daejeon: Springer,2012:464-477.

(Executive editor:Zhang Tong)

TP391.41 Document code:A Article ID:1005-1120(2015)01-0048-13

*Corresponding author:Shao Quan,Associate Professor,E-mail:shaoquan@nuaa.edu.cn.

How to cite this article:Shao Quan,Liang Binbin,Zhu Yan,et al.Adaptive human tracking across non-overlapping cameras in depression angles[J].Trans.Nanjing U.Aero.Astro.,2015,32(1):48-60.

http://dx.doi.org/10.16356/j.1005-1120.2015.01.048

(Received 13 November 2014;revised 7 January 2015;accepted 12 January 2015)

猜你喜欢
陈涛
Simulation of crowd dynamics in pedestrian evacuation concerning panic contagion: A cellular automaton approach
封二 春姑姑走啦
猴爸爸的百宝箱
陈涛吉祥物设计作品选登
急诊内科治疗老年重症心力衰竭临床疗效分析
扶贫猪
一道模考题引发的思考
让错误开出美丽的花朵
精神病新娘
女市长巧治“懒老公”