999精品在线视频,手机成人午夜在线视频,久久不卡国产精品无码,中日无码在线观看,成人av手机在线观看,日韩精品亚洲一区中文字幕,亚洲av无码人妻,四虎国产在线观看 ?

An Adaptive Superpixel Tracker Using Multiple Features

2019-11-25 10:22:30JingjingLiuBinZhangXuChengYingChenandLiZhao
Computers Materials&Continua 2019年9期

JingjingLiu,BinZhang,XuCheng,YingChenandLiZhao

Abstract:Visual tracking is a challenging issue in the field of computer vision due to the objects’ intricate appearance variation.To adapt the change of the appearance,multiple channel features which could provide more information are used.However,the low level feature could not represent the structure of the object.In this paper,a superpixel-based adaptive tracking algorithm by using color histogram and haar-like feature is proposed,whose feature is classified into the middle level.Based on the superpixel representation of video frames,the haar-like feature is extracted at the superpixel level as the local feature,and the color histogram feature is applied with the combination of background subtraction method as the frame feature.Then,local features are clustered and weighted according to the target label and the location center.Superpixel-based appearance model is measured by using the sum of the voting map,and the candidate with the highest score is selected as the tracking result.Finally,an efficient template updating scheme is introduced to obtain the robust results and improve the computational efficiency.The proposed algorithm is evaluated on eight challenging video sequences and experimental results demonstrate that the proposed method can get better performance on occlusion,illumination variation and transformation.

Keywords:Superpixel tracker,multiple features,adaptive update.

1 Introduction

Precise objection tracking is the most substantial component in a wide range of applications like activity analysis,visual surveillance,human computer interaction and medical imaging.With the development of high-powered computers and cheap high quality video cameras,the progress of object tracking techniques is fast in recent years.However,tracking objects is complex due to noise in images,partial and full object occlusions,scene illumination variation,real-time processing requirements and so on.Zhenguo et al.represent the object by low dimensional compact shape and color feature[Gao,Xia,Zhang et al.(2018)].In this paper,the target and its surroundings are represented by using small regions,called superpixels [Ren and Malik (2003)].Superpixels are generally defined as contracting and clustering uniform pixels in the image,which are local,coherent.Superpixels have been broadly used in various vision applications such as image segmentation,object recognition,anomaly detection and object tracking.The predominant integrity of superpixels provides a more natural and meaningful representation.Numerous works have been applied superpixels known as middle level visual cue and pleasant consequences have been obtained.Image superpixel segmentation approach using Lazy random walk (LRW) algorithm has been presented by Shen,which achieves the better performance by relocating the center positions of superpixels and splitting the large superpixels into small ones [Shen,Du,Wang et al.(2014)].Xuan constructs a video representation with a superpixel-based bag- of -features (B of) model for action recognition which includes histograms of optical flow (H of),histograms of oriented gradients (HOG),and motion boundary histograms [Dong,Tsoi and Lo (2014)].Superpixel-based statistical anomaly detection for sense and avoidance was presented by Pappas,which is based on superpixel image segmentation combined with subsequent statistical analysis and anomaly detection,can be applied to solve the problem of visually detecting possible aircraft collisions [Giordano,Murabito,Palazzo et al.(2015)].

In terms of object tracking based superpixels,significant advances have been witnessed with the development of adaptive algorithms and productive applications.Robust superpixel tracking (SPT) adopts a discriminative model based on superpixels to distinguish the target from the background,which can obtain the best candidate by maximum a posterior estimation in the target-background confidence map [Yang,Lu and Yang (2014)].Concerning to estimate the target position,Junqiu employs multiple hypotheses for superpixel matching and projects the matching results onto a displacement confidence map [Wang and Yagi (2014)].In consideration of recovering the object from the drifting scene,Xu proposes robust superpixel tracking with weighted multipleinstance learning which achieves robust and accurate performance [Cheng (2015)].Apart from making efforts on the discriminative appearance model,Yuxia proposes a superpixel tracking method via graph-based hybrid discriminative-generative appearance model to deal with occlusion and shape deformation [Wang and Zhao (2015)].

Though the aforementioned works can be efficient,these methods are hard to deal with the tracking adaptively.In this paper,our approach takes into account for the content and quality of the tracking objects and applies various update frequency according to the different sequences.In addition,the superpixel haar-like feature and the region of interest deviating from background subtraction are integrated to build the appearance model,which uses the local feature of the superpixel and the global feature of the region of interest.Moreover,the updating policy ensures the update frequency is adaptive.

The remainder of this paper is organized as follows.In Section 2,the proposed method is described.Section 3 presents the experiment which include experiment setups and experiment results.Then,the conclusion is made in Section 4.

2 Proposed algorithm s

The proposed image representation scheme and tracking algorithm will be described in this section.

2.1 Superpixel based appearance model

In this paper,the mid-level cue with structural information captured in superpixels and saliency feature with global color information derived from background subtraction are utilized to represent the object.Each input video frame is partitioned into a set of superpixels,which usually have more regular and compact shape with better boundary adherence.Superpixels are utilized as the primitives for the superpixel-level haar-like feature and combined with the background subtraction.

To construct the superpixel-level feature,Ntsuperpixels are achieved by segmenting the surrounding region of the object from a set ofntraining frames using the LSC (Linear Spectral Clustering) algorithm [Li and Chen (2015)].LSC produces compact and uniform superpixels with low computational costs,which terms asThen haar-like feature is applied to represent the superpixels,and noted aswhereandare the mean and the covariance of the regionr’scolor features,respectively.After collecting the superpixels of all the training frames,mean shift clustering algorithm is employed on the feature pooland thennvarious clusters are obtained.Each clusterhas its unique cluster centercluster radiusand its cluster members.Each cluster has its own scope to locate and corresponds to the particular place in the training frames.Among the clusters,weightis necessary to assign which is defined as

The clusters are sorted in descending order according to the cluster’s weight value and then assemble their cluster center factoras a dictionaryD.

Color histograms are extracted at frame level as global feature for appearance model.We only use the Mark area to calculate color histogram rather than the whole frame to reduce the computing cost.The algorithm transfers the image from the RGB (red,green,blue)color space to HSV (hue,saturation,value) color space and quantifies Hue space to 16 bins noted as a vector.Then the difference between two frames is the Euclidean distance between the corresponding vectors named asV.

2.2 Combining vote maps

When the new frame reaches,a surrounding region of the target is extracted and segmented into superpixels,and then the Mark area is labeled by the ViBe algorithm.The superpixels in the same cluster are scored according to 3 factors which are the cluster it belongs to,the distance between this superpixel and its corresponding cluster center,and the cluster’s label according to background or target.Additionally,clusters have weights in the feature space.The new frame’s cluster center is the other factor to affect the superpixel’s score.Therefore,the score of each superpixel is computed as follows:

whereω(r,i)enotes the weighting factor based on the distance between the cluster center in the new superpixel and the cluster center in the feature tool.λlis a penalty factor of the cluster and its value is decided by the specified regions which are the target,background and Mark area.By taking the three factors into account,is the score for superpixelrat thet-thframesp(t,r).

The confidence map for each pixel on the current frame is obtained based on the contribution.The pixel in the superpixelsp(t,r)is assigned with the superpixel scoreand the others are set to 0.After constructing the confidence map,the following will discuss the decision procedure of the likely location and the updating policy.

2.3 Motion model

Because of the short interval of the inter-frame,the displacement and the scope of the target in the current frame must move and change smoothly.Here the motion model is assumed as Gaussian distributed:whereФdenotes a diagonal covariance matrix,which represents the standard deviations for target’s location center and scale change.The most suitable candidate is obtained under the particle filter framework,and the mean of the random particles is the position of target’s center in the last frame.The candidate with the highest score is selected as the position of the current frame’s target among all the candidates.Candidates’ scores are measured using the weights for each pixel according to each superpixel and the difference between the histogram of color,which can described as following:

whereis the number of the pixels for each superpixel,Sum(k)denotes the score for a certainkcandidate andVrepresents as the difference of the color histogram between the current frame and the last frame.After all,the candidate with the highest score would be chosen and recorded its region and center as the tracking result.

2.4 Appearance model updated with occlusions

Updating the appearance mode is a significant and essential element in the tracking system,which can influence the efficiency,robustness and calculation speed.Updating the object appearance everyUframe is frequently-used whose performance is affected by the amount of the frame number.In order to acquire a compromise between the tracking errors and computational cost,a toy experiment is performed and is shown in the Fig.1.Based on the information in the Fig.1,tracking error is the least in most situations when the update interval is 1.However,the tracking error is too high when the frame number is greater than 300 and the frame number is from 200 to 220.For the update interval with 10 and 30,the tracking errors fluctuate and their mean values are roughly the same especially when the frame number is greater than 200.However,the average per-frame run time is increasing with the grow th of the update interval.In general,a constant update interval is set based on the experiment.However,constant update interval is silliness and adaptive updating scheme which is more effective is applied.

Figure1:The relationship of tracking errors and update interval

The policy of updating the model is according to threefold factors.Firstly,the score of the candidate is an essential aspect and the score’s tendency is noticed instead of the score itself.When the tendency of score is increasing or decreasing,the appearance model is updated.If the score fluctuates along with the arrival of the new frame,the model is kept.Secondly,the model is updated according to the deviation of the color histogram.If the difference is greater than 10% of the quadratic sum of the histogram for the origin model,the model updates and makes a caution as occlusion.If the difference is more than 20%,the object is considered as occluded.Thirdly,the updating frame is not accumulate the series of the e adjacent frames,the update is applied using the discrete frames but the adjacent continues frames,which inspired by the coding of the MPEG(Moving Picture Experts Group).For example,the 40thframe need update,the information of the frame 33th,35th,37th,38th,39th,40thare accumulated rather than the information from frame 33thto 40th.In this case,the information acquired from the discrete frames are more than the information from continues frames with the equal amount.Additionally,in order to guarantee the model’s effectiveness,the model is forced to update every 30 frames,which is an empirical value.In addition,the origin model is reserved for all the tracking schemes.

2.5 The proposed tracking algorithm

The proposed whole tracker would be presented in this part,which is summarized in the Algorithm 1.

Algorithm 1.Main steps of the e proposed tracking algorithm Initialization:For t=1 to m (e.g.,m=4) 1.Initialize arguments of the e proposed algorithm in frame t.2.Segment the surrounding region and extract their features.End Discriminative appearance model is achieved by Eq.(1) and Eq.(2).Tracking:For t=m+1 to the end of the sequence 1.Segment a surrounding region and extract their features.2.Compute the voting map using Eq.(3) and Eq.(4).3.Sample N candidates with the voting map.4.Select the candidate with the highest score as the tracking result.5.Detect the occluded station.6.Update the appearance model using our specified update scheme.End.

3 Experiments

The experimental setups and the exploratory results are depicted in this section.The algorithm is conducted in MATLAB and runs on Inter Core 2 Duo 2.93 GHz CPU with 2.96 GB memory.

3.1 Experimental setups

We utilize haar-like feature and information derived from background subtraction for each superpixel.The LSC algorithm is applied to segment frames into superpixels where the parameters are set as default values.The ViBe was called using the value in this paper:N=20,R=20,minimum number=2,Ф=16.To build the training dataset for initialization,the first 4 frames are processed using the simple tracker.The threshold for detecting occlusion is set to 0.2 and the forced updating frame is 30.

The proposed algorithm is evaluated on 8 challenging sequences with 8 different methods.The sequences chosen for evaluating the algorithms contain challenges like occlusion,the change of illumination and large movement of the target.The compared algorithms include IVT [Ross,Lim,Lin et al.(2008)],VTD [Kwon and Lee (2010)],M IL [Babenko,Yang and Belongie (2009)],L1 [Bao,Wu,Ling et al.(2012)],TLD[Kalal,M ikolajczyk and Matas (2012)],Frag [Adam,Rivlin and Shimshoni (2006)],SPT[Yang,Lu and Yang (2014)],and Cheng [Cheng (2015)],whose code can be obtained from the authors’ homepage.Then,for completeness of the e analysis,the proposed tracker is performed on the recent dataset VOT2014 [Kristan,Leonardis,Matas et al.(2015)],which contains 25 sequences.In the VOT2014,there are 38 trackers including generative trackers and discriminative trackers.

3.2 Experimental results

In the first experiment,8 sequences are used to evaluate the proposed algorithm and these are:bird2,basketball,deer,David,faceoce2,shaking,singer1,and woman.For fair comparison,the parameters of each tracker remain consistent with the code provided by the authors.The average result of 3 runs are adopted as the comparative results,or the results are obtained straightforwardly from the published papers.The two criterions which are center location errors and successful tracking frames are used in this paper,which are shown in Tab.1 and Tab.2.Center location error is the average of Euclidean distance between the tracking result and the ground truth and successful tracking frames are the count of successful frame based on evaluation metric of the PASCAL VOC object detection [Everingham,Van,Williams et al.(2010)].For presentation clarity,the top 2 tracking for each sequence are boldface.It is manifest from the tables that our algorithm achieves the better or similar performance comparing with the other trackers.

In the basketball sequence,the M IL,IVT and L1 methods drift to background area because of the deficient design of non-grid deformation.Our method achieves the second best results and deviates a little from the result of SPT.The Bird2 sequence has the nongrid deformation,pose change and occlusion,our algorithm and SPT works roughly the same.The algorithm from Cheng works the third best in the Bird2 sequence and we owe the success to the good attributes of Superpixels.In Fig.2(c),David Indoor sequence involves illumination and pose variations.L1 tracks the object in the small rectangle accurately because of its precise design for distance change.Although the LI and Cheng trackers achieve better result in this frame,our method achieves the minimum center location error which is shown in the Tab.1.

Table1:Tracking results:center location error (pixel)

Table2:Tracking results:successful tracking frames

To the problem of occlusion and heavy motion in the sequences of woman,deer,and faceocc2 (Figs.2(d)-2(f)),the algorithms from Cheng and ours perform well,whose results are almost equal.Yet VTD tracker acquires the best result in the Deer sequence.The reason is that it does not distinguish the target from the background and considers some background pixels as parts of the e target,thereby getting good results because of the inherent attribute in the sequence.On the other hand,Cheng and our tracker build the appearance model with the target area and miss the similar information of the target.But the trackers are more robust in the other sequences.

Shaking sequence contains illumination variation and pose change with the head shaking.Singer1 clip includes illumination and scale variations which make the trackers drift.As it is shown in the Fig.1,VTD,L1,and TLD can track the object quite well except for some errors in the Singer1,ours can also achieve the promising results and works almost in a bar with the design from Cheng.In the shaking sequence,our tracker is better than the traditional trackers which are less reliable except for VTD and Cheng’s.

Figure2:Tracking results with comparisons to the state of the art trackers

Figure3:The accuracy-robustness ranking plots with respect to the two experiments

The comparison in the VOT2014 is displayed by visualizing the ranking results in the Fig.3.Each tracker’s AR rank is explained as a point which stands for its independent scores for accuracy and robustness.The experiments with the overall performance in the Fig.3 contain the baseline and region noise experiments.A tracker is determined as good when it is close to the top-right corner.Except for the overall evaluations,per-visual-attribute experiments which contain camera motion,illumination change,motion change,occlusion,size change and no degradation are shown in the Fig.4.The proposed tracker shows superior results under the situation of baseline and region noise,although it is not the top results in the comparison.The top trackers are correlation filter-based algorithm or deep learning algorithm while ours performs well in comparison with the state- of -theart algorithms.For the videos with occlusion attribute,the proposed tracker ranks 1st.For the motion change and camera motion attributes,our tracker rank 5th.Generally speaking,ours cannot reach the top but gives a comprehensive result and shows great in the traditional trackers.

Figure4:The accuracy-robustness ranking plots with respect to different attributes

4 Conclusion

In this paper,an adaptive superpixel-based tracking approach is proposed.We build an effective appearance model with superpixel which provides the flexible and effective middle level cues and background subtraction which supports the interesting area.It distinguishes the foreground target from the background.The unique updating policy guarantees the adaptive performance.The tracking algorithm could handle the difficulties like occlusion,transformation and illumination change.The comparison results and evaluations demonstrate the proposed tracker performs robustly comparing with the state of -the-art trackers.

Acknowledgement:This research project is funded in part by National Natural Science Foundation of China (Nos.61673108;61571106).The authors would like to thank the anonymous reviewers and the associate editor for their valuable comments they provided.

主站蜘蛛池模板: 亚洲国产日韩在线成人蜜芽| 91美女在线| 国产自在自线午夜精品视频| 一级毛片网| 国产成人综合网在线观看| 国产亚洲精品自在线| 92午夜福利影院一区二区三区| 久久久久久久蜜桃| 18禁影院亚洲专区| 国产乱人乱偷精品视频a人人澡| 亚洲欧美另类中文字幕| 欧美性猛交一区二区三区| 久久精品亚洲热综合一区二区| 免费jizz在线播放| 国产在线精品人成导航| av在线无码浏览| 国产成人高精品免费视频| 国产99视频精品免费观看9e| 亚洲欧美精品日韩欧美| 色网站免费在线观看| 人人看人人鲁狠狠高清| a级毛片在线免费观看| 五月天天天色| 亚洲AⅤ综合在线欧美一区| 国产永久免费视频m3u8| 国产成人成人一区二区| 国产无码高清视频不卡| 99re在线免费视频| 成人日韩视频| 国产精品成人观看视频国产 | 国产亚洲高清在线精品99| 99久久人妻精品免费二区| 久久久成年黄色视频| 老司机久久99久久精品播放| 亚洲综合国产一区二区三区| 激情视频综合网| 在线观看欧美国产| 国产色婷婷| 2021国产精品自产拍在线观看| 亚洲国产天堂久久综合| 国产手机在线ΑⅤ片无码观看| 日韩经典精品无码一区二区| 在线看国产精品| 天天躁夜夜躁狠狠躁躁88| 波多野结衣一区二区三区AV| 中文成人在线| a天堂视频| 亚洲h视频在线| 国产欧美一区二区三区视频在线观看| 亚洲视频四区| 在线观看欧美精品二区| AV无码一区二区三区四区| 久久夜色精品| 亚洲娇小与黑人巨大交| 欧美日韩国产综合视频在线观看| 黄色污网站在线观看| 在线观看91精品国产剧情免费| 99热精品久久| 国产三级精品三级在线观看| 亚洲国产日韩一区| 欧美亚洲日韩中文| 欧美午夜性视频| 日韩AV手机在线观看蜜芽| 久久不卡国产精品无码| 黄色网页在线播放| 久久婷婷国产综合尤物精品| 日韩小视频在线播放| 毛片基地视频| 日韩欧美高清视频| 国产亚洲精品精品精品| 欧美一级高清免费a| 亚洲精品少妇熟女| 国产精品成人一区二区不卡| 午夜啪啪福利| 国产噜噜噜| 婷婷中文在线| 欧美国产日韩另类| 亚洲h视频在线| 亚洲床戏一区| 在线观看亚洲成人| 婷婷丁香在线观看| 日韩福利在线观看|