999精品在线视频,手机成人午夜在线视频,久久不卡国产精品无码,中日无码在线观看,成人av手机在线观看,日韩精品亚洲一区中文字幕,亚洲av无码人妻,四虎国产在线观看 ?

A Multi-Feature Learning Model with Enhanced Local Attention for Vehicle Re-Identification

2021-12-15 07:08:54WeiSunXuanChenXiaoruiZhang1GuangzhaoDaiPengshuaiChangandXiaozhengHe
Computers Materials&Continua 2021年12期

Wei Sun,Xuan Chen,Xiaorui Zhang1,,Guangzhao Dai,Pengshuai Chang and Xiaozheng He

1Jiangsu Collaborative Innovation Center of Atmospheric Environment and Equipment Technology,Nanjing University of Information Science&Technology,Nanjing,210044,China

2School of Automation,Nanjing University of Information Science&Technology,Nanjing,210044,China

3Engineering Research Center of Digital Forensics,Ministry of Education,Jiangsu Engineering Center of Network Monitoring,School of Computer and Software,Nanjing University of Information Science&Technology,Nanjing,210044,China

4Rensselaer Polytechnic Institute,Troy,NY,12180,USA

Abstract:Vehicle re-identification(ReID)aims to retrieve the target vehicle in an extensive image gallery through its appearances from various views in the cross-camera scenario.It has gradually become a core technology of intelligent transportation system.Most existing vehicle re-identification models adopt the joint learning of global and local features.However,they directly use the extracted global features, resulting in insufficient feature expression.Moreover,local features are primarily obtained through advanced annotation and complex attention mechanisms, which require additional costs.To solve this issue,a multi-feature learning model with enhanced local attention for vehicle re-identification(MFELA) is proposed in this paper.The model consists of global and local branches.The global branch utilizes both middle and highlevel semantic features of ResNet50 to enhance the global representation capability.In addition,multi-scale pooling operations are used to obtain multiscale information.While the local branch utilizes the proposed Region Batch Dropblock (RBD), which encourages the model to learn discriminative features for different local regions and simultaneously drops corresponding same areas randomly in a batch during training to enhance the attention to local regions.Then features from both branches are combined to provide a more comprehensive and distinctive feature representation.Extensive experiments on VeRi-776 and VehicleID datasets prove that our method has excellent performance.

Keywords: Vehicle re-identification; region batch dropblock; multi-feature learning; local attention

1 Introduction

Vehicle re-identification (ReID) is a core technology of intelligent transportation systems.With the development of artificial intelligence and big data technology, vehicle re-identification has a wide range of applications in suspect tracking, unmanned parking lot management, smart logistics,and unmanned driving.It can be regarded as an image retrieval problem, aiming to retrieve images of the query vehicles from a large gallery, where the images are taken by different cameras from multi-views.Especially if the license plate is blocked, removed, destroyed, or in other special scenarios, the technology will become the only option [1–3].

Many previous studies directly adopted convolutional neural networks (CNNs) to learn the robust representation of vehicle images [4–6].However, CNNs tend to focus on global observation while ignoring other discriminative vehicle parts.Specially, under different viewing angles, the appearances of vehicles change significantly, which leads to the instability of global features [7].Local location features contain more stable and distinguishable information, such as windows,headlights, license plates, etc [8,9].These features are critical for judging similar vehicles and will not change greatly with environmental changes, so the features are more robust.Hence,many studies introduce local location features to cope with illumination, posture, perspective,and occlusion [10–12].However, these methods require additional annotations, such as direction key point annotations, local position annotations, etc.To avoid additional annotations, some studies focus on applying attention mechanisms for vehicle re-identification to increase focus on local regions [13,14], but complex attention mechanism modules need to be designed additionally.The above issues motivate us to propose a new vehicle re-identification framework that can simultaneously extract global features and strengthen local feature learning, without pre-labeling and complex attention mechanisms.

In this paper, we propose a multi-feature learning model with enhanced local attention for vehicle re-identification(MFELA).This model is composed of two branches, including global and local feature enhancement branches.The middle layer of Layer3 and last layer of Layer4 in ResNet50 [15] are used to obtain middle and high-level semantic features simultaneously, and multi-scale pooling operations of GAP and GMP are integrated to enhance the global representation of the vehicle.However, this module does not consider the subtle differences between similar cars, especially cars of the same brand, model, and color.Therefore, a method called Region Batch Dropblock (RBD) is proposed to form the local feature enhancement branch.RBD is an improvement of existing person re-recognition work BDB [16].RBD first divides the global features into several regions which enforces the network to extract discriminative details in each region.Then RBD randomly drops the corresponding areas of multiple regions in batches, namely the same semantic car parts, to enhances the attentive feature learning of each region of the remaining area .While BDB drops the global features directly, which leads to insufficient attention to local details.

In summary, the main contributions of the study are as follows:

(1) The proposed MFELA model can obtain more representative global features.Both middle and high-level semantic features of ResNet50 are simultaneously used to extract more comprehensive global features.In addition, multi-scale pooling operations bring multi-scale information.By observing the overall appearances of the vehicles, the module attempts to maximize the separation of the identity in the feature space.

(2) RBD is proposed to learn the subtle differences between similar vehicles.RBD first adopts the idea of regional division to encourage the deep model to learn distinguishing features.Then, RBD drops the corresponding areas in a batch during training, which further reinforces the attentive feature learning of local regions.The method is efficient and straightforward, does not require additional labels and complex attention mechanisms.

(3) The global semantic and local subtle discriminative cues are jointly learned in the final embedding of the vehicle.Ablation studies and experiments on two mainstream vehicle ReID datasets demonstrate the effectiveness of MFELA.It significantly improves ReID accuracy over the baseline and outperforms the most existing vehicle Re-ID methods.

The rest of the paper is organized as follows:Section 2 reviews the relevant works and Section 3 introduces the proposed model of vehicle Re-ID.Extensive experimental results are presented and analyzed in Section 4 and finally, the conclusions are summarized in Section 5.

2 Related Work

With the rapid development of deep learning, vehicle ReID has gradually become a hot topic.Existing vehicle ReID methods based on deep learning can be roughly divided into the following four categories.

(1) Vehicle ReID based on global features.Global features only focus on the vehicle’s overall appearance, such as color, model type, etc.Among the earliest attempts for vehicle ReID that involve deep learning, Liu et al.[4] proposed to extract global features by using a convolutional neural network combined with traditional methods.Wang et al.[17] extracted richer features through adequate labeling information using additional attribute information, such as specific brands and models.Zhang et al.[5] reduced the redundancy of global features by applying the SE block to automatically obtain the importance of each channel feature to improve Densenet121.However, due to the influence of low resolution,illumination variation, and cross-camera perspective, the appearance of the same vehicle is visually prone to change, which makes it difficult to obtain complete and stable global features.

(2) Vehicle ReID based on local location features.Recent related researches use the local location features to improve representation capabilities.The RAM model proposed by Liu et al.[10]also adopted the idea of regional division to extract local features.However, the RAM model directly learns from the local region without in-depth mining of local information.Wang et al.[11] defined 20 key points of the car body and extracted local vehicle features based on the pre-defined key points.Besides, He et al.[12] predefined local region locations and used the target detection algorithm to extract features of local regions.Although these methods can extract stable local features, they all need to annotate key points and regional regions in advance.As the datasets under traffic scenarios become larger, annotations will consume a lot of time and effort, and the accuracy cannot be guaranteed.

(3) Vehicle ReID based on local attention mechanism.Some researchers have begun to pay attention to the application of attention mechanisms in vehicle ReID.Khorramshahi et al.[13]proposed a dual-path attention network, which can adaptively select and focus on key feature points and azimuth information.Zhang et al.[14] adopted the attention mechanism to assign higher weights to salient local areas to achieve higher attention.Whereas most attention mechanisms are complex and difficult to come up with.Meanwhile, it increases parameters of models which further leads to the difficulty of training.

(4) Vehicle ReID based on GAN.Another effective strategy is to introduce generative adversarial networks (GAN) [18] into vehicle ReID.Zhou et al.[19] employed GAN to realize effective multi-view feature reasoning and generate multi-view features from single-view.Lou et al.[20] utilized two GANs to generate multi-perspective and hard samples, respectively.It must be mentioned that the overall model structure and training process are also more complicated when using GAN for image generation.

3 Methodology

The entire pipeline of the proposed MFELA consists of two main parts:global feature extraction and local feature enhancement.The network structure is illustrated in Fig.1

Figure 1:The entire pipeline of the proposed MFELA (yellow dotted box is the global feature extraction module to extract macroscopic appearance features, another branch enhances local features to learn the subtle differences between similar cars, and the RBD method to extract local features)

In MFELA, we use ResNet-50 as the backbone network.The purpose of global feature extraction module is to extract macroscopic appearance features.Specifically, the features obtained by RESNET-50 Layer3 and Layer4 is defined asX3andX4, respectively.We respectively do global average pooling (GAP) and global max pooling (GMP) operation onX3to obtainfxgapandfx3gap.These two features are fused to get the first global featurefx3g.Similarly, the sa3me procedure is implemented onX4to get the second global featurefx4g.Another branch enhances local features to learn the subtle differences between similar vehicles, and the RBD method is proposed to extract local features.Concretely, divide the feature map intokregions from top to bottom:p1,...,pk, and zero out the units in the dropping area of each region to get local featuresfp1,...,fpk, respectively.The Global branches and local branches are combined for multi-feature learning.The model can simultaneously learn the overall structure and fine-grained distinction information of vehicles, and enhance the discriminative ability of the model.In the process of feature extraction, triples and softmax loss are used for training.In the following sections, we discuss the details of the network structure and model training.

3.1 Global Feature Extraction

GAP preserves overall data features, while GMP acquires texture features.In addition, the fusion feature of GAP and GMP is used to obtain multi-scale information to enhance vehicle representation.

3.2 Local Feature Enhancement

If different vehicles have a similar global appearance, the differences mainly exist in local areas, such as the number and location of annual inspection signs on the windshield and window decorations.As shown in Fig.2.Therefore, we design a local branch to enhance local features.

Figure 2:Examples of different vehicles with similar global appearance (each column shows two different vehicles, and the differences in local areas are highlighted with red circles)

Firstly, the feature maps extracted from Layer4 are divided intoknon-overlapping regions vertically, and thenp1,...,pkparts are obtained.Through experiments, ifkis 6, the network achieves the best performance.In this way, the whole feature map is divided into regions, and each region is studied separately, which improves the learning strength of local regions.Subsequently, the RBD method randomly drops the corresponding areas of these regions in a batch.In particular, for each non-overlapping regions, there is a mask of the same size multiplied by it then the units in the dropping area of the mask are zeroed out.GMP is done on each region to get local featuresfp1,...,fpk, and the BN layer and the FC layer follow closely.Finally, the prediction vectorsFp1,...,Fpkare obtained.When each region randomly drops the same area, the model pays more attention to the remaining parts to strengthen the local attention learning.A simple schematic of RBD is shown in Fig.3.

Figure 3:A simple schematic of Region Batch DropBlock (the two feature maps of the same batch are evenly divided into three parts, the same area in each region is dropped, and the crossed part in the figure indicates that it is dropped)

3.3 Training

The overall objective loss function is composed of global and local losses.Both branches are trained by the combined loss of hard triple loss [22] and cross-entropy loss.

The hard triplet loss is described as follows:

a,pandnare anchor, positive, and negative samples.?is the distance margin parameter.Da,pandDa,nare the Euclidean distances calculated from the features ofaandp/nin feature space.

Besides,lcross(y,)is cross-entropy loss:

yis the ground-truth vector,is the predicted probability vector,nrepresents the type of sample,that is, the ID class of the vehicle.

Hard triple loss of global feature extraction branch is calculated byfx3gap,fx3gmp,fx4gapandfx4gmp, and cross entropy loss is calculated by the prediction vectorFx3andFx4.The global loss is described as follows:

Simultaneously, Hard triple loss of local feature enhancement branch is calculated usingfp1,...,fpk, and cross entropy loss is calculated using the prediction vectorFp1,...,Fpk.The part loss is described as

The final loss of the two-branch network is defined as

In the test stage, global featuresfx3gandfx4gare combined with local featuresfp1,...,fpkto obtain the overall features of the vehicle.Finally, cosine distance is used to compare the features of each pair of query and test images to determine their similarity.

4 Experiments

4.1 Datasets and Evaluation Metrics

We evaluate the proposed model on two mainstream datasets, namely, VehicleID and Veri776.

VehicleID[23] is a large-scale vehicle ReID dataset.It contains 221, 763 pictures of approximately 26,267 vehicles.Images of the vehicle are taken from both front and rear perspectives.The test phase is divided into three scale datasets, namely large, medium, and small datasets.Cumulative Match Curve (CMC) for top 1 (CMC@1) and top 5 (CMC@5) matches are adopted as evaluation metrics for this dataset.

Veri776[4] is also widely used in the vehicle Re-ID.It provides images captured from 20 cameras at different perspectives, consisting of 499,357 images of 776 different vehicles.Therefore,this dataset can reflect the actual situation of real-world traffic scenarios.Evaluation metrics employed for Veri776 are mean Average Precision (mAP) and CMC@1.

4.2 Implementation Details

Our network uses RTX2070 GPU for training, and the batch size is 32.Each batch has eight identities, so each identity contains four instance images in a batch.We use ResNet50 as the backbone CNN for training, which has been pre-trained on ImageNet [24] and the last stride is set to 1.All image sizes are cropped to 256×256 and normalized.Random flip and random erasing [25] are used for data enhancement.The model is trained 70 epochs in total.We used a warm-up strategy [26], the initial learning rate is 3.5×10-6and changed to 3.5×10-4in the 10th, and then drops to 3.5×10-5and 3.5×10-6in the 30th, 55th.Adam optimization model is adopted in the training process.

4.3 Comparison with State-of-art Methods

We compare the MFELA model against the recent state of the art methods, which includes:LOMO [27], FACT [4], AGNet [17], ODJL [5], RAM [10], OIFE [11], PRN [12], AAVER [13],PGAN [14], VAMI [19], and EALN [20].The results of the comparison are presented in Tabs.1 and 2.

Table 1:The mAP and CMC@1 on VeRi776

Table 2:The mAP and CMC@1 on VehicleID

We observe that, compare with the RAM which also adopts region division, the mAP of VeRi-776 has increased by 20.4%, the CMC@1 has increased by 7.5%, and the VehicleID has also been greatly improved.The reason is that our local branch not only adopts region division but also employs RBD to increases the attention to each region.Compared to other methods,PGAN and our MFELA have achieved a significant performance improvement.It can be seen that the two methods are on the same level.Because both approaches focus on the saliency local area.Noteworthy, PGAN assigns higher weights to salient local areas to gain higher attention, which involves complex attention model design.However, the proposed MFELA achieves excellent results without additional tags and complex attention mechanisms, which is efficient and straightforward.

4.4 Ablation Study

4.4.1 Selection of Parameter k

In this paper, the Region Batch DropBlock method is proposed to extract local features.Specifically, the feature maps with the size of 16×16 need to be divided intokregions from top to bottom.To make sure that the region is a block rather than a strip,kranges from 1 to 8.We do value experiment on VERI-776.As shown in the Tabs.3, ifk=6, the model shows the best performance.For the VehicleID dataset, we also directly usek=6, and the experimental results are also satisfactory, reflecting the high robustness of the model.

Table 3:The selection of parameter k on VeRi776

4.4.2 Effectiveness of Each Branch

We conduct ablation experiments on two datasets to prove the effectiveness of each branch.The experimental results are shown in Tabs.4 and 5.The global branch is more effective than the baseline.Because our global branch incorporates middle and high level information and leverages different multi-scale pooling operations to enhance vehicle representation.In the absence of a global branch, the local branch still performs better than the baseline.It is proved that RBD has a strong ability to enhance the attention learning of local regions.Adding the global branch could further improve the performance.This suggests that the two branches are mutually reinforcing, and both are important to the final performance.The motivation behind the twobranch structure in the MFELA Network is that it obtains more representative global and local features simultaneously.

To better understand the influence of RBD in MFELA, we visualize heatmaps of vehicle images.We use class activation mapping (CAM) [28] to visualize global features.Through CAM,we can observe where the network focuses on.In Fig.4, it is observed that the global branch is better than the baseline.Because the global branch utilizes both Layer3 and Layer4 features and integrates GAP and GMP multi-scale pooling features to enhance the global representation of vehicles.The MELAE model is constructed based on the global branch.The Fig.4 shows the local area of concern is covered more widely and more accurately.Thus, RBD reflects a strong focus on local regions.

Table 4:Ablation study about each branch of MFELA for VeRi776

Table 5:Ablation study about each branch of MFELA for VehicleID

Figure 4:Class Activation Maps of different vehicles (the highlighted pixels indicate that they play a more important role in determining the similarity between vehicles)

To further illustrate the effectiveness and practicability of the proposed framework, we visualize some examples of query results on VeRi-776, as shown in Fig.5.True matches are green,and false matches are red.The left is the query results of the baseline, and the right is the query results of the proposed method MELAE.As we can see, the number of correct images returned by the query of the MELAE model far exceeds that of the baseline, which shows that our model queries are more accurate.

Figure 5:Visualization of the ranking list on VeRi-776 (the left-hand side is retrieval results obtained from baseline, while the right-hand side is obtained by the proposed method)

5 Conclusion

In this paper, we propose a multi-feature learning model with enhanced local attention for vehicle re-identification.This model could obtain more representative global and local features at the same time.Global features incorporate both middle and last layers to extract more comprehensive global features.Besides, the global branch integrates GAP and GMP multi-scale pooling features to obtain multi-scale information.In addition, the RBD adopts regional division and regional batch dropping to strengthen the attention learning of local regions, which is efficient and straightforward.Extensive experiments are conducted to show the effectiveness of our model.

However, most vehicle images in the VeRi-776 and VehicleID datasets are complete, with fewer occluded.In the future, we will attempt to build a vehicle dataset that includes occlusion conditions and further improve the proposed model to adapt to the occlusion environment and make the model more robust.At the same time, we will further lightweight the model to realize the re-ID in real-time traffic scenes.

Funding Statement:This work was supported, in part, by the National Nature Science Foundation of China under Grant Numbers 61502240, 61502096, 61304205, 61773219; in part, by the Natural Science Foundation of Jiangsu Province under grant numbers BK20201136, BK20191401; in part,by the Postgraduate Research & Practice Innovation Program of Jiangsu Province under Grant Numbers SJCX21_0363; in part, by the Priority Academic Program Development of Jiangsu Higher Education Institutions (PAPD) fund.

Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

主站蜘蛛池模板: 91久草视频| 91偷拍一区| 成人综合网址| 国产精品网拍在线| 男人的天堂久久精品激情| 国产综合在线观看视频| 青青草国产精品久久久久| 日韩福利视频导航| 91人人妻人人做人人爽男同| 就去吻亚洲精品国产欧美| 国产日韩精品欧美一区喷| 久久香蕉国产线看精品| 久久网综合| 国产中文一区a级毛片视频 | 欧美国产另类| 福利视频久久| 理论片一区| 尤物精品视频一区二区三区| 精品色综合| 91网红精品在线观看| 亚洲永久精品ww47国产| 亚洲欧美在线精品一区二区| 国产成人精品在线1区| 国产激情在线视频| 国产资源免费观看| 青青国产在线| 色香蕉网站| 欧美区一区二区三| 国产无套粉嫩白浆| 91成人精品视频| 成年人免费国产视频| 国产精品高清国产三级囯产AV| 黄色成年视频| 欧美综合区自拍亚洲综合绿色| 国产欧美亚洲精品第3页在线| 欧美日韩导航| 亚洲日本一本dvd高清| 国产真实乱人视频| 国产成人综合久久精品下载| 亚洲精品无码高潮喷水A| 国产麻豆精品久久一二三| 国产成人综合亚洲欧洲色就色| 欧美精品成人一区二区视频一| 少妇高潮惨叫久久久久久| 欧美国产成人在线| 高清乱码精品福利在线视频| 国产真实自在自线免费精品| 亚洲女同欧美在线| 日本欧美一二三区色视频| 91av国产在线| 青草视频久久| 欧美午夜视频| 国产欧美视频综合二区| 又爽又大又黄a级毛片在线视频| 亚洲成人免费在线| 日韩美一区二区| 国产精品成人啪精品视频| 精品国产一二三区| 久久国产精品娇妻素人| 国模沟沟一区二区三区| 日韩欧美中文| 欧美日韩午夜视频在线观看| 第一页亚洲| 天天干天天色综合网| 日韩高清一区 | 亚洲第一天堂无码专区| 婷婷久久综合九色综合88| 国产成人高清精品免费软件| 国产成人调教在线视频| 国产高清毛片| 日本精品αv中文字幕| 亚洲永久视频| 91九色视频网| 国产高清精品在线91| 一本综合久久| 毛片在线播放a| 精品国产中文一级毛片在线看 | 三级欧美在线| 精品久久久久久中文字幕女| 日韩最新中文字幕| 手机永久AV在线播放| 一本大道无码日韩精品影视 |