999精品在线视频,手机成人午夜在线视频,久久不卡国产精品无码,中日无码在线观看,成人av手机在线观看,日韩精品亚洲一区中文字幕,亚洲av无码人妻,四虎国产在线观看 ?

Ghost Module Based Residual Mixture of Self-Attention and Convolution for Online Signature Verification

2024-05-25 14:40:26FangjunLuanXuewenMuandShuaiYuan
Computers Materials&Continua 2024年4期

Fangjun Luan ,Xuewen Mu and Shuai Yuan,?

1Department of Computer Technology,School of Computer Science and Engineering,Shenyang Jianzhu University,Shenyang,110168,China

2Liaoning Province Big Data Management and Analysis Laboratory of Urban Construction,Shenyang Jianzhu University,Shenyang,110168,China

3Shenyang Branch of National Special Computer Engineering Technology Research Center,Shenyang Jianzhu University,Shenyang,110168,China

ABSTRACT Online Signature Verification(OSV),as a personal identification technology,is widely used in various industries.However,it faces challenges,such as incomplete feature extraction,low accuracy,and computational heaviness.To address these issues,we propose a novel approach for online signature verification,using a one-dimensional Ghost-ACmix Residual Network(1D-ACGRNet),which is a Ghost-ACmix Residual Network that combines convolution with a self-attention mechanism and performs improvement by using Ghost method.The Ghost-ACmix Residual structure is introduced to leverage both self-attention and convolution mechanisms for capturing global feature information and extracting local information,effectively complementing whole and local signature features and mitigating the problem of insufficient feature extraction.Then,the Ghost-based Convolution and Self-Attention(ACG)block is proposed to simplify the common parts between convolution and self-attention using the Ghost module and employ feature transformation to obtain intermediate features,thus reducing computational costs.Additionally,feature selection is performed using the random forest method,and the data is dimensionally reduced using Principal Component Analysis (PCA).Finally,tests are implemented on the MCYT-100 datasets and the SVC-2004 Task2 datasets,and the equal error rates (EERs) for small-sample training using five genuine and forged signatures are 3.07%and 4.17%,respectively.The EERs for training with ten genuine and forged signatures are 0.91% and 2.12% on the respective datasets.The experimental results illustrate that the proposed approach effectively enhances the accuracy of online signature verification.

KEYWORDS Online signature verification;feature selection;ACG block;ghost-ACmix residual structure

1 Introduction

Online Signature Verification(OSV)technology is a biometric authentication technique aimed at establishing a high-accuracy and robust system to determine whether a signature sample is genuine or forged.Unlike offline signatures,online signatures are real-time signals that vary with time and primarily capture dynamic information,such as signature trajectory,pressure,vertical and horizontal deviations,and velocity,which are collected using signature devices like electronic tablets[1–3].Online signatures are widely accepted in society due to their ease of collection,resistance to imitation,and high security [4,5].However,the instability of online signatures makes them susceptible to changes of the signer’s physiological state,psychological fluctuations,and surrounding environment.Additionally,the limited number of online signature samples results in significant challenges for achieving high verification accuracy [6–8].Therefore,achieving effective personal identification through OSV technology is crucial for determining the legal validity of handwritten signatures.In-depth research on online signature verification can address challenges related to stability and accuracy,thereby fostering its application and development in areas such as e-commerce and e-government.

Convolution and self-attention mechanisms,as potent techniques in deep learning,are frequently employed in current methods for handwritten signature verification.Consequently,methods for handwritten signature verification can be categorized into convolution-based and self-attention-based approaches.Each of these approaches has its unique advantages and limitations:Convolution demonstrates excellent performance in data processing,capturing local features and spatial information.However,it is constrained by the design of local receptive fields and parameter sharing,which may impede its ability to handle long-range dependencies and global information.On the other hand,the self-attention mechanism offers advantages in capturing global dependencies and dynamically computing attention weights,making it suitable for processing sequential data and global information.Nevertheless,when dealing with large-scale data,it may encounter significant computational overhead and might excessively focus on global information,potentially lacking the ability to handle local features.

To enhance the performance and accuracy of handwritten signature verification,while minimizing computational overhead and effectively handling local and global features,we propose a novel onedimensional Ghost-ACmix Residual Network (1D-ACGRNet) that integrates convolution and selfattention mechanisms.The main contributions are as follows:

First of all,we introduce a novel Ghost-ACmix residual structure that combines the strengths of convolution and self-attention mechanisms to extract feature information.This approach allows for a synergistic utilization of global and local signature features,effectively addressing the issue of insufficient feature representation in signatures.

Then,recognizing that the fusion of convolution and self-attention increases model complexity and computational overhead,we propose a Ghost-based Convolution and Self-Attention (ACG)module.In this module,during the underlying convolution operations of both convolution and selfattention,a straightforward feature transformation is employed to replace the conventional convolution process,acquiring more comprehensive feature information.The advantage of this approach lies in reducing feature redundancy and lowering computational costs,thereby,enhancing computational efficiency while maintaining model performance.

Finally,we employ the random forest algorithm for feature selection on global features,filtering out redundant information based on feature importance scores.We also adapt the Principal Component Analysis (PCA) method for feature dimensionality reduction,thereby,further reducing the computational load of the model.

2 Related Work

2.1 Previous Work

After decades of exploration and research,numerous online handwritten signature verification methods have been proposed.Generally,these methods can be categorized into two types according to the approach of feature extraction: Parameter-based methods and function-based methods.The parameter-based signature verification methods primarily extracted a fixed-length feature vector from the time-series data of the entire or partial sampling points of signatures.Then,they transformed the comparison between two signatures into a comparison between their respective feature vectors.Jain et al.[9] considered the stroke count as a global feature.Lee et al.[10] used average speed,average pressure,and the number of pen lifts during the signature as global features.Ibrahim et al.[11]extracted 35 global features,including maximum signature velocity,pen-down count,signature aspect ratio,etc.,for signature verification.This method showed acceptable overall verification performance,but some signers had relatively low verification accuracy.Fierrez-Aguilar et al.[12] introduced 100 global features and ranked them based on their discriminatory power in differentiating genuine and forged signatures.Vorugundi et al.[13] combined a one-dimensional convolutional neural network(1D-CNN)with a Siamese neural network(SNN)and used the extracted global feature set to represent online handwritten signatures.Additionally,some transform methods,such as wavelet transform[14],Fourier transform[15],i-vector[16],have also been applied for extracting global features from online handwritten signatures.

With the continuous development of deep learning and signature devices,many researchers have been devoted to adopting deep learning methods for online signature authentication.Recurrent Neural Networks(RNN)and Long Short Term Memory(LSTM)networks were widely used in this context.Tolosana et al.[17] investigated different types of RNN networks for signature authentication and proposed a multi-RNN Siamese network as well as a Siamese network combining LSTM and Gate Recurrent Unit (GRU).Greff et al.[18] conducted an analysis of eight different LSTM variants across three distinct tasks: Speech recognition,handwriting recognition,and polyphonic music modeling.Furthermore,Li et al.[19]proposed a Stroke-Based RNN model that combined signature stroke information with RNN networks in 2019.Lai et al.[20] proposed a method for synthesizing dynamic signature sequences to address the problem of lacking forged signatures during training.Vorugunti et al.[21]merged manually extracted features with features extracted by Autoencoders(AE)and applied the Depthwise Separable Convolutional Neural Network(DWSCNN)for the first time to online signature verification tasks.Additionally,a language-independent shallow model Shallow Convolutional Neural Network (sCNN) [22] based on a convolutional neural network (CNN) has been introduced.This model had a simple structure,using a custom shallow CNN to automatically learn signature features from the training data for signature authentication.

The function-based online signature verification methods treated the time-series data of all or partial sampling points as feature information and performed signature verification by calculating the distance between the template signature and the test signature.This approach retained information from all sampling points,providing more data and higher security.Kholmatov et al.[23]improved the Dynamic Time Warping (DTW) method by constructing a three-dimensional feature vector for each signature using the distance between the current sample and all reference samples.Song et al.[24] selected five stable local temporal features: x-axis velocity,y-axis velocity,pressure,y-axis acceleration,and centripetal acceleration,and used the DTW-SCC algorithm for signature verification.Lai et al.[25] proposed an online handwritten signature verification method based on RNN and signature Length Normalized Path Signature(LNPS)descriptor to address the problem of inconsistent signature lengths,but this method involved heavy model computation.In addition,the Information Divergence-Based Matching Strategy [26],Gaussian Mixture Model (GMM) [27],and other methods have also been applied to function-based signature verification.

2.2 Existing Challenges

From the above-mentioned research work,it can be concluded that researchers have made significant progress in the field of online signature authentication over the years.However,there are still some challenges that need to be addressed.Firstly,improving the accuracy of online signature verification is still a primary focus and a difficult task in current research.Previous studies mainly used traditional methods or convolutional methods for feature extraction from online signatures.However,online signatures are complex and contain multiple types of information,and susceptible to various external factors.In the process of feature extraction,the global or local information of the features is often overlooked,leading to limited effectiveness in feature extraction.Secondly,reducing the computational cost of signature verification,eliminating redundant information,and lightening the model structure pose significant challenges in online signature authentication methods.Moreover,signatures that have been deliberately forged over time are difficult to distinguish from genuine signatures.Therefore,distinguishing between proficient forgeries and genuine penmanship is a challenging aspect in online signature authentication.

2.3 Residual Network Architecture

ResNet,short for Residual Network,is a widely recognized and significant neural network architecture in deep learning,proposed by He et al.in 2016 [28].ResNet was designed to address the issues of vanishing and exploding gradients that occur in deep neural networks.In deep neural networks,as the number of layers increases,the information propagation may suffer from the problem of vanishing gradients.This leads to difficulties in network convergence of training very deep neural networks.ResNet addresses this issue by introducing a residual learning approach.It utilizes“residual blocks,”which allow the network to directly learn the residual part of the target function.By using residual blocks,even for a very deep network,the information propagation path becomes more direct with reducing gradient decay and making it easier to train and optimize deeper networks.The core idea of ResNet is to introduce skip connections or “shortcut connections”in the network,enabling the direct transfer of information between different layers.By incorporating skip connections,ResNet enables the flow of information across layers,facilitates the training of very deep neural networks and addresses the vanishing gradient problem,and ultimately leads to improved performance and convergence.

2.4 Integration of Self-Attention and Convolution

Convolution and self-attention are two powerful representation learning techniques that are considered as independent methods.However,they have strong inherent connections because much of their computations actually involve similar operations.Specifically,traditional convolution can be decomposed into smaller convolutions,which are then shifted and summed.Similarly,in the selfattention module,the projections of queries,keys,and values can be seen as multiple convolutions,followed by computation of attention weights and aggregation of values.As a result,the first stage of both modules involves similar operations.Moreover,the first stage is computationally more intensive compared with the second stage.Their operations can be used to perform an elegant fusion of both seemingly distinct methods,known as the hybrid model(ACmix)[29],which simultaneously leverages the advantages of both self-attention and convolution while having the minimum computational overhead compared with pure convolution or self-attention counterparts.

2.5 GhostNet

The Ghost block[30]is a lightweight convolutional block that achieves comparable performance to standard convolution with fewer parameters and lower training costs.Traditional convolutions require significant computations and parameters to extract feature maps,often resulting in feature maps containing rich or redundant information,with many feature maps being similar.The Ghost block obtains some “ghost” feature maps through linear transformations,which can be derived from other feature maps.The Ghost block divides the convolution process into two parts:Standard convolution and transformations(identity mapping and depth-wise convolution).During this process,the Ghost block introduces the concept of “Ghost”dividing the input feature maps into main and ghost parts.The main part consists of a small portion of feature maps generated by convolutions,preserving the primary feature information,while the ghost part consists of feature maps generated by identity mapping,assisting computations for providing additional information.This design allows the Ghost block to maintain low computational and memorial overhead while offering high model expressiveness and accuracy.

Another advantage of the Ghost block is its adaptability and pluggability.It can be integrated into various convolutional neural network structures and model architectures.By applying the Ghost block to different parts of a network model,lightweight and efficient goals can be achieved without the need to redesign the entire model.

3 Method

3.1 Whole Framework of the Proposed Method

The overall framework of our proposed online signature verification method based on the 1DACGRNet model is illustrated in Fig.1.It primarily consists of two main components: Feature extraction and selection,as well as the Ghost-ACmix residual module for signature verification.In the first part of feature extraction and selection,the original dataset undergoes feature extraction,resulting in a 54-dimensional global feature vector.Subsequently,a random forest algorithm [31] is employed to determine the importance of each feature dimension,leading to the removal of redundant features with importance scores below 0.Additionally,PCA[31]is used to reduce the dimensionality of 54 global features.The resulting four global features are combined with the features extracted to form the feature information.The second part is the 1D-ACGRNet model signature authentication.The stable global features processed in the first stage are input into a convolutional layer,followed by LeakyReLU activation and Average Pooling.The output is then fed into four Ghost-ACmix residual modules for feature modeling,and then output to the classifier composed of the full connection layer and SoftMax function for classification,and finally obtains the signature authenticity.Table 1 provides the specific parameters of the model.

Table 1: Detailed parameters of 1D-ACGRNet

Figure 1: Whole framework of the proposed method

3.2 Ghost-ACmix Residual Structure

In traditional deep neural networks,with the network depth increases,the gradient signal becomes very small,making training difficult or even impossible.This paper employs the Ghost-ACmix residual structure as shown in Fig.2,which introduces skip connections to allow information to bypass certain layers,effectively overcoming the problem of vanishing gradients.

Figure 2: Ghost-ACmix residual structure

Considering that conventional online signature verification methods often overlook either global or local information,leading to incomplete feature extraction and insufficient accuracy,given that online signature authentication techniques are often deployed on small-capacity embedded devices,requiring a reduction in overall model parameters and computational complexity,we adopt the ACG block combining convolution and self-attention with the Ghost method to replace the standard convolutional part in the residual structure.This approach aims to improve model performance while remaining lightweight.Assuming that the input to the residual structure is denoted as x,the output y(x)can be represented by the following equation:

3.3 ACG Block

The convolutional and self-attention modules typically follow different design paradigms.Traditional convolutional operations utilize aggregation functions over local receptive fields,applying weighted summation to the data within each window based on the convolution filter weights.In contrast,self-attention modules apply weighted averaging operations based on the context of input features,where attention weights are dynamically computed as similarity functions between relevant pairs of pixels.This flexibility allows the attention module to adaptively focus on different regions and capture more diverse information features.The specific operation is shown in Fig.3.

For the convolution operation,given a convolution kernel K of size 1×k,it can be decomposed into k sets of 1×k convolution kernels,followed by subsequent shifting and summation operations.The specific decomposition process is as follows.

Figure 3: ACG block

The convolutional kernel K of size 1 × k is decomposed intokindividual 1 × 1 convolutional kernels,denoted asKp,wherep∈{0,1,...,k-1}.The input feature map isF∈RCin×W,and the output feature map isG∈RCout×W.The tensorsf i∈RCinandgi∈RCoutrepresent the feature tensors of pixelsiinFandGrespectively.Utilizing these 1×1 convolutional kernels,convolution operations are performed with the corresponding positions offi,resulting in k feature maps denoted asgip.Due to the high computational complexity of tensor displacement,a fixed-kernel depth convolution is used to perform the tensor displacement operation.Shift and sum the k feature maps to obtain the pixel valuegiin the final output featureG.

By such decomposition,the standard convolution can be achieved through a series of smaller 1×1 convolutional kernels and a two-step process of shifting and summation.

For self-attention operations,the attention module can also be decomposed into two stages.In the first stage,the input features are projected into query q,key k,and value v using a 1×1 convolution.Similarly,let the input features be represented asF∈RCin×W,and the output features asG∈RCout×W.This facilitates the transformation off i∈RCinandgi∈RCoutrepresentingFandGcorresponding pixelifeature tensor.qi,ki,virespectively denote tensors of query,key,and value for pixeli.Wq,Wk,Wvcorrespondingly refer to the weights associated with query q,key k,and value v.

The second stage involves computing attention weights and performing a weighted sum.The attention weights A are computed by taking the dot product of the query q and key k,followed by scaling the result using the scaling factor 1/√dto prevent gradient vanishing or exploding issues.drepresents the feature dimension ofqi.

Therefore,by weighting and summing the value vectors v using attention weights,the pixelgiin the output featureGcan be computed.

Decomposing the Self-Attention into two stages makes the computation process more efficient,as a significant portion of computations is concentrated in the first stage,similar to the computation pattern in traditional convolutions.Since both Convolution and Self-Attention apply convolutional operations in the first stage,it is possible to perform the first-stage computation only once and then separately pass the intermediate features to the convolution and self-attention paradigms for computation,finally combining them through weighted summation.This integration fully leverages the advantages of both,while reducing computational costs.

Although this processing approach reduces the parameter count to some extent,there still exists a significant amount of redundancy in the intermediate features.To address this issue,Fig.4 represents the idea of the Ghost block introduced in this paper.In the first stage of obtaining intermediate features,the first set of intermediate features is generated through a 1×1 convolution.Subsequently,the Ghost block is employed to generate the remaining sets of intermediate features.By using the Ghost block,the overall parameter count of the model is reduced,and the problem of feature redundancy is effectively avoided.

Figure 4: Ghost block

3.4 Classifier

As depicted in Fig.1,the classifier of the 1D-ACGRNet model is composed of several layers including Flatten layer,Dropout layer,Relu activation function layer,and three fully connected layers[128,64,2].In the Flatten layer,the features produced by four Ghost-ACmix residual structures are flattened into a one-dimensional array and fed into the fully connected layers.Considering that the front-end network has already captured significant features relevant to signature authentication through max-pooling layers,we introduce a single Dropout layer (0.5) in the first fully connected layer to eliminate redundant feature information.Ultimately,the SoftMax activation function is employed to predict the authenticity of each signature sample.The design of the entire classifier aims to fully exploit effective features and prevent overfitting,thus,enhancing the performance of signature verification.

4 Experimental Preparation

4.1 Datasets and Preprocessing

The experimental datasets in this study include the MCYT-100 datasets and the SVC-2004 Task2 datasets,which are described as follows:

The MCYT-100 datasets are a publicly available datasets commonly used for research in online signature verification.The datasets are released by the BiDA Laboratory at the Autonomous University of Madrid.The MCYT-100 Spanish datasets are consist of 100 users,and each user has 25 genuine and forged signature samples.The signature feature information includes five time-series features:Abscissa,ordinate,pen pressure,pen horizontal angle,and pen vertical angle.

The SVC-2004 Task2 Chinese-English datasets are subset datasets provided at the First International Signature Verification Competition held by the Hong Kong University of Science and Technology in 2004.The datasets contain a relatively small number of signatures and consist 40 users,with each user having 20 genuine and forged signature samples.The signature feature information includes seven time-series features: Abscissa,ordinate,pen pressure,time,pen horizontal angle,pen vertical angle,and pen-up/pen-down flag.For the sake of model generalization,the time and penup/pen-down flag time-series features in the dataset were not used during the experiment.

To mitigate the impact of input device variations and the influence of the number,size,and location of sampling points on each signature input,we employed smoothing and normalization on the signature feature sequence X and Y coordinates.This is done to eliminate data errors originating from data acquisition equipment and individual user factors.We utilized a five-point cubic smoothing filter to process the feature data.The time sequence of the signature is represented as T,where n denotes the number of sampling points,X and Y represent the abscissa and ordinate trajectory feature sequence,P signifies the pressure sequence feature,and AZ and AL denote the horizontal and vertical angular features,respectively.For the signature feature sequence{T1,T2,...,Tn},where in Ti ∈{X,Y},corresponds to the smoothed data points,i=(1,2,...,n),while n represents the count of signature samples.The smoothing process is as follows:

Subsequently,the selected features are normalized to scale the data of the five feature sequences within the range of[0,1].The formula for normalization is as follows:

After computation,a total of 54 global signature features are extracted,as detailed in the following Table 2.

Table 2: Signature global features

4.2 Feature Selection

In the process of using multiple global features,the issue of feature redundancy inevitably arises.These redundant features not only increase the complexity of model training but may also negatively impact the model performance in authenticating genuine and forged signatures.To mitigate this drawback,this study employs a random forest feature selection method based on ensemble learning strategies to select features from the 54 extracted global features.This method assigns importance weights to each feature,and features with an importance weight of zero are removed.Simultaneously the PCA method is adopted to reduce 54 global features to four dimensions for filtering out effective features from the numerous global features.

4.3 Training Configuration

This study primarily conducted network hyperparameter optimization,training,and validation experiments using Python 3.8 and the TensorFlow 2.3 deep learning library.The adaptive moment estimation(Adam)optimizer is employed in this research.To ensure the convergence of the network model and mitigate overfitting,the model weight with the highest validation rate is selected as the final trained model after multiple iterations of training.Table 3 shows the results of hyperparameter optimization during the model learning process.

Table 3: Hyperparameter results

4.4 Evaluation Methods

In this research,a variety of metrics are employed to assess the model.The formulas for these evaluation metrics are as follows:

Accuracy (ACC) is utilized to measure overall correctness.The False Acceptance Rate (FAR)quantifies the rate of incorrectly accepting forged signatures,while the False Rejection Rate (FRR)gauges the rate of incorrectly rejecting genuine signatures.The Equal Error Rate (EER) represents the point where FAR equals FRR,with lower values indicating superior performance.The Receiver Operating Characteristic(ROC)curve illustrates the trade-off between FAR and FRR.The Confusion Matrix(CM)presents detailed classification outcomes,including True Positives,True Negatives,False Positives,and False Negatives.

5 Results and Analysis

5.1 Analysis of Model Performance

To assess the overall performance of the model,this section presents a comprehensive analysis of signature authentication performance on the MCYT-100 and SVC-2004 Task2 datasets.This analysis includes the evaluation of authentication performance for all users.During the experimental process,the selected feature set after feature selection is employed as the input to the network.The experiment employed five genuine and forged signatures for training.

To further validate the effectiveness of the proposed method,we conducted experiments on both the 1D-ACGRNet model and the 1D-RNet model,comparing their EER and plotting ROC curves for comparison.From Fig.5,it can be observed that on the MCYT dataset,the EER for signature verification with the 1D-RNet model is 4.37%,and the one for the 1D-ACGRNet model is 3.07%,and the improved 1D-ACGRNet model shows a significant decrease in EER by 1.30%.On the SVC dataset,the 1D-RNet model EER for signature verification is 4.76%,while the 1D-ACGRNet model EER is 4.17%,which indicates a reduction of 0.59%compared with the former.

Figure 5: The ROC curves of the MCYT dataset and the SVC dataset for two model architectures

In this subsection,we conduct comprehensive performance experiments using both the SVC dataset and the MCYT dataset,and the experimental results are presented in Table 4.It can be observed that with the increase in training data,FAR,FRR,and EER all show a decreasing trend on both datasets,while ACC shows improvement.The detailed data of the confusion matrix corresponding to Table 4 is illustrated in Fig.6.The experimental results prove that on both the MCYT and SVC datasets,the 1D-ACGRNet model exhibits substantial overall performance enhancement in signature verification,confirming the effectiveness of the 1D-ACGRNet model.

Table 4: Results for the two datasets

Figure 6: Confusion matrix of SVC and MCYT datasets

5.2 Analysis of Ablation Experiments

We employ “ablation experiments” to qualitatively and quantitatively analyze the impact of various components within the research model.To ensure the robustness of the model,the experiment adopts a stratified k-fold cross-validation method,dividing the signature data of each user into k equal parts,with one part as the training data,conducting k rounds of training and validation to ensure that each data subset participates in the testing phase.In this experiment,we ensure that the training data volume for each round is 5.Table 5 provides the specific results of the experiment.In conclusion,the enhanced model employed in this study has demonstrated significant performance improvements.This model not only enhances accuracy but also reduces computational costs,rendering it more suitable for online signature verification.

Table 5: Comparison of ablation experiment results

5.3 Comparison with Similar Work

To further validate the effectiveness of the proposed method,this section compares the performance of our approach with other methods.Table 6 presents experimental results obtained by different methods proposed by predecessors on the MCYT dataset and SVC dataset.When training with 5 genuine and forged signatures,as well as with 10 genuine and forged signatures,the equal error rates(EER)are consistently lower than those of other methods.This further validates the effectiveness of the proposed online handwritten signature verification method based on the one-dimensional Ghost-ACmix Residual Neural Network(1D-ACGRNet)introduced in this study.

Table 6: Comparison of EER across different methods

6 Conclusion

This paper proposes an online signature verification method using a 1D-ACGRNet,which integrates convolution and self-attention mechanisms into a residual network.It is the first application of a residual network embedded with a mixture of convolution and self-attention mechanisms in this field.Convolution and self-attention are proficient in extracting local and global information from data,respectively.By combining them,the signature information becomes more comprehensive,thereby enhancing the overall accuracy of the model.Simultaneously applying feature selection removes redundant global features,which are combined with feature dimensionality reduction methods to obtain effective feature information.And the ghost method is further introduced to reduce the model’s computational complexity.The experimental results indicate that,when training samples are employed with 5 genuine and forgery signature data on the MCYT-100 and SVC-2004 Task2 datasets,the equal error rates of this study are 3.07%and 4.17%,respectively.When using 10 genuine and forgery signatures,the equal error rates are 0.91%and 2.12%.Compared with previous research,the proposed method in this paper has implemented some improvement in performance.

Acknowledgement:The authors would like to present appreciation to National Natural Science Foundation of China and Liaoning Provincial Science and Technology Department Foundation.

Funding Statement:This work is supported by National Natural Science Foundation of China(Grant No.62073227)and Liaoning Provincial Science and Technology Department Foundation(Grant No.2023JH2/101300212).

Author Contributions:The authors confirm contribution to the paper as follows: Study conception and design:F.Luan,X.Mu;data collection:X.Mu;analysis and interpretation of results:F.Luan,X.Mu,S.Yuan;draft manuscript preparation:X.Mu,S.Yuan.All authors reviewed the results and approved the final version of the manuscript.

Availability of Data and Materials:Due to the nature of this research,participants of this study did not agree for their codes to be shared publicly,and supporting datasets are available.The SVC-2004 Task2 database is openly at: https://cse.hkust.edu.hk/svc2004/download.html.The MCYT-100 database is openly at:http://atvs.ii.uam.es/atvs/mcyt100s.html.

Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

主站蜘蛛池模板: 成人91在线| 青青草一区| 在线免费亚洲无码视频| 一区二区三区精品视频在线观看| 国产精品.com| 91麻豆久久久| 亚洲有无码中文网| 中文字幕亚洲另类天堂| 国产成人精彩在线视频50| 国产亚洲精久久久久久无码AV| 亚洲熟妇AV日韩熟妇在线| 欧美视频免费一区二区三区| 国产精品视频公开费视频| 欧洲日本亚洲中文字幕| 亚洲天堂高清| 丁香六月激情综合| 日韩高清成人| 9999在线视频| 黄片一区二区三区| 美女啪啪无遮挡| 国产特级毛片aaaaaaa高清| 久久精品免费国产大片| 精品少妇人妻av无码久久 | 欧美一区中文字幕| 久久精品娱乐亚洲领先| 国产成人免费视频精品一区二区| 免费观看男人免费桶女人视频| 国产成人精品亚洲77美色| 国产青榴视频在线观看网站| 亚洲AⅤ无码国产精品| 农村乱人伦一区二区| 欧美激情网址| 小13箩利洗澡无码视频免费网站| 成人精品区| 美臀人妻中出中文字幕在线| 欧美一区二区丝袜高跟鞋| 91人妻日韩人妻无码专区精品| 国产区91| 福利小视频在线播放| 精品一区二区三区视频免费观看| 亚洲香蕉久久| 亚洲日韩Av中文字幕无码| 日本一区二区不卡视频| 亚洲αv毛片| 国产精品黄色片| 婷婷综合亚洲| 国产福利微拍精品一区二区| 国产成人亚洲欧美激情| 欧美国产日韩一区二区三区精品影视| 97青草最新免费精品视频| 欧美日韩va| 亚洲国产精品人久久电影| 亚洲国产成熟视频在线多多| 国产成人综合网在线观看| 亚洲精品成人福利在线电影| 国产午夜看片| 日本手机在线视频| 深夜福利视频一区二区| 成人免费一级片| 日韩欧美国产区| 久热中文字幕在线| 呦女精品网站| 欧美精品综合视频一区二区| 国内精品久久人妻无码大片高| 色哟哟国产精品| 欧美区一区| 色综合中文字幕| 国产制服丝袜91在线| 欧美亚洲综合免费精品高清在线观看 | 91丝袜在线观看| 国产成人禁片在线观看| 日本人妻丰满熟妇区| 成人午夜久久| 黑人巨大精品欧美一区二区区| 无码AV日韩一二三区| 大陆精大陆国产国语精品1024| 亚洲精品无码AⅤ片青青在线观看| 色婷婷丁香| 国产男女XX00免费观看| 伊人网址在线| 99无码中文字幕视频| 色婷婷电影网|