999精品在线视频,手机成人午夜在线视频,久久不卡国产精品无码,中日无码在线观看,成人av手机在线观看,日韩精品亚洲一区中文字幕,亚洲av无码人妻,四虎国产在线观看 ?

Compact Bat Algorithm with Deep Learning Model for Biomedical EEG EyeState Classification

2022-11-11 10:45:14SouadLarabiMarieSainteEatedalAlabdulkreemMohammadAlamgeerMohamedNourAnwerMustafaHilalMesferAlDuhayyimAbdelwahedMotwakelandIshfaqYaseen
Computers Materials&Continua 2022年9期

Souad Larabi-Marie-Sainte,Eatedal Alabdulkreem,Mohammad Alamgeer,Mohamed K Nour,Anwer Mustafa Hilal,Mesfer Al Duhayyim,Abdelwahed Motwakel and Ishfaq Yaseen

1Department of Computer Science,College of Computer and Information Sciences,Prince Sultan University,Saudi Arabia

2Department of Computer Sciences,College of Computer and Information Sciences,Princess Nourah Bint Abdulrahman University,Riyadh,11671,Saudi Arabia

3Department of Information Systems,College of Science&Art at Mahayil,King Khalid University,Saudi Arabia

4Department of Computer Sciences,College of Computing and Information System,Umm Al-Qura University,Saudi Arabia

5Department of Computer and Self Development,Preparatory Year Deanship,Prince Sattam bin Abdulaziz University,AlKharj,Saudi Arabia

6Department of Natural and Applied Sciences,College of Community-Aflaj,Prince Sattam bin Abdulaziz University,Saudi Arabia

Abstract: Electroencephalography (EEG)eye state classification becomes an essential tool to identify the cognitive state of humans.It can be used in several fields such as motor imagery recognition, drug effect detection,emotion categorization, seizure detection, etc.With the latest advances in deep learning (DL)models, it is possible to design an accurate and prompt EEG EyeState classification problem.In this view, this study presents a novel compact bat algorithm with deep learning model for biomedical EEG EyeState classification(CBADL-BEESC)model.The major intention of the CBADL-BEESC technique aims to categorize the presence of EEG EyeState.The CBADL-BEESC model performs feature extraction using the ALexNet model which helps to produce useful feature vectors.In addition, extreme learning machine autoencoder (ELM-AE)model is applied to classify the EEG signals and the parameter tuning of the ELM-AE model is performed using CBA.The experimental result analysis of the CBADL-BEESC model is carried out on benchmark results and the comparative outcome reported the supremacy of the CBADL-BEESC model over the recent methods.

Keywords: Biomedical signals; EEG; EyeState classification; deep learning;metaheuristics

1 Introduction

Brain-Computer Interface (BCI)is the growing field in human-computer interaction (HCI).It permits users to communicate with computers via brain-activity[1].Usually,these kinds of activities are measured by using an electroencephalography(EEG)signal.EEG is a non-invasive physiological method for recording brain electrical activity based on electrodes put on distinct locations on the scalp.classification of Eye condition,the task of identifying the state of eye in case it is opened or closed,is a generic time series problem to identify human cognitive condition[2,3].Classifying human cognitive state is greater for medical concern in daily lives.Significant use-cases to detect eye condition results in the recognition of eye blinking rate,according to its possibility to forecast diseases like Parkinson’s disease or subject suffered from Tourette syndrome[4].The precise recognition of the eye condition with EEG signal is the challenge however indispensable task for the healthcare sector as well as in our daily lives[5,6].Different machine learning(ML)based methods were introduced for classifying the EEG-based signal in several applications.Many of them had presented the conventional method.However,it is necessary for correct and accurate classification models that could effectively categorize the eye conditional with electroencephalogram signal[7].

Tahmassebi et al.[8]presented an explainable structure with capability of real time forecast.For demonstrating the analytical power of structure, a test case on eye state recognition utilizing EEG signal was utilized for investigating a deep neural network(DNN)technique that generates a forecast and that predictive was interpreting.The authors in [9] developed the temporal order of data from the place.It can generate several CNN network techniques and choose optimum filter and depth.In CNN feature techniques were effectual became concerned issue dependent and subject independent eye state EEG classifiers.Islam et al.[10]presented 3 frameworks of DL technique utilizing ensemble approach to eye state detection(open/close)in EEG directly.The analysis was implemented on freely accessible publicly EEG eye state data set of 14980 instances.The individual efficiency of all classifiers is detected,and also application of detection efficiency of an ensemble network is related to the present prominent manners.

The authors in[11]established a hybrid classification technique for eye state recognition utilizing EEG signals.This hybrid classifier technique was estimated with other standard ML approaches,8 classifier techniques (Pre-possessed+Hypertuned), and 6 recent approaches for assessing their suitability and exactness.This presented classifier technique introduces an ML based hybrid method to the classification of eye states utilizing EEG signals with higher accuracy.Nkengfack et al.[12]presented classification model containing approaches dependent upon Jacobi polynomial transforms(JPTs).Discrete Legendre transforms (DLT)and discrete Chebychev transform (DChT)primarily remove the beta(β)and gamma(γ)rhythm of EEG signal.Afterward,various measures of difficulty are calculated in the EEG signal and its removed rhythm and executed as input of least-square support vector machine(LS-SVM)technique with RBF kernel.

In [13], a robust and unique artificial neural network (ANN)based ensemble approach was established in that several ANN was trained separately utilizing distinct parts of trained data.The resultants of all ANNs are joined utilizing other ANN for enhancing the analytical intelligence.The resultant of this ANN has been regarded as vital forecast of user eye state.The presented ensemble technique needs lesser trained time and takes extremely accurate eye state classifier.Shooshtari et al.[14] presented 8 confidence connected property from EEG and eye data that is considered descriptive of determined confidence level from random dot motion (RDM).Since a fact, the presented EEG and eye data property were able of identifying over 9 different levels of confidence.Amongst presented features,the latency of pupil maximal diameter with stimulation performance has introduced that one of the connected to confidence levels.

This study presents a novel compact bat algorithm with deep learning model for biomedical EEG EyeState classification(CBADL-BEESC)model.The major intention of the CBADL-BEESC technique aims to categorize the presence of EEG EyeState.The CBADL-BEESC model performs feature extraction using the ALexNet model which helps to produce useful feature vectors.In addition,extreme learning machine autoencoder(ELM-AE)model is applied to classify the EEG signals and the parameter tuning of the ELM-AE model is performed using CBA.The experimental result analysis of the CBADL-BEESC model is carried out on benchmark results

The rest of the paper is organized as follows.Section 2 offers the proposed model, Section 3 validates the results,and Section 4 draws the conclusion.

2 The Proposed CBADL-BEESC Model

In this study, a novel CBADL-BEESC technique has been developed for biomedical EEG EyeState classification.The main goal of the CBADL-BEESC technique is to recognize the presence of EEG EyeState.The CBADL-BEESC model involves AlexNet feature extractor,ELM-AE classifier,and CBA parameter optimizer.The parameter tuning of the ELM-AE model is performed using CBA.Fig.1 depicts the overall process of proposed CBADL-BEESC technique.

Figure 1:Overall process of CBADL-BEESC technique

2.1 Feature Extraction Using AlexNet Model

At the initial stage, the AlexNet model can be utilized to generate useful set of feature vectors.AlexNet is a kind of CNN that includes several layers like max pooling, input, dense, output, and convolution layers which are their fundamental structure blocks.It resolves the issue of images classifier in which the input image is one of 1000 distinct classes and the outcome is vector of individual classes[15].Thekthcomponent of resultant vectors are assumed that possibility the input image goes tokthclass.It is noticeable the sum of probability of the total resultant vectors is always equivalent to 1.AlexNet gets an RGB image as input containing the size of 256 * 256 that implies each image from the train and test set required to contain the size of 256*256.When the input image fails from equivalent to the typical image size,afterward it requires that changed to the typical size,for instance,256 * 256 before utilize for training the networks.Once the input image utilized is grayscale image,next it can be changed to RGB with replicate the single channel as to 3 channels RGB images.The infrastructure of AlexNet is reformed in the CNN technique that is utilized to computer vision issues and is significantly greater than CNN.AlexNet is 60 million parameters and 650,000 neurons that get a long period to train.

2.2 Classification Using ELM-AE Model

During classification process, the ELM-AE model is applied to classify the EEG EyeState.Autoencoder (AE)is an ANN approach that is usually employed from deep learning (DL).AE is an unsupervised neural network(NN),the resultants of AE are identical to inputs of AE,and AE is a variety of NNs that reproduce the input signals as possible.The ELMAE projected by Kasun et al.is a novel approach of NN that reproduces the input signal and AE.This technique of ELMAE established input, single-hidden, and resultant layers [16].Anjinput layer node,nhidden layer (HL)node,jresultant layer node,and the HL activation function g(x).Due to the resultant of HL representing the input signal,ELMAE was separated as to 3 distinct demonstrations as follows.

? j=n: Equal Dimension Representations: it can be implied feature in an input signal space dimensional equivalent to feature space dimensional.

? j>n: Compressed Representations: it signifies features in a superior dimension input signal space to lesser dimension feature space.

? j<n:Sparse Representations:it demonstrates features in a minimum dimension input signal space to maximum dimension feature space.

There are 2 variances amongst ELMAE and typical ELM.Primarily, ELM is a supervised NN and the resultant of ELM was labeled,however,ELMAE is an unsupervised NN and the resultant of ELMAE is similar to the input of ELMAE.Secondary, the input weight of ELMAE is orthogonal and the bias of HLs of ELMAE is also orthogonal[17],however,ELM is not so.In order toNvarious instances,xi∈Rn×Rj,(j=1,2,...,N),the resultants of ELMAE technique HL was formulated as(1),and the numerical connection amongst the resultant of HLs and the resultant of resultant layer is written as(2):

Employing ELMAE for attaining the resultant weightVis also separated as to 3 steps, yet the computation technique of resultant weightVof ELMAE in Step 3 has distinct in the computation method of resultant weightsVof ELM.Fig.2 demonstrates the structure of ELM method.

For compressed and sparse ELMAE representation,the resultant weight V can be computed by Eqs.(3)and(4).If the amount of trained instances are greater than the amount of HL nodes,

Once the amount of trained instances are lesser than the count of HL nodes,

To equivalent dimensional ELMAE representation,resultant weightsVcan be computed as:

Figure 2:ELM structure

2.3 Parameter Tuning Using CBA

For tuning the parameters involved in the ELM-AE model,the CBA is applied to it.The purpose of the compact process is for stimulating the operation of population-based BA method [18] in a form with small stored memory.The actual population of solution of BA is converted to the compact process through creating a dispersed data structure, i.e., perturbation vector(PV).PVdenotes the probabilistic method to population of solution.

Whereasδandμdenote variables of standard deviation and mean of vectorPV,andtindicates the existing time.Theδandμvalues are organized in the possibility density function(PDF)and are truncated from 1 and 1.The PDF amplitude can be standardized by keeping its region equivalent to one through gaining around satisfactory in well it is the standard distribution with a complete shape.

A real-value prototype vector was utilized to maintain sampling probability for making arbitrary components of candidate solutions.The vector process is distributed-based evaluated distribution approach (EDA)[19].The probability is that the evaluated distribution will trend, driving novel candidate forwarding to the FF.The candidate solution is probabilistically created from the vector,and the component from the best solution is utilized for making slight variations to the probability in the vector.The candidate solutionχiequivalent to the position of the virtual bat is created as(μi,Δi).

WhereasP(x)denotes the likelihood distribution ofPVwhich formulate a truncated Gaussian PDF related toμandδ.New candidate solutions are created as biased iteratively to a potential region of optimum solution.Then attain all the components of likelihood vector through learning the preceding generation.Theerfdenotes the error function.The codomain ofCDFis ordered within 0 to 1.TheCDFdetermine the real-value arbitrary parameterxwith provided likelihood distribution,and the gained values are lesser than or equivalent toxi.

Also,CDF specifies the distribution of multi-variate arbitrary parameters.Therefore,the relations of PDF andCDFare described byPV accomplishes the sampling parameterxithrough creating an arbitrary value within(0,1).This corresponds to obtaining the parameter through the computation of inverse operation ofχjset to inverse(CDF).

For finding an optimal individual from the procedure of compact approach, a comparison of these parameters is implemented.These parameter agents of bat are 2 sampling individuals that accomplished from PV.The “winner” specifies the vector with fitness score is maximum when compared to others,and the“loser”shows that low fitness assessment.The parameters,winner,and loser, are from main function valuation which compared a candidate solution with the preceding optimal global.To updatePV,μandδare taken into account on the basis of subsequent rules.When the mean value ofμwas assumed as 1, the updating rule for the element is,set forwarding to,as follows:

WhereasNpsignifies virtual population.Regardingδvalue,the updating rule of element as:

Generally,a probabilistic method for compact BA was applied for representing the bat solution set in which the position or velocity is saved,but a recently created candidate is saved.A variableωis utilized as a weight for controlling the likelihood ofμisampling in PDF among left[-1,μi]i.e.,PL(x),for-l ≤x≤μj,and right[1]i.e.,PR(x),forμi≤x ≤1.The extended edition of PDF for sampling method is employed.The generated new candidate of bat is applied by sampling fromPVfor example ifr <ωit produces coefficientxi∈[1,0]forPL(x),if notxi∈[1,0]forPR(x)

3 Experimental Validation

The presented CBADL-BEESC model is simulated using the benchmark database from UCI repository (available at https://archive.ics.uci.edu/ml/datasets/EEG+Eye+State).It comprises 14980 samples with two classes namely eye closed with 6723 samples (class 1)and eye open (class 0)with 8257 samples.

The confusion matrix offered by the CBADL-BEESC model with five distinct runs is portrayed in Fig.3.The figures demonstrated that the CBADL-BEESC model has resulted in effective EEG EyeState classification on all runs.For instance,with run-1,the CBADL-BEESC model has identified 6626 instances under class 1 and 8148 instances under class 2.In addition,under run-2,the CBADLBEESC model has recognized 6624 instances under class 1 and 8156 instances under class 2.Along with that, run-5, the CBADL-BEESC model has identified 6627 instances under class 1 and 8163 instances under class 2.

Figure 3:Confusion matrix of CBADL-BEESC technique different runs

The classifier results obtained by the CBADL-BEESC model under distinct runs are portrayed in Tab.1 and Fig.4.The results demonstrated that the CBADL-BEESC model has resulted in enhanced outcomes under every run.For instance,under run-1,the CBADL-BEESC model has obtainedprecnof 0.9838,recalof 0.9856,Fscoreof 0.9847,andaccuyof 0.9862.Eventually,under run-3,the CBADLBEESC method has gainedprecnof 0.9856,recalof 0.9851,Fscoreof 0.9853, andaccuyof 0.9868.Meanwhile,under run-1,the CBADL-BEESC approach has obtainedprecnof 0.9860,recalof 0.9857,Fscoreof 0.9859,andaccuyof 0.9873.

Table 1:Result analysis of CBADL-BEESC technique with distinct runs interms of various measures

Figure 4:Result analysis of CBADL-BEESC technique interms of various measures

Fig.5 illustrates the ROC analysis of the CBADL-BEESC system on the test dataset.The figure revealed that the CBADL-BEESC technique has gained improved outcomes with the increased ROC of 99.9391.

Tab.2 provides a brief comparative analysis of the CBADL-BEESC model with recent methods under distinct folds.

Fig.6 investigates the classifier results analysis of the CBADL-BEESC model with recent models under two-fold.The results indicated that the CBADL-BEESC model has resulted in enhanced outcomes over the other ML models.In addition,the SVM,BGA,and RF models have accomplished lower classification outcomes over the other methods.Moreover, the ET and KNN methods have obtained moderately closer classification outcomes.Furthermore, the proposed CBADL-BEESC model has accomplished superior outcome with the higherprecnof 96.42%,recalof 98.40%,Fscoreof 97.65%,andaccuyof 98.24%.

Figure 5:ROC analysis of CBADL-BEESC technique

Table 2: Comparative analysis of CBADL-BEESC technique with recent approaches under distinct folds

Fig.7 examines the classifier results analysis of the CBADL-BEESC system with recent techniques under five-fold.The results demonstrated that the CBADL-BEESC model has resulted in enhanced outcomes over the other ML technique.Besides, the SVM, BGA, and RF models have accomplished lower classification outcomes over the other algorithms.In addition, the ET and KNN techniques have obtained reasonably closer classification outcomes.Eventually, the projected CBADL-BEESC methodology has accomplished superior outcome with the superiorprecnof 97.97%,recalof 98.52%,Fscoreof 98.34%,andaccuyof 98.63%.

Figure 6:Comparative analysis of CBADL-BEESC technique under two-fold

Figure 7:Comparative analysis of CBADL-BEESC technique under five-fold

Fig.8 defines the classifier results analysis of the CBADL-BEESC technique with recent approaches under ten-fold.The outcomes referred that the CBADL-BEESC model has resulted in maximum outcome over the other ML models.Likewise, the SVM, BGA, and RF models have accomplished lesser classification outcomes over the other methods.Followed by,the ET and KNN approaches have gained moderately closer classification outcomes.At last, the presented CBADLBEESC method has accomplished higher outcome with the higherprecnof 96.55%,recalof 98.80%,Fscoreof 97.64%,andaccuyof 98.50%.

Finally, a comparative analysis of the CBADL-BEESC model is made with recent methods in Tab.3 and Fig.9.The results indicated that the SVM and BAG models have reached lower average accuracy of 88.43% and 89.68% respectively.In line with, the RF model has resulted in slightly increased average accuracy of 90.48%.Along with that, the KNN and ET models have accomplished moderately improved average accuracy of 93.51%and 92.65%respectively.However,the CBADL-BEESC model has outperformed the other methods with an average accuracy of 98.44%.By looking into above mentioned tables and figures, it is ensured that the CBADL-BEESC model has outperformed other models in terms of different measures.

Figure 8:Comparative analysis of CBADL-BEESC technique under ten-fold

Table 3: Average accuracy analysis of CBADL-BEESC technique with existing approaches

Figure 9:Average accuracy analysis of CBADL-BEESC technique with recent methods

4 Conclusion

In this study, a novel CBADL-BEESC technique has been developed for biomedical EEG EyeState classification.The main goal of the CBADL-BEESC technique is to recognize the presence of EEG EyeState.The CBADL-BEESC model involves AlexNet feature extractor,ELM-AE classifier,and CBA parameter optimizer.At the initial stage,the AlexNet model can be utilized to generate useful set of feature vectors.During classification process,the ELM-AE model is applied to classify the EEG EyeState.For tuning the parameters involved in the ELM-AE model, the CBA is applied to it.The parameter tuning of the ELM-AE model is performed using CBA.The experimental result analysis of the CBADL-BEESC model is carried out on benchmark results and the comparative outcome reported the supremacy of the CBADL-BEESC model over the recent methods.Therefore,the CBADL-BEESC model has appeared as an effective tool for EEG EyeState classification.In future,the classification outcome can be boosted by hybrid DL approaches.

Funding Statement:The authors extend their appreciation to the Deanship of Scientific Research at King Khalid University for funding this work under grant number(RGP 2/180/43).Princess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2022R161),Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.The authors would like to thank the Deanship of Scientific Research at Umm Al-Qura University for supporting this work by Grant Code:(22UQU4310373DSR04).The authors would like to acknowledge the support of Prince Sultan University for paying the Article Processing Charges(APC)of this publication.

Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

主站蜘蛛池模板: 久久久久亚洲av成人网人人软件| 欧美亚洲综合免费精品高清在线观看| 91久久天天躁狠狠躁夜夜| 久久综合AV免费观看| 欧美翘臀一区二区三区| 91精品专区国产盗摄| 亚洲日本中文字幕乱码中文| 亚洲精品亚洲人成在线| 欧美激情伊人| 国产成人一区免费观看| 亚洲日韩第九十九页| 国产一级妓女av网站| 国产精品久久久久鬼色| 黄色片中文字幕| 亚洲欧美成人综合| 国产精品亚洲天堂| 日日拍夜夜操| 欧美激情福利| 日本手机在线视频| 亚洲欧洲国产成人综合不卡| 黄色三级毛片网站| 色综合天天视频在线观看| 精品三级在线| 在线亚洲精品自拍| 全部毛片免费看| 98超碰在线观看| 全部毛片免费看| 国产原创自拍不卡第一页| 中国丰满人妻无码束缚啪啪| 日韩欧美在线观看| 成人在线不卡视频| 日韩欧美在线观看| 毛片免费高清免费| 狠狠综合久久| 57pao国产成视频免费播放| 精品一区二区三区视频免费观看| 无码视频国产精品一区二区| 91精品人妻一区二区| 亚洲 欧美 日韩综合一区| 欧美性天天| 亚洲中文字幕久久无码精品A| 人妻无码中文字幕一区二区三区| 久久久久国产精品熟女影院| 国产91小视频| 欧美人与牲动交a欧美精品 | 久久精品亚洲中文字幕乱码| 久久伊人操| 国产aaaaa一级毛片| 亚洲 日韩 激情 无码 中出| 欧美午夜视频在线| 六月婷婷激情综合| 日韩av电影一区二区三区四区 | 国产福利一区二区在线观看| 久久精品一品道久久精品| 国产无码制服丝袜| 亚洲午夜久久久精品电影院| 无码专区国产精品第一页| a级毛片视频免费观看| 亚洲swag精品自拍一区| 91精品专区国产盗摄| 一级毛片在线免费看| 一级黄色欧美| 亚洲一区网站| 天堂亚洲网| 色婷婷狠狠干| 色欲综合久久中文字幕网| 中文毛片无遮挡播放免费| 日韩久久精品无码aV| 国产日韩精品欧美一区灰| 亚洲成人网在线播放| 国产情侣一区| 欧洲高清无码在线| 国产乱子伦精品视频| 日韩人妻少妇一区二区| 999精品免费视频| 国产亚洲精品自在线| 国产网站免费观看| 亚洲美女视频一区| 国产成人一区| 国产在线第二页| 国产黄在线观看| 欧美人在线一区二区三区|