999精品在线视频,手机成人午夜在线视频,久久不卡国产精品无码,中日无码在线观看,成人av手机在线观看,日韩精品亚洲一区中文字幕,亚洲av无码人妻,四虎国产在线观看 ?

Intelligent Sign Language Recognition System for E-Learning Context

2022-11-11 10:47:30MuhammadJamilHussainAhmadShaoorSulimanAlsuhibanyYazeedYasinGhadiTamaraalShloulAhmadJalalandJeongminPark
Computers Materials&Continua 2022年9期

Muhammad Jamil Hussain,Ahmad Shaoor,Suliman A.Alsuhibany,Yazeed Yasin Ghadi,Tamara al Shloul,Ahmad Jalal and Jeongmin Park

1Department of Computer Science,Air University,Islamabad,44000,Pakistan

2Department of Computer Science,College of Computer,Qassim University,Buraydah,51452,Saudi Arabia

3Department of Computer Science and Software Engineering,Al Ain University,Al Ain,15551,UAE

4Department of Humanities and Social Science,Al Ain University,Al Ain,15551,UAE

5Department of Computer Engineering,Korea Polytechnic University,Siheung-si,237,Gyeonggi-do,Korea

Abstract: In this research work, an efficient sign language recognition tool for e-learning has been proposed with a new type of feature set based on angle and lines.This feature set has the ability to increase the overall performance of machine learning algorithms in an efficient way.The hand gesture recognition based on these features has been implemented for usage in real-time.The feature set used hand landmarks, which were generated using media-pipe(MediaPipe)and open computer vision(openCV)on each frame of the incoming video.The overall algorithm has been tested on two well-known ASLalphabet(American Sign Language)and ISL-HS(Irish Sign Language)sign language datasets.Different machine learning classifiers including random forest, decision tree, and na?ve Bayesian have been used to classify hand gestures using this unique feature set and their respective results have been compared.Since the random forest classifier performed better, it has been selected as the base classifier for the proposed system.It showed 96.7%accuracy with ISL-HS and 93.7%accuracy with ASL-alphabet dataset using the extracted features.

Keywords: Decision tree; feature extraction; hand gesture recognition; landmarks;machine learning;palm detection

1 Introduction

Human computer interaction(HCI)applications are now getting more and more attention from researchers around the world.This field deals with various interactions of humans with computers.Modern world applications require humans to interact with computers in many different ways to easily perform their everyday tasks[1,2].Speech and vision signals are increasingly being used as inputs to many HCI speech recognition and computer vision-based applications.Human machine interaction(HMI)deals with the ways in which humans can interact with the machines.The machines are trained on speech data and therefore are usually interacted with using speech commands (e.g., Siri, Alexa).The same logic applies to the systems trained on vision-based inputs(e.g.,autonomous vehicles)[3].In vision-based applications, the given scenes are understood intelligently by the machines, such as,in crowed anomaly systems [4] or human object interaction recognition systems [5].Among visionbased inputs, there are many methods which humans use to interact with computers.These inputs are widely accepted and are most convenient when humans intend to give signals to machines.In these applications, human body shapes, facial expressions [6], hand gestures, and posture changes generate inputs for the systems to recognize and perform tasks [7] accordingly.This understanding or recognition of human gestures by machines is known as human gesture recognition(HGR)[8].In noisy environments,such as,factory floors,the best way of interacting with machines is through hand gestures and vision-based systems.Also, there are some other environments, such as, hospitals and educational institutions where the speech inputs are inappropriate and can conflict with the regulations[9].In these cases,the best and most convenient way to interact is through gestures[10].Sign language through human hand gesture recognition is considered as one of the most effective ways in which human can interact with machines.This method is also useful for a community consisting of special individuals with speaking disabilities.

With these motivations, an intelligent sign language recognition tool has been developed to recognize hand gestures in real time [11].The proposed system effectively extracts gestures from colored video frames.First,preprocessing is performed on every frame of the input video collected at the rate of 18 to 35 frames per second using webcam.Then,media-pipe hands pipelines detects palms and generates landmarks on hands.These landmarks are then used to extract the features which are discussed in detail in the following sections.The feature extraction process consists of two types of features based on angles and lines obtained from 21 landmarks.They are light weight in a sense that they do not require high computational power and ML classifiers are used to train the proposed model on these features,making it useful for real-time situations.

The rest of the paper is organized as follows: Section 2 describes the related research work and Section 3 presents the proposed methodology.Section 4 discusses the details of the used datasets and the performed experiments.The conclusion of the paper and the future work is discussed in Section 5.

2 Related Work

Hand gesture recognition is not a new idea in the field of HCI.There are two types of gestures:static and dynamic gestures[12].Gestures that remain the same for a given time frame are called static gestures while dynamic gestures keep changing in a given time frame,such as,waving.Some common approaches for hand gesture recognition include the use of electromyography (EMG)[13], cameras[14], and wearable gloves [15].Asif et al.[16] used a convolutional neural network (CNN)to detect and recognize hand gestures.After 80-100 epochs,the system begins to understand the gestures from the superficial electromyography(sEMG)data recorded from 18 subjects.In[17],a signal processing and sliding window approach was used to recognize hand gestures through electromyography(EMG)and artificial neural network (ANN)with an accuracy of 90.7%.In [18], a real-time hand gesture recognition model was introduced that collected data from a Myo armband worn on the subject’s forearm.The model is based on k-nearest neighbor(KNN)and dynamic time warping algorithms.The model performed with a high accuracy of 86%with 5 classes of gestures.In[19],a unique approach was presented to minimize the unnecessary or redundant information of EMG and also to increase the performance of real-time recognition of hand gestures through principal component analysis(PCA)and generalized regression neural network (GRNN)with 95.1% accuracy.The recognition of hand gestures through ultrasonic sensors and smart phones has attracted the attention of many researchers during the past few years[20].An online machine learning solution,known as“wisture”,can recognize hand gestures on smartphones.It has been trained using a recurrent neural network (RNN)with an accuracy of up to 93% with three hand gestures [21].Panella et al.[22] discussed the problems in recognizing the hand gestures through hand segmentation in devices like smart phones with less resources.They introduced a new and efficient ML algorithm, which is capable of recognizing the hand gestures through Hu image moments[23].

Extracting features from hand is quite a difficult task.Numerous researchers proposed different methods and techniques for this purpose.Oprisescu et al.[24]suggested that the gestures input should be taken from depth and time-of-flight(Tof)cameras.The gestures are recognized using the decision tree classifier and an accuracy of 93.3%was achieved on 9 different gestures.Yun et al.[25]proposed a multi-feature fusion and template matching technique for classification of hand gestures.Ahmed et al.[26] used dynamic time wrapping to recognize 24 hand gestures of Indian sign language with an accuracy score of 90%.Pansare et al.[27] proposed an alphabet American sign language recognizer(A-ASLR)based on real-time vision and the ASL alphabets dataset which obtained an accuracy of 88.26%.Ansar et al.[28] used a point-based full hand-based feature extraction method.Gray wolf optimizer was used to optimize features and genetic algorithm was used to classify hand gestures.Various landmark extraction techniques are being developed these days.Shin et al.[29] extracted features using 21 hand landmarks.These features were angle and distance based through landmark extracted through MediaPipe (media-pipe).A combination of support vector machine (SVM)and light gradient boosting machine(GBM)was used to classify the hand gestures.Costa et al.[30]also used media-pipe for landmark extraction but they obtained bounding boxes around the hands by selecting the top and bottom landmarks.For classification,they used SVM and achieved 90%accuracy.However, the methods which are distance and area based become problematic when tested in realworld situations using 2D cameras.There is a high chance that there would be very diverse feature values for distance and area-based methods for a single gesture.

3 Method and Materials

This section explains the proposed methodology in detail.Apart from the two datasets,a camera has also been used to record hand gestures and the obtained videos have been converted into image frames.The frames have been passed through a pre-processing phase using open computer vision(openCV)library which reads the video frame by frame.These frames have been sent to media-pipe which locates landmarks in each frame.The feature extraction module has used these landmarks to extract features.Then the extracted feature set has been passed to train classifiers.An overview of the system is shown in Fig.1.

3.1 Preprocessing

The OpenCV library [31] has been used to read the dataset image frames.The ASL-alphabet(American Sign Language)and ISL-HS(Irish Sign Language)datasets has been selected for this work.A publicly available pipeline known as MediaPipe[32]has been used for generation of landmarks on hands.This pipeline uses two models: a palm detection model that works on the whole frames and returns the region of interest(ROI).The ROI contains the cropped hand in the returned frame from detection model.It becomes input for landmark model (also known as joint locator model)which works on this ROI and generates 3D landmarks on hand via regression.The MediaPipe models are trained on 30000 colored images.This pipeline returns the values of pixels on which landmarks are located by models,i.e.,x and y coordinates on image as shown in Fig.2.

Figure 1:A general overview of the proposed system design

Since21 landmarks are generated on each hand,each landmark could be accessed via a specific id from 0 to 20.The MediaPipe assigns a token or id on each landmark as shown in Fig.3.These landmarks can help extract different features including the distance and angle between any two landmarks.

3.2 Features Extraction

The angles and lines are the feature extracted in feature extraction module.

3.2.1 Angle Features

For extracting the angle features,the slopes between each pair of landmarks are calculated using Eq.(1).

where(xi,yi),(xj,yj)is a pair of landmarks andSi,jis the slope between them.These slopes are used to calculate the angles of landmarks using Eq.(2).

whereθi,jis the angle between the slope and the x-axis.For every slope,the angle is calculated using the above equation in the proposed method.Fig.4 shows the calculation of 3 different angles.

Figure 2:Media-pipe results on(a)Closed fist as“A”gesture,(b)“V”gesture and(c)“P”gesture

Figure 3:Landmark generation on hand and finger from 0 to 20

Figure 4:Angle feature representations including(a)Angles between landmarks 0 and 5,(b)Angles between landmarks 0 and 9 and(c)Angles between landmarks 0 and 6

3.2.2 Lines Features

The fingers are considered as lines in this type of feature extraction.The fingers are labeled from 0 to 4 as shown in Fig.5.The slopes of fingers are calculated using bottom and top landmarks for each finger using Eq.(1).

Figure 5:Line representation of 5 fingers

Angles between fingers are calculated based on slopes of fingers using Eq.(3).These angles between fingers are used as new features.Fig.6 shows the fingers as lines.Eq.(3)is used for calculation of angle between linesiandj.

whereθi,jis the angle between the slopesSiandSj(slopes of fingeriandj).The value ofiandjis 0,1,2,3,4 representing fingers.

Algorithm 1:Extraction of angle and line features between every pair of landmarks Input:dataset ASL-alphabet/ISL-HS Output:Extracted feature from hand pose for n=0:m(for every frame)1 Take each Frame and generate landmarks via media pipe.for i=0:p(for every possible pair)2 Compute slopes between a pairi of land mark.Using Eq.(1)3 Take slopes and compute angle via Eq.(2)and store it.4 Take slopes and calculate lines via Eq.(3)and store it.end end return()Extracted Features of angles and lines

Figure 6:Angles between different lines,(a)Angle between lines 0 and 3(b)Angle between lines 1 and 2 and(c)Angle between lines 1 and 3

After extracting the angle and line features,a dimensionality reduction technique is applied[33].More than half of the dimensionality of the feature vector is reduced in this step.Features were highly correlated(i.e.,angle from landmark 0 to landmark 1 and angle from landmark 1 to landmark 0).

3.3 Model Training

This feature extraction module is applied to ASL-alphabet and ISL-HS datasets for feature extraction.The generated features are loaded for model training.These ML classifiers are selected over deep learning methods because they are more interpretable.Interpretability is the important factor in this research work since we want to investigate the performance state-of-art ML classifiers with the proposed feature set.

3.4 Classifiers

The methodology is evaluated using different machine learning classifiers such as na?ve Bayes,decision trees,and random forest.

3.4.1 Na?ve Bayes

The na?ve Bayes is the first classifier used to train the model.It works best on independent features.For example,if feature B has a certain value,then it determines the probability of class‘A’using Eq.(4).

In our case,B is the feature set(angle and line values)and A is the gesture.Since B is the features set(f1,f2,f3,f4,f5...f440)according to the proposed scenario,Eq.(4)can be rewritten for class A as Eq.(5).

3.4.2 Decision Tree

The decision tree takes features as nodes of a tree called decision nodes.The leaf nodes of a decision tree are the classes.The root node is the feature with the highest information gain to reduce the interclass similarities.For every class,there is a different path followed on the tree to give accurate output.The decision at each node is dependent upon the value of the nodes or features traversed(e.g.,the value of F3 is less than equal to 1.2 and the next decision node is F2)for a particular gesture as shown in Fig.7.

Figure 7:Decision tree with feature set <F1,F2,F3,F4,and F6>,F1 is the root node with value 0.7

Information gain for root node can be calculated using Eq.(6).

where gain is the information gain for node or split andiis the child of the root node.It is calculated by subtracting entropy of all child nodes from the entropy of the node for which the information is being calculated (root node in above Eq.(6)).Entropy measures the purity of split.The higher the entropy,the harder it is to draw any conclusion from the available information.The entropy of a node can be calculated using Eq.(7)

where c represents a class andpirepresents the probability of that class with respect to the node on which the entropy is being calculated.

3.4.3 Random Forest

Random forest classifier is selected as the base classifier for classification of hand gestures in the proposed method.The random forest consists of multiple decision trees and in this approach,the tree count is 100.It is an ensemble-based classifier which does majority voting after collecting the recognized labels from each decision tree against a given sample.

4 Experimental Setting and Results

This section gives a brief description of the performance of the proposed system using five different classifiers on the two datasets.All these experiments were performed using pycharm integrated development environment (ide)for python.The material provided in this section of the paper will validate the system.First,it provides the description of the datasets and then gives the details of the various experiments conducted to establish the high performance of the proposed system on the basis of accuracy,precision,recall score,F1 score,learning curves,and heat maps.

4.1 Datasets Description

The two publically available well known datasets for sign language are selected.These include ASL alphabet dataset[34]and ISL-HS dataset[35].

4.1.1 ASL Alphabet Dataset

This dataset[17]contains 28 gestures:the alphabets of the English language(A,B... Y,Z)and the two signs of space and delete.The gestures are single handed(from right hand)performed under similar lighting conditions and backgrounds.Performing real-time hand gesture recognition on this dataset is challenging because there is high interclass similarity between different classes.The dataset contains 3000 colored images per gesture(a total of 87000 images for all 28 gestures).Few samples of this dataset as shown in Fig.8.

Figure 8:Samples of ASL alphabet dataset[34]includes gestures of A,B,C,D,E and F

4.1.2 ISL Dataset

This dataset[18]contains hand gestures for 26 alphabets of the English language.Every sign has 18 videos,so there are a total of 468 colored videos for all gestures and each of them is roughly 3 s long.We have taken only the first 60 frames from every video to generate features.The rationale behind this was that the orientation of landmarks will change with time between the samples of dynamic datasets and produce diverse values for a single gesture or class.This effects the overall performance of algorithm.Therefore, only the first 60 frames are taken so that orientation remains almost the same for every sample.Few samples of this dataset are available as shown in Fig.9.

Figure 9:Samples of ISL-HS dataset[35]includes gestures of A,B,C,D,E and F

4.2 Results

4.2.1 Learning Curves

With 10-fold cross validations, the learning curves of different classifiers are plotted.Fig.10 represents the learning curves for three different classifiers on ISL-HS of random forest; namely,random forest,decision tree and na?ve Bayes.

Figure 10: Accuracy score on y-axis and training examples on x-axis for (a)Random forest curve,(b)Decision tree curve and(c)Naive bayes curve

The cross-validation score in each of the curves is low and increases with time.For random forest,the curve shows very high performance as compared to the decision tree and the na?ve Bayes classifier.The training score is nearly maximum in random forest and decision tree but it decreases over time in case of na?ve Bayes.The learning curves for ASL-alphabet dataset as shown in Fig.11.

The cross-validation score in each of the curves is low and increases with time.For random forest,the curve shows very high performance as compared to the decision tree and the na?ve Bayes classifier.The training and cross-validation scores of the models are low on the ASL dataset as compared to the ISL dataset.This is because there is high interclass similarity in the ASL dataset as compared to the ISL dataset.

Figure 11: Accuracy score on y-axis and training examples on x-axis for (a)Random forest curve,(b)Decision tree curve and(c)Naive bayes curve

4.2.2 Recognition Accuracy,Precision,Recall,F1

Tabs.1 and 2 show the accuracy, precision, recall, and F1 scores obtained using five different classifiers;namely,Na?ve Bayes,decision tree,random forest,k-nearest neighbors(KNN)and support vector machine(SVM)over the ISL-HS and ASL alphabet datasets respectively.

Table 1: Precision,accuracy,recall and F1 score of different classifiers over the ISL dataset

Table 2: Precision,accuracy,recall and F1 score of different classifiers over the ASL dataset

From Tabs.1 and 2,it is conclusive that the random forest classifier works better with the proposed algorithm as compared to the other four classifiers.Also,it can be seen that Na?ve Bayes achieved the lowest accuracies because it works best with independent features.In the proposed method,however,the features are interdependent.For example,when the angle between landmark 0 and landmark 12(12 is on the tip of the middle finger)is changed,then its effect is also reflected over the other features in a way that the angle between landmark 12 and landmark 4 is also changed.Tabs.3 and 4 show the confusion matrices depicting the accuracies of the selected ASL alphabet and ISL-HS datasets respectively.

Table 3: Confusion matrix of random forest classifier on ASL dataset

Table 4: Confusion matrix of random forest classifier on ISL dataset

Table 4:Continued

In Figs.12 and 13, the results are validated through heat maps for ASL-alphabet and ISL-HS datasets respectively.

Figure 12:Heat map for ASL-alphabet dataset

While in Tab.5,the proposed system has been compared with other state-of-the-art methods and has shown the most promising results.

Figure 13:Heat map for ISL-HS dataset

Table 5: Comparison of the proposed method with other state-of-the-art methods on datasets

5 Conclusion

In this research work, an efficient system has been proposed which performed well on ASL alphabet and ISL-HS datasets.Among the five chosen classifiers,andom forest has shown the highest accuracy on the selected datasets.It achieved an accuracy of 93%on ASL alphabet dataset and 96.7%on ISL dataset.The proposed system is light weight and will perform well in variable environments.However, it shows limitations when it comes to dynamic datasets since the orientations of hand landmarks with each other in multiple frames of a dynamic class sample generate very different values for a single hand gesture label.

The authors are planning to add more features in addition to angles and lines to increase the system’s performance for dynamic and versatile environments.Moreover,they are also trying to devise a strategy to handle dynamic datasets.

Funding Statement:This research was supported by a Grant(2021R1F1A1063634)of the Basic Science Research Program through the National Research Foundation (NRF)funded by the Ministry of Education,Republic of Korea.

Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

主站蜘蛛池模板: h网站在线播放| 91在线播放免费不卡无毒| 亚洲日韩第九十九页| 亚洲成AV人手机在线观看网站| 无码内射在线| 福利一区在线| 亚洲国产成人综合精品2020| 少妇精品久久久一区二区三区| 亚洲一区第一页| 成人字幕网视频在线观看| 91青青草视频在线观看的| 亚洲永久色| 一级爆乳无码av| 亚洲一区第一页| 青青草一区| 日本欧美中文字幕精品亚洲| 亚洲美女一区| 欧美色视频网站| 亚洲最大福利网站| 69视频国产| yjizz国产在线视频网| 精品国产美女福到在线不卡f| 亚洲国语自产一区第二页| 国产麻豆精品在线观看| 四虎国产在线观看| 国产日本视频91| 亚洲成人精品久久| 成人在线第一页| 国产性生大片免费观看性欧美| 欧美日韩国产精品综合| 91久久精品日日躁夜夜躁欧美| 国产免费久久精品99re不卡| 亚洲狠狠婷婷综合久久久久| 国产精品太粉嫩高中在线观看| 欧美成人在线免费| 亚洲成肉网| 国产91丝袜在线播放动漫| 亚洲日本www| 国产成人a在线观看视频| 91探花国产综合在线精品| 真实国产乱子伦视频| 97se亚洲综合在线天天| 中文字幕人成人乱码亚洲电影| 国产一区二区色淫影院| 欧美在线视频不卡第一页| 久久精品无码国产一区二区三区| 精品少妇人妻无码久久| 呦女亚洲一区精品| 国产精品美女网站| 精品国产免费观看| 亚洲欧洲国产成人综合不卡| 亚洲手机在线| 在线视频一区二区三区不卡| 九九免费观看全部免费视频| 女人18毛片久久| 日韩第一页在线| 欧美国产三级| 亚洲va欧美va国产综合下载| 91午夜福利在线观看| 亚洲色欲色欲www在线观看| 直接黄91麻豆网站| 99这里只有精品在线| 久久久久免费精品国产| 亚洲综合色区在线播放2019| a毛片在线播放| 日韩成人在线一区二区| 亚洲欧洲日产国产无码AV| 99这里只有精品6| 99久久精品免费看国产免费软件| 亚洲最大情网站在线观看 | 婷婷六月色| 青青网在线国产| 欧美一级特黄aaaaaa在线看片| 亚洲国产中文欧美在线人成大黄瓜| 亚洲国产精品无码久久一线| 国产91全国探花系列在线播放| 日本久久网站| 青青草原国产免费av观看| 精品超清无码视频在线观看| 97se亚洲综合在线天天| 成人国产免费| 精品一區二區久久久久久久網站|