999精品在线视频,手机成人午夜在线视频,久久不卡国产精品无码,中日无码在线观看,成人av手机在线观看,日韩精品亚洲一区中文字幕,亚洲av无码人妻,四虎国产在线观看 ?

An Intelligent Gestational Diabetes Diagnosis Model Using Deep Stacked Autoencoder

2021-12-15 07:07:24SumathiMeganathanandVijilaRavisankar
Computers Materials&Continua 2021年12期

A.Sumathi,S.Meganathan and B.Vijila Ravisankar

1SRC,SASTRA Deemed to be University,Kumbakonam,Tamil Nadu,612001,India

2Consultant Obstetrician and Gynecologist,Anbu Hospital and Nalam Clinic,Kumbakonam,Tamil Nadu,612001,India

Abstract: Gestational Diabetes Mellitus (GDM) is one of the commonly occurring diseases among women during pregnancy.Oral Glucose Tolerance Test (OGTT) is followed universally in the diagnosis of GDM diagnosis at early pregnancy which is costly and ineffective.So,there is a need to design an effective and automated GDM diagnosis and classification model.The recent developments in the field of Deep Learning(DL)are useful in diagnosing different diseases.In this view,the current research article presents a new outlier detection with deep-stacked Autoencoder(OD-DSAE)model for GDM diagnosis and classification.The goal of the proposed OD-DSAE model is to find out those mothers with high risks and make them undergo earlier diagnosis,monitoring,and treatment compared to low-risk women.The presented ODDSAE model involves three major processes namely, preprocessing, outlier detection, and classification.In the first step i.e., data preprocessing, there exists three stages namely,format conversion,class labelling,and missing value replacement using k-nearest neighbors (KNN) model.Outliers are superior values which considerably varies from other data observations.So, it might represent the variability in measurement,experimental errors or novelty too.So, Hierarchical Clustering (HC)-based outlier detection technique is incorporated in OD-DSAE model, and thereby classification performance can be improved.The proposed model was simulated using Python 3.6.5 on a dataset collected by the researcher themselves.A series of experiments was conducted and the results were investigated under different aspects.The experimental outcomes inferred that the OD-DSAE model has outperformed the compared methods and achieved high precision of 96.17%,recall of 98.69%,specificity of 89.50%,accuracy of 96.18%,and F-score of 97.41%.

Keywords: Gestational diabetes mellitus; data classification; outlier detection;deep learning; hierarchical clustering

1 Introduction

Gestational Diabetes Mellitus (GDM) is characterized by uncertain glucose level in blood stream at the time of pregnancy [1].In earlier gestation fasting, the blood glucose level becomes low and insulin secretion also gets reduced gradually.This phenomenon is applied in the progressive improvement of insulin resistance in next trimesters with a marginal enhancement of insulin generation or hyperinsulinemia.Further, insulin resistance exists in case of placental hormones too.Pre-defined pathophysiologic systems, along with pregnancy, tend to change the metabolism in human body and enables maximum amount of postprandial maternal glucose.Pregnancy is one of the hyperinsulinemic conditions which creates an imbalance in glucose secretion, when insulin secretion becomes irregular to equalize pregnancy-based insulin protection.GDM can be described as chronic low-grade subclinical infection classified by irregular generation of cytokines,mediators followed by initiation of inflammatory signaling pathways.Even though GDM is considered to be resistant to insulin, the accurate process behind this model is still unexplored and challenging.Therefore, the increased level of insulin resistance in pregnancy is defined as cortisol and gestational hormones; however, the recent studies confirmed the involvement of cytokines too in this application [2].The presence of type 2 diabetes, among females with previous GDM (pGDM), is diagnosed during postpartum period with a maximum threshold.Further, the developers have found that the females with pGDM have high chances of developing type 2 DM within a limited period after pregnancy.Diabetes is one of the irreversible diseases which is accompanied by CVD mostly.In addition to the above, women with GDM are also prone to heart attacks, obesity, hypertension, dyslipidemia, and subclinical atherosclerosis.Those patients under metabolic anomalies with GDM are likely to have type 2 diabetes and it is a result of natural course of disease that eventually causes high risk CVD.

Early diagnosis and prediction tend to reduce the existence of GDM and commit minimum adverse pregnancy results [3].But, the survey conducted earlier mentioned that GDM cases are confirmed in a limited period of time using Oral Glucose Tolerance Test (OGTT).This leaves an optimal window of contribution for both fetal as well as placental deployment.OGTT is mostly recommended during early pregnancy across the globe.However, it is expensive and provide inaccurate results in most of the cases.GDM manifests during the period of mid-to-late pregnancy [4].So, earlier detection is essential to reduce further health risks.The establishment of simple prediction model, using previous medical data, among women who are at high risks of GDM helps in identifying mothers who require earlier diagnosis and treatment to obviate OGTTs.The currently-developed prediction models for GDM were introduced under the application of traditional regression model.In this scenario, Machine Learning (ML), a data analysis model that develops methods to predict results by ‘learning’from data, has been emphasized as a competing alternative to regression analysis.Further, ML is capable of performing well than regression,in terms of capturing nonlinearities as well as complicated communications between prediction attributes.Therefore, the current study employed ML methods for GDM prediction and no other models were selected that are relevant to Logistic Regressions (LR).

1.1 Outlier Detection Problem

Outlier or anomaly prediction is one of the significant objectives in ML and data mining methods.In the literature [5], the prediction of an outlier is considered to be the problem of pattern identification in data, which is not the actual behavior.Outlier prediction domains are also involved in intrusion detection, credit fraud investigation, video tracking, climate detection,identification of cybercrimes in electronic commerce, and so on [6].Additionally, some other types of outliers exist such as point outliers, contextual outliers and collective outliers.The prediction of point outlier has been employed in various other domains too.A data set has multiple points whereas a point is referred to as an outlier when it is differentiated from a massive number of points.In order to predict the outliers, some of the effective models have been applied earlier namely, classification modules, nearest neighbor schemes, clustering, statistical approach, distancebased approaches, and so forth.In classification-based outlier prediction, there are two classes present such as multiclass and single-class anomaly prediction approaches.Initially, multi-class classifiers have adopted the training dataset with labeled normal class label points.Next, the learner who applies the supervised learning trains another method under the application of labeled data.Classifier is generally applied to differentiate normal class from the remaining class.Alternatively, in single-class outlier’s prediction, only a single normal class classification contributes in learning and predicting the boundary of a normal class.When a test point is irrelevant to the boundary, then it is considered as an outlier.

1.2 Paper Contribution

The contribution of this research work is as follows.This research article presents a new outlier detection model i.e., deep stacked Autoencoder (OD-DSAE) model for GDM diagnosis and classification.The presented OD-DSAE model aims at identifying high-risk mothers who mandatorily need early diagnosis, monitoring, and treatment compared to low-risk women.The presented OD-DSAE model performs GDM diagnosis following three sub-processes such as preprocessing, outlier detection, and classification.Primarily, data preprocessing occurs in three stages namely, format conversion, class labelling, and missing value replacement using k-nearest neighbors (KNN) model.Next to that, Hierarchical Clustering (HC)-based outlier detection technique is incorporated in OD-DSAE model due to which classification performance gets improved.The detailed simulation analysis was conducted to verify the superior performance of the presented OD-DSAE model.

2 Related Works

This section reviews several diagnosis models used in the detection of GDM.Xiong et al.[7]aimed at developing a first-19 weeks’risk detection method with numerous potential for GDM detectors and used Support Vector Machine (SVM) as well as Light Gradient Boosting Machine(lightGBM).Zheng et al.[8] developed a simple method to predict GDM among Chinese women during their earlier pregnancy with the help of biochemical markers and ML model.In the literature, Shen et al.[9] examined the possibility of the best AI method for GDM examination in a setting that requires limited clinical equipment and clinicians.The study also developed an app based on AI scheme.

In the literature [10], the detection of GDM with diverse ML methods has been implemented on PIMA dataset.The accuracy of diverse ML methods was verified with measures.The importance of ML techniques was depicted under the application of confusion matrix, Receiver Operating Characteristic (ROC) and AUC values in managing diabetes PIMA dataset.Srivastava et al.[11] proposed a statistical method for evaluation of Gestational Diabetes Mellitus under the application of Microsoft Azure AI services.It is an ML Studio that yields excellent performance while its algorithm works on the perception of drag and drop.The classifier used in this process to detect the existence of GDM relied on the aspects occur during earlier phases of pregnancy.Cost-Sensitive Hybrid Model (CSHM) and five traditional ML models were applied in the development of prediction methods [12] to capture the upcoming threats of GDM in temporally-collected EHRs.After the completion of data cleaning, few data was recorded and gathered for a data set.

In the study conducted earlier [13], Radial Basis Function Network (RBFNetwork) was designed, estimated for performance and compared with ANN model to find feasible cases of GDM that produce harmful effects to pregnant women and fetus.In Ye et al.[14], parameters have been trained in various ML and classic LR methods.In Du et al.[15], three different classifiers were used to predict the target in case of future infection.The prediction accuracy guides the physician to make a better decision and regular prevention.Finally, it is identified that the DenseNet method predicts the target as gestational diabetes with minimum flexibility.

The major limitation of predominant classification models is that these frameworks depend upon exact labels, for normal class labels, which are highly complicate to use in real-time scenarios.In case of nearest neighbor-related outlier prediction schemes, some of the considerations are‘points’that belong to dense regions whereas the outliers come under sparse regions.Local Outlier Factor (LOF) mechanism is one of the well-established models to date.The basic principle of LOF depends on local density estimation of score points.Every point is allocated to a score in the ratio of mean local density of k-nearest neighbors (k-NN) of the point against the local densities of data point.Thus, the major constraints of this model are O(n2) complexity.Clustering-relied outlier prediction models make use of clustering technologies to collect the data as clusters.But,these points do not come under the clusters named ‘outliers’.

The main aim of clustering is to identify the clusters; thus, the outliers are considered to be the result of clustering operation which is not optimized appropriately.The root cause behind this strategy is the complex nature of clustering approaches in O(n2).In case of statistical outlier prediction, these approaches rely upon normal data point in high probability sites of stochastic method while the anomaly exists in low possibility regions.Generally, the statistical models are suitable for Gaussian distribution which is a combination of parametric statistical distribution that offers data and statistical inference test used to estimate the unknown sample under this method.But, the major disadvantage of this approach is the interruption of data points.Hence, in case of maximum dimension of data, the hypothesis becomes false.For distance-based outlier’s prediction,a point has been assumed to be an outlier when it is composed of minimum sufficient points than a defined threshold value [16].

3 The Proposed OD-DSAE Model

Fig.1 shows the working principle of OD-DSAE model.The presented OD-DSAE model involves three major processes namely, preprocessing, outlier detection, and classification.Initially,data preprocessing occurs in three stages namely, format conversion, class labeling, and missing value replacement.In addition, Hierarchical Clustering (HC)-based outlier detection technique is employed to remove the unwanted instances present in the dataset.Outliers are superior values and it varies considerably from other data observation.So, it might represent variability in measurement or experimental errors or in terms of novelty.So, outlier detection technique is incorporated in the OD-DSAE model, thereby classification performance can be improved.Finally,DSAE is applied as a classification model to determine an appropriate class label for GDM.

Figure 1:Overall process of OD-DSAE model

3.1 Data Preprocessing

At this stage, the input medical data is preprocessed to improve data quality in three ways.Firstly, the data conversion process occurs during when the input data in .xls format is converted into .csv format.Secondly, class labeling process is carried out during when the data instances are allocated to corresponding class labels.Thirdly, missing value replacement is performed using KNN technique.KNN is a simple and efficient technique that stores all the existing cases while at the same time, it categorizes the new cases based on a similarity measure.In this scenario, KNN model is deployed as a tool for data imputation.The principle behind KNN method is provided herewith [17]:

? Evaluate the variable k:Here, the value of the parameter k is set as 5.When the value of k is minimum, then maximum noise occurs which in turn mitigates the accuracy during classification.At the same time, when it is high, it results in value reduction and indirectly influences the accuracy.

? The Euclidian distance between the missing values should be determined with the help of Eq.(1).

whered(x,y)implies Euclidian distance, J means data attribute withj=1,2,3,...s, s refers to data dimensions,xajdefines value from j-attribute along with missing data, andybjrepresents the value from j-attribute with no missing data,

? According to the distance details attained, the lower Euclidian distance relies on a parameter k which is calculated as an imputed value in missing data.Hence, the value of imputation is determined using Weight Mean Estimation technique, as given in Eq.(2).

wherexjdenotes the Weight Mean Estimation, K implies the variable count applied with k=5,wkrefers to the nearest neighbour observation measure, andvkindicates the value from complete details on attribute with missing data related to variable k.In this model, the function ofwkhas been measured using Eq.(3).

whered(x,y):Euclidian distance of every variable k.

3.2 HC-Based Outlier Detection Technique

Once the data is pre-processed, HC-based outlier detection technique is executed to remove the outliers exist in the data.The proposed HC algorithm operates in a bottom-up manner.Here,the clusters are merged till an individual cluster is attained and the entire procedure gets iterated to determine the optimal cluster count that fits the data [18].It begins with the consideration of every instance as a single cluster, and consequently necessitate |X|-1 merging steps.To reduce the number of merging steps, HC algorithm normally utilizes a low complexity initial clustering in the generation ofclusters,?|X|.The initial clustering produces small-sized clusters and considerably reduces the computation complexities of HC technique.

Fig.2 shows the processes involved in HC.In HC model, the clusters are merged in such a way that the quadratic mutual information gets maximized.At this point, two approaches are available for merging such as agglomerative clustering and split-and-merge clustering.

Figure 2:Processes involved in HC

3.2.1 Agglomerative Clustering

In this method, the modifications are estimated on quadratic mutual data after the integration of pair of clusters and identification of optimal clusters.The pair is selected which enhances the quadratic mutual information.The clusters with minimum distortion have emerged which are applied in the optimization of distortion-rate function.Consider that the clustersχAandχBare combined to generateχc=χA∪χBwhile the modifications in quadratic mutual data,are determined as,

?

3.2.2 Split and Merge Clustering

In contrast to agglomerative clustering, split-and-merging clustering technique is applied to detect the cluster in a hierarchy for elimination process.The cluster exerts a poor impact on quadratic mutual data.This denotes that when the clusters are eliminated, then mutual information could be enhanced.Assume the clusterχAis composed of worst impact on mutual data and the quadratic mutual information gets changed whileis determined as:

The instances of poor cluster are allocated to residual clusters of clustered space depending on minimum Euclidean distance, where the closest samples are allocated initially.These processes are followed until one remaining cluster is reached.Followed by, on the basis of drastic extensions in quadratic mutual data from various hierarchiesthe actual number of clusters is estimated.Hence, the pseudo code for split-and-merge clustering model is defined in algorithm 2.

Algorithm 2:Pseudo code for the split and merge clustering 1:Initialization of Clustering, |?X|=N 2:for r=1:N-1 do 3:Determine Δ?I(t+1)A for every pair 4:Discard cluster χA, where χA=maxAΔ?I(t+1)A 5:for i=1:|χA|do 6:Assign samples xi=mini,kd(xi,?x ∈χk) to cluster χk 7:end for 8:end for 9:Estimate # of clusters

3.3 DSAE-Based Classification Technique

Figure 3:Structure of AE

DSAE model is executed to allocate the class labels of input data, i.e., presence of GDM or not.Basically, autoencoder (AE) is a type of unsupervised learning and is composed of three layers namely, input, hidden, and output layers as depicted in Fig.3.The operation of AE training is comprised of two portions namely, Encoder and Decoder.Generally, encoder is applied in mapping the input data to hidden implication, while decoder is considered for redevelopment of input data from hidden representation.It is provided as unlabeled input datasetwhere xn∈Rm×1,hnindicates the hidden encoder vector estimated from xn, anddenotes the decoder vector of the resultant layer [19].Therefore, the encoding operation is as given below.

where f implies the encoding function, W1refers to the weight matrix of the encoder, and b1defines a bias vector.The decoder operation is described as follows

where g represents the decoding process, W2implies the weight matrix of the decoder, and b2indicates a bias vector.The parameter setting of AE undergoes optimization to reduce the reformed error:

where L implies a loss functionAs depicted in Fig.4 [20], the architecture of SAEs undergoes stackingnAEs intonhidden layers using unsupervised layer-wise learning method and fine-tuned using a supervised model.

Figure 4:Structure of SAE

Thus, SAE-relied models can be classified using three procedures as given herewith

? The first AE is trained with the help of input data and gain the learned feature vector

? The feature vector of the previous layer is applied as input for the next layer and is followed till the training process is completed.

? Once the hidden layers are trained, Backpropagation (BP) model is applied to maximize the cost function and upgrade the weights with labelled training set and accomplish the fine-tuning process.

3.3.1 Dropout

Dropout is an efficient principle that limits the overfitting process involved in NN training.Most of the overfitting issues exist only when the training set is minimum and it intends to generate minimum accuracy on the test set.Dropout affects the neurons randomly with a hidden layer and the energy gets drained during training process; however, the weights of the neurons are maintained.Moreover, dropout is capable of accomplishing the output date of hidden neurons to 0 and the neurons could not be related to forward propagation task.The developers have sampled the impact of dropout to reduce overfitting issues for minimum training set.It is employed to maximize the feature extraction and classification accuracy of SAEs during GDM diagnosis.

3.3.2 ReLU Function

In case of classical activation functions like sigmoid and hyperbolic tangent functions, the gradients get reduced immediately with training error and propagate it to forward layers.Rectified LLinear units(ReLU) achieves maximum concentration from developers in the recent times since the gradient tend not to reduce the independent parameter.Thus, the system with ReLU is free from gradient diminishing problem.Therefore, ReLU function can be represented as given herewith.

4 Performance Validation

The presented model was simulated in Python 3.6.5 with additional packages such as tensorflow-gpu==1.14.0, pyqt5==5.14, pandas, scikit-learn, matplotlib, prettytable, seaborn, tqdm,numpy==1.16.0, and h5py==2.7.0.The snapshots of the results yielded during the simulation process are shown in the appendix.The presented OD-DSAE model was experimentally validated using the GDM dataset constructed by researcher themselves.It has a total of 3525 instances with the existence of 15 features.In addition, a set of two class labels exists in the dataset.Particularly, 2153 instances belong to class 0 and 1372 instances belong to class 1.Tab.1 shows the information related to the dataset.Fig.5 shows the frequency distribution of attributes that exist in GMD dataset.

Figure 5:Frequency distribution of GDM dataset

Table 1:Dataset description

Fig.6 depicts the confusion matrix generated by DSAE model when classifying GDM dataset.The figure portrays that the DSAE model classified 2032 instances under class 0 and 1111 instances under class 1.These values inferred the proficient classification outcome of the presented DSAE model.

Figure 6:Confusion matrix of DSAE

Fig.7 illustrates the confusion matrix generated by OD-DSAE model when classifying GDM dataset.The figure shows that the DSAE model classified 2107 instances under class 0 and 716 instances under class 1.These values exhibit that the presented OD-DSAE model effectually classified more number of instances in comparison with DSAE model.In order to emphasize the outlier detection performance of C technique, the results were examined in terms of correctlyand incorrectly-clustered samples as shown in Tab.2.From the table, it is evident that the HC technique has correctly clustered 2135 instances under class 0 and 800 instances under class 1.At the same time, the number of incorrectly clustered instances under classes 0 and 1 were 18 and 572 respectively.Thus, the HC technique is inferred to have correctly clustered a total of 2935 instances and incorrectly clustered 590 instances.

Figure 7:Confusion matrix of OD-DSAE

Tab.3 provides the results of comparative analysis of the presented OD-DSAE model against traditional ML models using GDM dataset.Fig.8 showcases recall, precision, and specificity analyses of the presented OD-DSAE method on applied GDM dataset.The attained values signify that the Voted Perceptron model produced minimal classification performance with a recall of 65.1%, precision of 86.9%, and specificity of 63.86%.In line with this, the Logit boost model accomplished a somewhat higher recall of 77.61%, precision of 84.6%, and specificity of 75.46%.At the same time, the DT methodology yielded an even better recall of 79.02%, precision of 81.4%, and specificity of 75.13%.Along with that, the LR model achieved a moderate diagnostic outcome with a recall of 79.27%, precision of 88%, and specificity of 75.32%.Also, the NN model gained superior recall of 79.84%, precision of 83.2%, and specificity of 76.78%.However, it failed to outperform the proposed DSAE and OD-DSAE methods.Finally, the presented DSAE model accomplished an effective recall of 94.38%, precision of 88.62%, and specificity of 80.98%.But,the OD-DSAE model attained the maximum recall of 98.69%, precision of 96.17%, and specificity of 89.5%.

Table 2:Outlier detection results

Table 3:Performance evaluation of different methods on GDM dataset

Figure 8:Result of OD-DSAE model using GDM dataset in terms of precision, recall, and specificity

Fig.9 depicts the accuracy and F-score analyses of the presented OD-DSAE model on applied GDM dataset.The obtained values signify that the Voted Perceptron model accomplished the least classification performance with an accuracy of 65.1% and F-score of 78.86%.Afterward,the DT model showcased a slightly higher accuracy of 73.82% and F-score of 80.19%.Simultaneously, the Logit Boost model yielded an even better accuracy of 74.08% and F-score of 80.96%.Along with that, the NN model produced a moderate diagnostic outcome with an accuracy of 75.39% and F-score of 81.48%.Moreover, the LR model attained a higher accuracy of 77.21%and F-score of 83.41%.However, these models failed to outperform the presented DSAE and OD-DSAE models.At last, the presented DSAE model obtained an effective accuracy of 89.16%and F-score of 91.41% whereas the OD-DSAE model accomplished an even higher accuracy of 96.18% and F-score of 97.41%.

Figure 9:Accuracy and F-score analysis of OD-DSAE model on GDM dataset interms of accuracy and F-score

Tab.4 and Fig.10 illustrate the detailed results of comparative analysis of OD-DSAE model upon the applied GDM dataset [21].The resultant values in the table demonstrate that the ODDSAE model outperformed all other methods.The existing KNN model produced an inferior classifier outcome over other methods by accomplishing the least accuracy of 67.6%.At the same time, the NB, ELM, and SGD models obtained moderate accuracy values of 74.9%, 75.72%, and 76.6% respectively.Followed by, the MLP model exhibited somewhat better accuracy i.e., 81.9%.

Table 4:Comparison with recent methods on GDM dataset in terms of accuracy

Figure 10:Accuracy analysis of OD-DSAE model on GDM dataset

Eventually, hybrid and J48 (pruned) models produced reasonable and closer accuracy values of 86.6% and 89.93% respectively.Besides, the AMMLP and HPM models attempted to produce reasonable outcomes and achieved accuracies of 89.93% and 92.38%.Concurrently, K-means +LR model presented a near-optimal accuracy value of 95.42%.However, the presented OD-DSAE model showcased its supremacy and achieved the maximum accuracy of 96.18%.After observing the above-discussed tables and figures, it is evident that the presented OD-DSAE model has effective diagnostic outcomes over the compared methods.Therefore, it can be employed as an appropriate tool for diagnosis and classification of GDM.

5 Conclusion

The current research work presented a new DL-based GDM diagnosis and classification model i.e., OD-DSAE.The goal of the proposed OD-DSAE model is to identify high-risk mothers who need early diagnosis, monitoring, and treatment compared to low-risk mothers.The presented OD-DSAE model has three major processes namely preprocessing, outlier detection, and classification.At first, data preprocessing was performed in three stages namely, format conversion,class labeling, and missing value replacement.The presented model made use of HC-based outlier detection technique to remove the unwanted instances from the dataset.Finally, DSAE was applied as a classification model to determine the appropriate class label of GDM.A detailed simulation analysis was performed on our dataset and the results were investigated under different aspects.The presented OD-DSAE model outperformed other models and achieved a precision of 96.17%, recall of 98.69%, specificity of 89.50%, accuracy of 96.18%, and F-score of 97.41%.These experimental outcomes infer that OD-DSAE model is a promising candidate for diagnostic applications.As a part of future work, the classification performance can be further enhanced using feature selection models.

Funding Statement:The authors received no specific funding for this study.

Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

Appendix

Result Generation of DSAE

Result Generation of Outlier Detection Model using Hierarchical Clustering

Result Generation of OD-DSAE

主站蜘蛛池模板: 亚洲人成电影在线播放| 四虎亚洲精品| 亚洲av片在线免费观看| a毛片基地免费大全| 亚洲精品无码成人片在线观看| 无码有码中文字幕| 99久久精品国产自免费| 欧美综合区自拍亚洲综合天堂| 亚洲精品国产综合99久久夜夜嗨| 久久香蕉欧美精品| 亚洲国产天堂久久综合| 色婷婷亚洲综合五月| 国产精品综合久久久 | 国产欧美日韩va另类在线播放| 精品综合久久久久久97超人该| 熟女成人国产精品视频| 波多野结衣亚洲一区| a毛片在线播放| 欧美翘臀一区二区三区| 国产黄色免费看| 婷婷丁香在线观看| 在线精品亚洲一区二区古装| 91麻豆国产视频| 国产免费怡红院视频| 久久福利片| 一级全黄毛片| 波多野结衣国产精品| 黑人巨大精品欧美一区二区区| 亚洲区欧美区| 无码精品福利一区二区三区| 中文字幕av一区二区三区欲色| 日本不卡在线播放| 欧美精品亚洲精品日韩专区va| 国产毛片高清一级国语| 国产白浆视频| 久久国产精品无码hdav| 久久精品最新免费国产成人| 999精品色在线观看| 丝袜美女被出水视频一区| 亚洲国产欧美目韩成人综合| 国产精品lululu在线观看| 99热国产这里只有精品无卡顿"| 亚洲精品午夜无码电影网| 婷婷中文在线| 国产成人综合亚洲欧美在| 91精品国产麻豆国产自产在线| 四虎精品国产AV二区| 欧美www在线观看| 国产精品久久自在自线观看| 欧美在线综合视频| 婷婷亚洲视频| 男人天堂亚洲天堂| 国产乱码精品一区二区三区中文 | 国产迷奸在线看| 国产成人综合在线观看| 久草网视频在线| 久久亚洲黄色视频| 亚洲国产综合精品一区| 国产亚洲精品资源在线26u| 国产女人在线观看| 国产精品天干天干在线观看| 国产毛片不卡| 日本精品一在线观看视频| 欧美精品1区| 亚洲中文无码av永久伊人| 2024av在线无码中文最新| 2021国产v亚洲v天堂无码| 日韩毛片基地| 国产女人水多毛片18| 欧美一级在线看| 色偷偷男人的天堂亚洲av| 国产成人精品一区二区秒拍1o| 青青青国产精品国产精品美女| 国产老女人精品免费视频| 亚洲一区波多野结衣二区三区| 国产成人超碰无码| 久久99国产乱子伦精品免| 国产精品原创不卡在线| 无码专区第一页| 老司国产精品视频91| 成人国产免费| 亚洲青涩在线|