Hepatitis is a major public health problem all around the world. This paper proposes an automatic disease diagnosis system for hepatitis based on Genetic Algorithm (GA) Wavelet Kernel (WK) Extreme Learning Machines (ELM). The classifier used in this paper is single layer neural network (SLNN) and it is trained by ELM learning method. The hepatitis disease datasets are obtained from UCI machine learning database. In Wavelet Kernel Extreme Learning Machine (WKELM) structure, there are three adjustable parameters of wavelet kernel. These parameters and the numbers of hidden neurons play a major role in the performance of ELM. Therefore, values of these parameters and numbers of hidden neurons should be tuned carefully based on the solved problem. In this study, the optimum values of these parameters and the numbers of hidden neurons of ELM were obtained by using Genetic Algorithm (GA). The performance of proposed GAWKELM method is evaluated using statical methods such as classification accuracy, sensitivity and specivity analysis and ROC curves. The results of the proposed GAWKELM method are compared with the results of the previous hepatitis disease studies using same database as well as different database. When previous studies are investigated, it is clearly seen that the high classification accuracies have been obtained in case of reducing the feature vector to low dimension. However, proposed GAWKELM method gives satisfactory results without reducing the feature vector. The calculated highest classification accuracy of proposed GAWKELM method is found as 96.642 %.
1. Introduction
In medicine, the inflammation occured in liver is called hepatitis. It can be caused by infections with viruses, bacteria, fungi, exposure to toxins such as alcohol and autoimmunity. The hepatitis damages to cells of liver and tenderness, swelling and inflammation in the liver are some symptoms of this disease. Commonly, the liver can handle significant amounts of damage, and the liver function is still effective. However, it will decline if the disease is not fully controlled at an early stage. The hepatitis can be acute or chronic and it is a common disease over the world. The different types of hepatitis are caused by different things, but they all produce inflammation of the liver
[1]
. Viral hepatitis refers to several common contagious diseases caused by viruses that attack the liver. The most important types of viral hepatitis are hepatitisA, hepatitisB, and hepatitisC
[2]
. New forms of viral hepatitis such as D, E and G are also discovered. HepatitisA mainly infects by fecal contaminated substances taken by mouth. As a result of poor hygienic conditions, this results in epidemics based on water or food, especially in developing countries
[1]
. In infection of hepatitis A, any indication may be observed. Therefore, the patients could not become aware of the hepatitisA. HepatitisB virus could not infect easily such as hepatitisA. But, it is commonly transmitted to another person by transfusion. To decrease infection risk of hepatitisB virus, transfusion should be strictly controlled. Other infection ways of hepatitisB virus are drugs, sexsual intercourse and birth.
HepatitisC virus is also responsible from 80% of hepatitiss taking place after transfusion and it is the most severe type. It assaults the liver and caused swelling and redness in liver
[3]
. Nonviral forms of hepatitis can be caused by toxic agents (drugs or chemicals), alcohol, or autoimmune processes. Another form of hepatitis is toxic hepatitis
[1]
.
For inexperienced practitioners, one of the most important problems is to detect hepatitis virus
[2]
. The most common problems in the field of automatic diagnostic are selection of necessary parameter set for right diagnostics and forming of powerful algorithm which doesn’t require long time to run
[4]
. In addition, the doctors couldn’t correctly diagnosis diseases due to fatigue, inexperience and thoughtfulness. In diagnosis process, it can be useful to use an auxiliary tool which minimizes human based wrong diagnosis and helps inexperienced practitioners. In literature, different automatic classification systems have been proposed for diagnosis of hepatitis disease.
Different methods in literature such as Multilayer Neural Network (MLNN) with backpropagation (BP) and Radial Basis Function (RBF) with OLS in
[2]
, MLNN with Levenberg Marquardt (LM) in
[5]
, combined diagnosis system based genetic algorithm (GA) and artificial neural network (ANN) in
[6]
, Linear Discriminant Analysis (LDA) and Adaptive Network based on Fuzzy Inference System (ANFIS) in
[7]
, the application of ANN for forecasting the incidence of hepatitis A by using ARIMA model in
[8]
, (PCAANN ) in
[3]
, a hybrid diagnosis system using feature selection (FS) with fuzzy weighted preprocessing and artificial immune recognition in
[9]
, PCALSSVM are presented for automatic diagnosis of hepatitis disease in
[10]
.
NNs have been widely used in pattern recognition and regression. The NNs are commonly trained by backpropagation based on a gradientbased learning rule
[11]
. Up to now, the gradientbased learning methods have been widely applied for learning of NNs
[12

13]
. However, it has several shortcomings such as difficult setting of learning parameters, slow convergence, training failures due to local minima, and repetitive learning to improve performance of NNs. Also, it is clear that gradient descentbased learning methods are generally very slow
[14]
.
In Ref
[6]
, the GA is used for selection of feature subset for the input of ANN. Proposed method is not suitable for realtime implementation. Besides, the feature vector is randomly reduced to a lower dimension in Refs.
[3
,
7
,
9
,
10]
. ANFIS structure might not a good performance if a huge amount of data exists.
Recently, a new learning algorithm called Extreme Learning Machine (ELM) which randomly selected all the hidden nodes parameters of generalized Singlehidden Layer Feedforward Networks (SLFNs) and analytically determines the output weights of SLFNs is proposed in
[15

21]
. Altough output weights are analitically calculated, there is no rule in determination of number of hidden neurons and type of the kernel function. To obtain a good classification performance of ELM, these parameters should be determined properly. In here, the hepatitis diseases database is used
[22

25]
.
This paper proposes an automatic hepatitis diagnosis system for hepatitis using pattern recognition based on Genetic Algorithm (GA) – Wavelet Kernel (WK) based Extreme Learning Machines (ELM). The classifier used in this paper is single layer neural network (SLNN) and it is trained by ELM learning method. In Wavelet Kernel Extreme Learning Machine (WKELM) structure, there are three adjustable parameters of wavelet kernel. These parameters and the numbers of hidden neurons play a major role in the performance of ELM. Therefore, values of these parameters and numbers of hidden neurons should be tuned carefully based on the solved problem. In this study, the optimum values of these parameters and the numbers of hidden neurons of ELM were obtained by using Genetic Algorithm (GA). The hepatitis disease datasets are obtained from UCI machine learning database. The performance of proposed GAWKELM method is evaluated through statical methods such as classification accuracy, sensitivity and specivity analysis and ROC curves. In here, the numbers of hidden neurons of ELM and parameters of wavelet kernel function are optimized by GA. In GA structure, a individual composes of total of 20 bits. These are:
 The first four bits (1
^{st}
, 2
^{nd}
, 3
^{nd}
and 4
^{th}
bits) of each of these individuals represent the w parameter values (between 1 and 16) of the wavelet kernel functions.
 The second four bits (5
^{th}
, 6
^{th}
, 7
^{th}
and 8
^{th}
bits) of each of these individuals represent the x parameter values (between 1 and 16) of the wavelet kernel functions,
 The third four bits (9
^{th}
, 10
^{th}
, 11
^{th}
and 12
^{th}
bits) of each of these individuals represent the y parameter values (between 1 and 16) of the wavelet kernel functions,
 Rest of 20 bits represents number of hidden neurons (between 5 and 132).
The 40 number of these individuals are randomly chosen for initial population. Thus, it is purposed to obtain the best possible performance from ELM classifier. The training and testing dataset for proposed method is obtained from UCI dataset. These dataset compose of 155 data. The randomly selected 100 of 155 data are used for training of classifier whereas remaining data is used for testing of classifier. For different kernel function and number of hidden neurons, the results of proposed method are given. Further, a comparison is performed with previous studies to show validity of proposed method. From results, the proposed method is quite powerful tool for automatic diagnosis of hepatitis and may work in realtime systems.
In ref.
[27]
, Wang et al. claim that wavelet kernel functions can widen the kernel function selection range of Kernel ELM (KELM) and support its further development. The important reason of using wavelet kernel function is having adjustable w, x, and y paramaters. Thus, the wavelet kernel function gains adaptive features for best classification and test results in different data types [28]. In Wavelet Kernel Extreme Learning Machine (WKELM) structure, there are three adjustable parameters of wavelet kernel. These parameters and the numbers of hidden neurons play a major role in the performance of ELM. Therefore, values of these parameters and numbers of hidden neurons should be tuned carefully based on the solved problem. In this study, the optimum values of these w, y, x parameters and the numbers of hidden neurons of ELM were obtained by using Genetic Algorithm (GA). According to the results of this study, the wavelet kernel ELM has a better performance than classical ELM for automatic diagnosis system of hepatitis disease. At the same time, the wavelet kernel functions perform very well when the algorithm is compared with the classical kernel functions in ELM model. The comparing results are given in
Table 4
.
The paper is organized as follows. Section 2 presents pattern diagnosis concept. In section 3, Classic ELM and in section 4, Wavelet Kernel (WK) based Extreme Learning Machines (ELM) are briefly presented respectively. In section 5, the GA is explained. The description of hepatitis dataset used is given in Section 6. The application of GAWK ELM for hepatitis diagnosis is presented in Section 6. Finally, Section 7 provides conclusions of this study.
2. Pattern Recognition Concept
Pattern recognition could simply be defined as the process of categorizing the input signal. This process can be divided into two main groups as classification and regression. In the classification process, input signal is seperated into two or more classes while pattern recognition system tries to find the appropriate mapping between the input and the output variables in the regression process. Generally, regression problems are harder to solve than classification problems due to several reasons
[22]
.
Fig. 1
shows conventional pattern recognition concept for classification. As shown in
Fig. 1
, the pattern recognition concept composes of two stages. They include feature extraction and classification stages. The feature extraction stage is the most important part of pattern diagnosis. The appropriate features are extracted by a feature extractor. If the appropriate features are not selected, the classification performance will be poor even though using the best classifier. Therefore, feature extractor should reduce the dimension of pattern vector to a lower. The reduced feature vector should also cover useful information of the original vector. In last stage, reduced feature vector is given to inputs of classifier for the classification
[20
,
21]
.
The block diagram of the pattern recognition concept
3. Classic Extreme Learning Machine
The NNs have been widely used in pattern recognition and regression problems. Commonly, the learning of NN has been performed by using gradientbased learning algorithms. However, such methods has several drawbacks such as difficult setting of learning parameters, slow convergence, slow learning and training failures
[20
,
21]
.
To deal with the drawbacks of gradientbased learning methods, ELM was proposed by Huang, et al.
[20]
. In the ELM, the output weights of a singlehidden layer feedforward network (SLFN) are analytically computed by using the MoorePenrose (MP) generalized inverse instead of iterative learning scheme.
Fig. 2
shows structure of a SLFN using ELM. In this figure,
l
_{1m}
,
l
_{2m}
and
l
_{rm}
, are weights vector connecting the i
th
hidden neuron and the input neurons, and
w
is the weight vector connecting the i
th
hidden neuron and output neuron and
f
(.) is kernel function.
Structure of SLFN
The most important properties of ELM are given as follows:
 The learning speed of ELM is extremely fast. Therefore, SLFNs can be trained by ELM much faster than classical learning methods.
 The ELM tends to reach both the smallest training error and the smallest norm of weights. Thus, the ELM tends to have good performance for neural networks.
 The ELM learning algorithm can be used to train SLFNs with nondifferentiable kernel functions.
 The ELM tends to reach the solutions straightforward without such trivial issues
[20]
.
The outputs of a SLFN with m neurons and
f
kernel function are written as follows:
That standart SLFN with m hidden neurons and
f
kernel function can approximate these m samples with zero error means that
, i.e., there exist
w
,
l
and b such that
[11]
,
[20]
;
The m set of equations can be written as follows:
where; H is the hidden layer output matrix of the SLFN.
ELM method for SLFNs called ELM can be briefly explained as follows
[20]
:
Given a training set, kernel function
f
(.) and the number of hidden nodes m:
 Randomly assign input weight w and bias b, (i=1, 2, …, m).
 Find the hidden layer output matrix H .
 Find the output weight
w
.
4. Wavelet Kernel Based Extreme Learning Machines
The ELM learning algorithm has fast learning speed and good generalization performance. In recently years, number of researchers has increased
[4
,
7]
. The initial parameters of hidden layer need not be tuned in ELM algorithm. In this algorithm, all nonlinear piecewise continuous functions are used as the hidden neurons. Because of this, for M optional different sample data set
, the output function in ELM by using
K
hidden neurons is
In here,
v
= [
v
_{1}
(
r
),
v
_{2}
(
r
), …,
v_{K}
(
r
)] is the output vector of the hidden layer with respect to the input
r
.
S
= [
S
_{1}
,
S
_{2}
, …,
S_{K}
] is the vector of the output weights between the hidden layer of K neurons and the output neuron. v vector converts the data from input space to the ELM feature space
[7]
. In ELM structure, the training error and the output weights should be simultaneously minimized for decreasing the training error. Thus, generalization performance of neural networks improves.
There, Eq. (6) can be solved by using Eq.(7).
In here,
E
is the regulation coefficient,
A
is the hidden layer output matrix and
C
is the expected output matrix of samples. Thus, the output function of the ELM learning algorithm can be found as below:
If the feature vector
v
(
r
) is unknown, the kernel matrix of ELM based on Mercer’s conditions can be given as below:
thus, the output function
u
(
r
) of the wavelet kernel based extreme learning machine (WKELM) can be shown as:
In here,
D
=
AA^{T}
and
b
(
r
,
g
) is the kernel function of Extreme Learning Machine. In ELM literature, some kernel functions, which are linear kernel, polynomial kernel, Gaussian kernel, and exponential kernel, appropriate to the Mercer condition. The readers can be found more details in
[7
,
9
,
27
, 28]. In this paper, wavelet kernel function, which is given on Eq.(11), is used for simulation and performance analysis for WKELM.
The training and testing performance of the wavelet kernel function given in Eq. (11) is better than the performances of linear kernel, polynomial kernel, Gaussian kernel, and exponential classical kernel functions respectively. In wavelet kernel, the adjustable parameters w, x and y play a major role in the performance of neural networks. Therefore, values of these parameters should be tuned carefully based on the solved problem. Moreover, in WKELM algorithms, the hidden layer feature mapping need not be known and the number of hidden neurons need not be chosen. However, the WKELM learning algorithm obtains similar or better generalization performance. Most important WKELM is more stable compared to traditional ELM and it is faster than support vector machine (SVM)
[7
,
9]
.
5. Genetic Algorithm
Genetic Algorithms is good at taking large, potentially huge search spaces and navigating them, looking for optimal combinations of things, solutions you might not otherwise find in a lifetime
[22]
.
The GA provides efficient, effective techniques for optimization and machine learning applications and they have been widely used in business, scientific and engineering fields.
Fig. 3
shows components of a GA. As shown in this figure, components of a GA are encoding technique, initialization procedure, evaluation function, selection of parents, genetic operators, parameter settings. The GA begins with a set of solutions which are represented by individuals (chromosome). The individuals could be bit strings (0101 ... 1100), real numbers (43.2  33.1 ... 0.0 89.2), permutations of element (E11 E3 E7 ... E1 E15), lists of rules (R1 R2 R3 ... R22 R23), program elements (genetic programming) or any data structure. In reproduction, individuals of population are randomly selected with selection chances biased in relation to population evaluations. The evaluation step decodes a chromosome and assigns it a fitness measure. Also, the evaluation is the only link between a classical GA and the problem it is solving
[23]
. In genetic algorithm, the iterative process is repeated as long as during the new population is better than the old one. The higher the fitness value of an individual is, the more likely this individual, is reproduced for the next population. The iterative process is finished when some conditions (for example, number of individuals in the population, etc.) are satisfied
[24]
.
Components of a GA.
The basic genetic algorithm is summarized as follows:
Step 1
: A random population of n individuals is generated which is a suitable solution for the problem. The numbers of n are 40 in this study.
Step 2
: The fitness f(x) of each individual x is evaluated in the population [28]. In these experimental studies, each of individuals in population is randomly formed.
Step 3
: It is chosen two parental individuals from among the individuals has higher the fitness value in population. Then, the cross over operator is performed to these parental individuals. The purpose of cross over operator is obtaining the different individuals, which have higher fitness values.
Step 4
: It is cross overed the parents with a crossover probability to form new individuals. If crossover is not performed, individual will be the exact copy of parents.
Step 5
: It is mutated each new individual with a mutation probability at each locus, which is the position in the individual.
Step 6
: The new individuals are reconstructed from the new population.
Step 7
: The genetic algorithm is stopped, if the end conditions is satisfied, and return the best solution in current population.
Step 8
: It is gone to step 2 and used the newly generated population for a further run of the algorithm.
6. Data Description
The hepatitis data is obtained from UCI machine learning database. It contains 20 attributes including the class attribute
[26]
. The attribute information of dataset is given in
Table 1
.
The attribute information of dataset
The attribute information of dataset
The output shows whether patients with hepatitis are die or alive. Hepatitis dataset contains 155 data belonging to two different target classes. There are 20 features, 13 binary and 6 attributes with 68 discrete values. Also, the class distribution contains 32 cases for die and 123 cases for alive.
7. Application of GAWKELM for Diagnosis of Hepatitis
This GAWKELM method composes of three layer. In first layer of GAWKELM, the hepatitis data is obtained from UCI machine learning database mentioned in Section 5. In second layer of GAWKELM, the numbers of hidden neurons of ELM and parameters of wavelet kernel function are optimized by GA.
In this study, the used parameters, which are coding type, mutation rate, mutation type, crossover rate, crossover type, population size, and chromosome size for GA design are given in
Table 2
. These parameter values are the best values found after much trial.
The used parameters for GA design
The used parameters for GA design
The realized crossover sample in this study by using two point crossover can be given as below:
The realized crossover sample in this study
The realized mutation sample in this study by using bit inversion can be given as below:
The realized mutation sample in this study
In GA structure, a individual composes of total of 20 bits. These are:
 The first four bits (1
^{st}
, 2
^{nd}
, 3
^{nd}
and 4
^{th}
bits) of each of these individuals represent the w parameter values (between 1 and 16) of the wavelet kernel functions.
 The second four bits (5
^{th}
, 6
^{th}
, 7
^{th}
and 8
^{th}
bits) of each of these individuals represent the x parameter values (between 1 and 16) of the wavelet kernel functions,
 The third four bits (9
^{th}
, 10
^{th}
, 11
^{th}
and 12
^{th}
bits) of each of these individuals represent the y parameter values (between 1 and 16) of the wavelet kernel functions,
 Rest of 20 bits represents number of hidden neurons (between 5 and 132).
The 40 number of these individuals are randomly chosen for initial population. Thus, it is purposed to obtain the best possible performance from ELM classifier. The training and testing dataset for proposed GAWKELM method is obtained from UCI dataset. These dataset compose of 155 data. The randomly selected 100 of 155 data are used for training of classifier whereas remaining data is used for testing of classifier. For the optimimum numbers of hidden neurons of ELM and parameters values of wavelet kernel function, the results of proposed GAWKELM method are given. Further, a comparison is performed with previous studies to show validity of proposed GAWKELM method. From results, the proposed GAWKELM method is quite powerful tool for automatic diagnosis of hepatitis and may work in realtime systems.
The block diagram of the proposed GAWKELM method is given in
Fig. 6
. In these applications, a 3fold crossvalidation schema was applied where the twofifth data were used for training the proposed GAWKELM method and the remaining other data were used as the test data set. This strategy was applied for three times and the average values were calculated for determining the performance of proposed GAWKELM method.
The block diagram of GAWKELM based optimal diagnosis system
In genetic algorithm, the mean training error was used as fitness function. This mean training error was obtained from result of training of ELM for each of individuals by using parameters represented by these individuals. In the test process of GAWKELM, the numbers of hidden neurons of ELM and parameters of wavelet kernel function are optimized by GA. As said previously, the dataset has 19 relevant features except for class attribute and includes a total of 155 cases. Thus, it is a matrix with dimension of 155 x 19. The parameters of ELM are optimized by GA. Firstly, 40 random individuals which are formed from a total of 20 bits are chosen as the initial population. Coding for parameters of wavelet kernel functions and the number of hidden neurons are given in
Table 3
and
Table 4
, respectively.
Coding for kernel functions
Coding for kernel functions
Coding for number of hidden neurons
Coding for number of hidden neurons
An example for individuals of the population
An example for individuals of the population is shown as follows. The first four bits (1
^{st}
, 2
^{nd}
, 3
^{nd}
and 4
^{th}
bits) of each of this individual represent the w parameter values (between 1 and 16) of the wavelet kernel functions. The second four bits (5
^{th}
, 6
^{th}
, 7
^{th}
and 8
^{th}
bits) of each of this individual represent the x parameter values (between 1 and 16) of the wavelet kernel functions. The third four bits (9
^{th}
, 10
^{th}
, 11
^{th}
and 12
^{th}
bits) of each of this individual represent the y parameter values (between 1 and 16) of the wavelet kernel functions, the rest of 20 bits of this individual represents number of hidden neurons (between 5 and 132).
Performance of the proposed GAWKELM method is calculated by three evaluation methods as classification accuracy, sensitivity and specificity analysis and Receiver Operating Characteric (ROC) curves. The classification accuracies for the datasets are found using the following equation.
where, C is the set of hepatitis data to be classified (the test set), c.m is the class of item c and classify(c) returns the classification of the by ELM. The sensivity and specificity analysis are obtained as follows:
Where; TP, TN, FP and FN are true positives, true negatives, false positives and false negatives, respectively.
8. Obtained Results
In this study, an automatic intelligent system for diagnosis of hepatitis disease is presented. Performance of the proposed GAWKELM method is also evaluated by classification accuracy, sensitivity and specificity analysis and ROC curve.
The correct hepatitis diagnosis performances of classic ELM classifiers, which have different types of kernel function and the number of hidden neurons, are compared with the performance of the proposed GAWKELM method, which has optimum values of wavelet kernel w, x, y parameters and numbers of hidden neuron in these experimental studies. The results of these performance comparisons are given in
Table 4
. As shown in this table, the best classification accuracy of the proposed GAWKELM method is found as 97.5623% in case of values of w, x, y wavelet kernel function parameters, which are 15, 3 and 10, and hidden neuron with 86.
To show validity of proposed GAWKELM method, compared results with previous studies using same dataset are also given in
Table 5
. As shown in this table, the highest correct hepatitis diagnosis rate is obtained as 97.56% by using proposed GAWKELM method. Because, the optimum values of these parameters and the numbers of hidden neurons of ELM were obtained by using Genetic Algorithm (GA). Also, it is clearly shown that the obtained results from GRNN and RBF have been deteriorated in case of without GA. Similarly, other methods presented in Ref
[3
,
6
,
7
,
9]
and
[10]
use reduced feature vector for input of classifier. In this study, after finding optimum parameters, GA does not need to use and then the WKELM can be directly used. The correct hepatitis diagnosis performance comparing of GAWKELM method with Kernel ELM classifiers, which have different types of kernel function and the number of hidden neurons is given in
Table 5
.
The correct hepatitis diagnosis performance comparing of GAWKELM method with Kernel ELM classifiers, which have different types of kernel function and the number of hidden neurons.
The correct hepatitis diagnosis performance comparing of GAWKELM method with Kernel ELM classifiers, which have different types of kernel function and the number of hidden neurons.
The comparison results of proposed GAWKELM method and previous studies using same dataset.
The comparison results of proposed GAWKELM method and previous studies using same dataset.
The obtained hepatitis diagnosis accuracies by statistical evaluation criterias are given in
Table 7
.
The obtained hepatitis diagnosis accuracy by statistical methods.
The obtained hepatitis diagnosis accuracy by statistical methods.
A comparison with previous studies including different database is also given in
Table 8
. As shown in this table, the high classification accuracy has been obtained by proposed GAWKELM method
[22

25]
.
Comparing the correct diagnosis rate of GAWKELM based expert system and previous methods, which used different hepatitis database.
Comparing the correct diagnosis rate of GAWKELM based expert system and previous methods, which used different hepatitis database.
In this study, ROC curves and AUC values are calculated by using TP, TN, FP and FN are true positives, true negatives, false positives and false negatives, respectively
[22

27]
. A ROC curve is a graphical plot that illustrates the performance of a binary classifier system as its dis crimination threshold is varied. The curve is created by plotting the true positive rate against the false positive rate at various threshold settings. (The truepositive rate is also known as sensitivity in biomedical informatics or recall in machine learning. The falsepositive rate is also known as the fallout and can be calculated as 1  specificity). The ROC curve is thus the sensitivity as a function of the fallout.
In general, if the probability distributions for both detection and false alarm are known, the ROC curve can be generated by plotting the cumulative distribution function (area under the probability distribution from to) of the detection probability in the yaxis versus the cumulative distribution function of the falsealarm probability in the xaxis.
ROC analysis provides tools to select optimal models and to discard suboptimal ones independently from the class distribution or the cost context. ROC analysis is related in a direct and natural way to cost/benefit analysis of diagnostic decision making. In here, the ROC curve of GAWKELM is given by using obtained best TP, TN, FP and FN values in
Fig. 6
. The calculated AUC value of ROC curves by using GAWKELM classifier can be given as 0,9624.
9. Discussion and Conclusion
This paper proposes an GAWKELM based optimal intelligent system for diagnosis of hepatitis. The proposed GAWKELM hepatitis diagnosis system has advantages such as directly using of feature vector, fast training and testing time as well as generalization capability over conventional neural networks with back propagation. The proposed GAWKELM method is carried out in two stages as classification and optimization of classifier’s parameters. The feature vector from hepatitis dataset is used as input to WKELM classifiers. In Wavelet Kernel Extreme Learning Machine (WKELM) structure, there are three adjustable parameters of wavelet kernel. These parameters and the numbers of hidden neurons play a major role in the performance of ELM. Therefore, values of these parameters and numbers of hidden neurons should be tuned carefully based on the solved problem. In this study, the optimum values of these parameters and the numbers of hidden neurons of ELM were obtained by using Genetic Algorithm (GA). The output of WKELM decides make a classification. To obtain the best possible classification performance, the optimum values of parameters of classifier are determined by GA. The feasibility of the developed technique has been tested by data set of 192 test cases. The proposed GAWKELM technique shows effective classification performance when compared with previous studies depending on directly using of feature vector and training time as shown in Tables VVIII and
Fig. 8
.
The ROC curve of proposed GAWKELM method for intelligent hepatitis diagnosis.
BIO
Derya Avci She received B.S and Master degree in Computer Science from Firat University. She received Ph.D degree in Electrical and Electronic Engineering from Firat University. Her research interests are artificial intelli gence, image processing and signal processing.
Dogantekin Esin
,
Dogantekin Akif
,
Avci Derya
(2009)
Automatic hepatitis diagnosis system based on Linear Discriminant Analysis and Adaptive Network based on Fuzzy Inference System
Expert Systems with Applications
36
11282 
11286
DOI : 10.1016/j.eswa.2009.03.021
Dogantekin Esin
,
Dogantekin Akif
,
Avci Derya
2009
“Automatic hepatitis diagnosis system based on linear discriminant analysis and adaptive network based on fuzzy inference system”
Expert Systems with Applications
36
(8)
11282 
11286
DOI : 10.1016/j.eswa.2009.03.021
Jilani Tahseen A.
,
Yasin Huda
,
Yasin Madiha Mohammad
2011
“PCAANN for Classification of HepatitisC Patients”
International Journal of Computer Applications
14
(7)
Ultrasound and Magnetic Resonance Medical Images Textures Analysis Using Neural Networks
http://www.intermag.kiev.ua/projects/neuraln.html
(last accessed: 20 January 2012)
Bascil M. Serdar
,
Temurtas Feyzullah
(2011)
“A Study on Hepatitis Disease Diagnosis Using Multilayer Neural Network with Levenberg Marquardt Training Algorithm”
J Med Syst
35
433 
436
DOI : 10.1007/s1091600993782
Mahdieh Adeli
,
Zarabadipour Hassan
(2011)
“Automatic disease diagnosis systems using pattern recognition based genetic algorithm and neural networks”
Int. J. Phys. Sci.
6
(25)
6076 
6081
Li Bin
,
Rong Xuewen
,
Li Yibin
“An Improved Kernel Based Extreme Learning Machine for Robot Execution Failures”
Hindawi Publishing Corporation The Scientific World Journal
http://dx.doi.org/10.1155/2014/906546
Article ID 906546
2014
7 
DOI : 10.1155/2014/906546
Peng Guan
,
Huang DeSheng
,
Zhou BaoSen
2004
Forecasting model for the incidence of hepatitis A based on artificial neural network
China World Journal of Gastroenterol
10
(24)
3579 
3582
DOI : 10.3748/wjg.v10.i24.3579
Polat Kemal
,
Gunes Salih
(2006)
Hepatitis disease diagnosis using a new hybrid system based on feature selection (FS) and artificial immune recognition system with fuzzy resource allocation
Digital Signal Processing
16
889 
901
DOI : 10.1016/j.dsp.2006.07.005
Çalisir Duygu
,
Dogantekin Esin
2011
A new intelligent hepatitis diagnosis system: PCALSSVM
Expert Systems with Applications
38
(8)
10705 
10708
DOI : 10.1016/j.eswa.2011.01.014
Qeethara Kadhim AlShayea
(2011)
“Artificial neural net works in medical diagnosis”
International Journal of Computer Science Issues
8
(2)
150 
154
Helge Malmgren
,
Borga Magnus
,
Niklasson Lars
2012
Artificial Neural Networks in Medicine and Biology
Springer Science & Business Media
Proceedings of the ANNIMAB1 Conference
Göteborg, Sweden
1316 May 2000
Cho JaeHoon
,
Lee DaeJong
,
Chun MyungGeun
2007
“Parameter Optimization of Extreme Learning Machine Using Bacterial Foraging Algorithm”
EESRI
742 
747
Huang GuangBin
(2012)
“Extreme learning machine for regression and multiclass classification.”
Systems, Man, and Cybernetics, Part B: Cybernetics, IEEE Transactions on
42
(2)
513 
529
DOI : 10.1109/TSMCB.2011.2168604
Huang GuangBin
(2011)
“Extreme Learning Machines: A Survey,”
International Journal of Machine Leaning and Cybernetics
2
107 
1221
DOI : 10.1007/s130420110019y
Huang GuangBin
,
Chen Lei
,
Siew CheeKheong
(2006)
“Universal Approximation Using Incremental Constructive Feedforward Networks with Random Hidden Nodes”
IEEE Transactions on Neural Networks
17
879 
892
DOI : 10.1109/TNN.2006.875977
Huang GuangBin
“Extreme Learning Machine: A New Learning Scheme of Feedforward Neural Networks,”
2004 International Joint Conference on Neural Networks (IJCNN'2004)
Budapest, Hungary
July 2529, 2004
Liang NanYing
(2006)
“A Fast and Accurate Online Sequential Learning Algorithm for Feedforward Networks”
IEEE Transactions on Neural Networks
17
1411 
1423
DOI : 10.1109/TNN.2006.880583
Guo Ling
2011
“Automatic feature extraction using genetic programming: An application to epileptic EEG classification”
Expert Systems with Applications
38
(8)
10425 
10436
DOI : 10.1016/j.eswa.2011.02.118
Biswal Birendra
2009
“Nonstationary power signal processing for pattern recognition using HStransform”
Applied Soft Computing
9
(1)
107 
117
DOI : 10.1016/j.asoc.2008.03.004
UCI Machine Learning Repository
Available from: (last accessed: 20 September 2014)
Wang JiuYao
2013
“Wavelet kernel extreme learningclassifier”
Microelectron Comput
10
73 
7680