Advanced
Polynomial Fuzzy Radial Basis Function Neural Network Classifiers Realized with the Aid of Boundary Area Decision
Polynomial Fuzzy Radial Basis Function Neural Network Classifiers Realized with the Aid of Boundary Area Decision
Journal of Electrical Engineering and Technology. 2014. Nov, 9(6): 2098-2106
Copyright © 2014, The Korean Institute of Electrical Engineers
  • Received : January 27, 2014
  • Accepted : April 24, 2014
  • Published : November 01, 2014
Download
PDF
e-PUB
PubReader
PPT
Export by style
Article
Author
Metrics
Cited by
TagCloud
About the Authors
Seok-Beom Roh
Dept. of Electronic and Information Engineering, Wonkwang University, Korea. (nado@wonkwang.ac.kr)
Sung-Kwun Oh
Corresponding Author: Dept. of Electrical Engineering, The University of Suwon, Korea. (ohsk@suwon.ac.kr)

Abstract
In the area of clustering, there are numerous approaches to construct clusters in the input space. For regression problem, when forming clusters being a part of the overall model, the relationships between the input space and the output space are essential and have to be taken into consideration. Conditional Fuzzy C-Means (c-FCM) clustering offers an opportunity to analyze the structure in the input space with the mechanism of supervision implied by the distribution of data present in the output space. However, like other clustering methods, c-FCM focuses on the distribution of the data. In this paper, we introduce a new method, which by making use of the ambiguity index focuses on the boundaries of the clusters whose determination is essential to the quality of the ensuing classification procedures. The introduced design is illustrated with the aid of numeric examples that provide a detailed insight into the performance of the fuzzy classifiers and quantify several essentials design aspects.
Keywords
1. Introduction
Radial basis function (RBF) networks have been widely studied and applied to various tasks of regression and classification problem, cf. [16 , 19] . Since the concepts of RBF neural networks were introduced in the literature [20] , there have been a number of interesting and useful generalizations of the generic topology of these networks and their learning methods, cf. [1 , 4 , 12 , 14] . The visible feature of RBF neural networks comes with their fast two-phase training method. During this learning process, the values of the parameters of the radial basis functions are determined independently from the weight values of the output layer. Typically, the parameters of the basis functions (referred to as receptive fields) are estimated by some relatively fast and general methods of unsupervised learning applied to input data. After the basis functions have been determined, the output layer’s weights are obtained as the least-squares solution to a system of linear equations (e.g., by using the Moore-Penrose pseudo-inverse [2] ). Compared to the nonlinear optimization that is usually considered in the training of neural networks, this two-stage method is usually much faster and can help avoid local minima and eliminate difficulties with the convergence of the overall learning process [13] .
Let us look in more detail at this two-step design of RBF neural networks by highlighting the diversity of the optimization tools available there.
  • (a) Optimization of the hidden layer: We encounter a significant variety of radial basis functions being used and face with diverse ways of their development. Discussed are such typical forms of RBFs as Gaussian functions. Other analytical versions of such functions are also available; see[5]. An alternative way of dealing with the formation of the RBFs (receptive fields) is to exploit various clustering techniques including its commonly encountered representatives such as K-means and Fuzzy C-Means (FCM)[19]. Furthermore, the optimization method such as Particle Swarm Optimization can be used to position the RBFs in the input space, refer to[21].
  • (b) Optimization of the output layer: For learning scheme for the linear neuron located at the output layer of the network, gradient-based methods and Expectation Maximization (EM)-based training method are in common usage, see[9,11].
Let us consider a way in how to locate the receptive fields of RBF neural networks (i.e., how to analyze and describe the input space which is inherently related with the output space through some unknown function whenever the output space is the real numbers space present in case of regression problems or the space of integers used in classification tasks). As noted above, various unsupervised clustering methods such as K-means and Fuzzy C-Means have been proposed to construct receptive fields. In particular, for regression problems, Pedrycz [18] has pointed at a certain drawback of the original objective function based clustering techniques such as Fuzzy C-Means clustering. This shortcoming, which is commonly encountered when using the clustering methods based on the minimization of the objective function to form linguistic terms of a fuzzy model over the input space, is that all of those terms are formed in a completely unsupervised manner even though there is some component of supervision available which comes in the form of dependent (output) variables. To alleviate this shortcoming and take into account information about the output space, the Conditional Fuzzy C-means (c-FCM) clustering has been proposed. Given that the information about the output is used in the method, it brings some component of supervision to the clustering process.
In this study, we develop a concept of RBF neural networks based on the supervisory clustering (supervision-augmented clustering) which relates with supervisory clustering realized for regression problems. Furthermore, when dealing with classification problems, the supervisory clustering has to be activated within the boundary area occupied by patterns to be classified. We define the boundary area as a certain region of the input space where the data (patterns) belonging to different classes are located. Given a mixture of data coming from different classes or associated with a substantial variety of output’s values, the boundary region can be regarded as a source of useful discriminatory information. In contrast, the regions of the input space associated within the core of each class (where by the core we mean a region of the input space being predominantly occupied by patterns belonging to the same class) might be a limited source of discriminatory information.
In order to activate the supervised clustering within the boundary area, we describe this area by using several linguistic terms (quantified in terms of fuzzy sets). This approach is legitimate considering that fuzzy sets are naturally geared to describe concepts (here classes) exhibiting overlap with elements belonging to other classes. After determining the boundary area, we invoke supervisory clustering to analyze the structure of the space. The performance of the proposed classifier is contrasted with the results produced by polynomial Fuzzy Radial Basis Function Neural Networks (pFRBF NNs). To show the classification abilities of the proposed classifier preferred to the various types of classifiers, we compare the generalization ability of the proposed classifier with the well-known classifiers.
This study is organized as follows. In Section 2, we review the generic architecture of the generic RBF NNs and the extended RBF NNs. Next, in Section 3, we propose and elaborate on the pFRBF NNs classifiers focused on the boundary decision area and conditional fuzzy clustering. Extensive experimental studies are covered in Section 5 while Section 6 offers some concluding comments.
2. Architecture of the Extended pRBFNNs
Several researches have said that the generic pFRBF NNs exhibit some advantages including global optimal approximation and classification capabilities as well as rapid convergence of the underlying learning procedures, see [6 , 8] . The generic topology of pFRBF NNs is depicted in Fig. 1 .
PPT Slide
Lager Image
General architecture of the generic pFRBF Neural Networks
In Fig. 1 , Γ i , i=1, 2,…, c denotes receptive fields (radial basis functions), while “m” stands for the number of the input variables. The output of the generic pFRBF NN comes as a linear combination of the outputs ( Γ( x ) ) of the corresponding nodes at the hidden layer with the connection weights w 1 , w 2 , ⋯ , wc as shown below
PPT Slide
Lager Image
Where x = [ x 1 x 2 xm ] ∈ ℜ m and Γ i ( x ) is the activation level of the i-th node present at the hidden layer.
Generally, the Gaussian type pFRBFs are used as receptive fields
PPT Slide
Lager Image
where v i and σi are the apex (center) and the spread of the i th receptive field, respectively.
There are two major differences between the extended pFRBF NNs and the generic version of pFRBF NNs. The first one concerns the type of the underlying receptive fields. In the extended pFRBF NNs, the prototypes of the receptive fields (i.e., the nodes of the hidden layer) are determined by running fuzzy clustering. The output of each node in the hidden layer is an activation level of the corresponding linguistic term (fuzzy set)
PPT Slide
Lager Image
The second difference arises in terms of the type of the connection (weights) between the hidden layer and output layer. In the extended pFRBF NNs, we use linear functions or the 2 nd order polynomials rather than confining ourselves to some fixed numeric values. The architecture of the extended pFRBF NN and the type of connection weights considered above is shown in Fig. 2 .
PPT Slide
Lager Image
Architecture of the extended pFRBF Neural Networks
In Fig. 2 , fi denotes the connection (weight) between the i th node of hidden layer and the node in the output layer. The connection fi is expressed as a linear function or the 2 nd order polynomial. More specifically, we have
PPT Slide
Lager Image
Here, a i = [ a i0 a i1 aim ] T ∈ ℜ (m+1)
PPT Slide
Lager Image
Here,
PPT Slide
Lager Image
The activation level of each node in the hidden layer is determined using (3). The normalized activation level uik follows the expression
PPT Slide
Lager Image
The following relationship holds
PPT Slide
Lager Image
For the output node we obtain
PPT Slide
Lager Image
3. The Development of The pFRBFNN Classifier Activated Within the Boundary Area
When we consider a two class problem, as elaborated on in the introduction, we use the extended pFRBF NNs to be used as the primary classifier whose receptive fields are constructed by running supervised clustering (i.e., the conditional fuzzy C-Means).
The basic conjecture coming with the proposed classifier is that more ambiguous information is present within the boundary area than with the core area (generally speaking, the core area has homogeneous patterns belonging to the same class whereas the boundary area typically embraces patterns belonging to several classes).
The boundary surface is formed within boundary area. In this paper, the boundary surface (area) for each class is determined by using the extended pFRBF NNs as already presented in Section II. The output of extended pFRBF NNs is aggregated through a certain linear combination of the local models, which describe the relationship between the input variables and the output variables present within the related local areas. The local model (i.e., the linear function or the 2 nd order polynomial) of pFRBF NNs defines the local boundary surface, which is formed within the local area defined by the receptive field.
We anticipate that the improvement of classification performance becomes associated with the use of the receptive fields that are positioned within the boundary area.
- 3.1 Defining the boundary area
Let us recall that the boundary area pertains to the region in the input space in which we encounter patterns belonging to different classes. In contrast, the core area (region) is highly homogeneous where there are data belonging to the same class. Fig. 3 illustrates some core and boundary areas formed for the two-class data.
PPT Slide
Lager Image
Examples of core and boundary areas
In order to define the boundary areas by using linguistic terms, the data patterns involved in each class are previously analyzed by the Possibilistic C-Means (PCM) clustering. As far as data set is concerned, we consider a finite set of “n” input-output data coming in the form of the ordered pairs{ x k , gk } , k = 1, 2, ⋯ , n , x k ∈ ℜ m , while gk ∈ {1, 2, ⋯, l }, l is the number of classes. The output variable gk is the class label. Denote by Li the set of indices of the data pattern involved in the i-th class.
PPT Slide
Lager Image
The original FCM uses the probabilistic constraint meaning that the membership grades for the same data sum up to one. While this is useful in forming the partition, the membership values resulting from the FCM and the related methods, however, may not always correspond to the intuitive concept of degree of belongingness, compatibility or typicality (commonality) as noted in the literature. Krishnapuram and Keller relaxed this constraint and introduced possibilistic clustering (PCM) by minimizing the following objective function
PPT Slide
Lager Image
where ηi is a certain positive number, and “p” is a fuzzification coefficient that should be determined as any real number greater than 1 which is the same parameter as that is used in the ordinary FCM.
The first term requires that the distances from data points to the prototypes be as low as possible while the second term forces the values of uik to be as large as possible, thus avoiding running into a trivial solution. It is recommended to select ηi as discussed in [10] , that is
PPT Slide
Lager Image
Typically, the value of K is chosen to be equal to 1. The update of the prototypes is realized in the same way as this has been done in the FCM algorithm,
PPT Slide
Lager Image
The membership degree (partition matrix) in the PCM is calculated as follows
PPT Slide
Lager Image
We determine the prototypes and the activation levels for each class separately. The prototypes of class “j” (i.e. gk = j ) is calculated as follows
PPT Slide
Lager Image
PPT Slide
Lager Image
is the prototype of the i-th cluster of the j-class, nj is the number of elements of the index set Lj , and L j{k} means the k-th element of the index set Lj .
The activation level of the i-th cluster for the j-class is calculated as follows.
PPT Slide
Lager Image
Where
PPT Slide
Lager Image
is the specific ηi available in the j-th class.
Note that these expressions are the modified versions of (12) and (13).
The higher the activation levels (15) are, the more visibly the data is involved in the core area of the corresponding class.
After calculating the activation levels and prototypes for all classes, we define the boundary area as follows.
PPT Slide
Lager Image
Here T stands for some t-norm and S denotes a certain t-conorm (s-norm). In this study, the t-norm is realized as the minimum operator and the t-conorm is specified as the probabilistic sum.
PPT Slide
Lager Image
and
PPT Slide
Lager Image
denote the activation levels of the i-th cluster of “1” class and the j-th cluster of “2” class, respectively. As shown in Fig. 4 , with 2 classes where each class is composed of 2 clusters, the boundary area is defined in the following form
PPT Slide
Lager Image
Examples of the Boundary Area associated with the corresponding values of α-cuts of the fuzzy clusters (membership functions)
PPT Slide
Lager Image
- 3.2 Conditional fuzzy C-Means clustering within boundary area
The idea of Conditional Fuzzy C-Means (c-FCM, for short) clustering proposed in [18] was applied to the design of pFRBF neural networks as presented in [19] . To elaborate on the essence of the method, let us consider a set of patterns X = { x 1 , x 2 , ⋯ x N }, x k ∈ ℜ m (where m stands for the dimensionality of the input space) along with an auxiliary information granule, which is defined as the boundary area. Each element of X is then associated with the auxiliary information granule (fuzzy set) B given by (16).
In conditional clustering, the data pattern x k is clustered by taking into consideration the conditions (auxiliary information expressed in the form given by B ( x 1 ), B ( x 2 ), ⋯ , B ( x n ) ) based on some linguistic term expressed as a fuzzy set B ( B : ℜ→[0,1] ). The objective function used in the conditional fuzzy clustering is the same as the one used in the FCM, namely
PPT Slide
Lager Image
where J is the objective function, uik is the activation level associated with the linguistic term B defining the boundary area, v i is the i th cluster and c is the number of rules (clusters) formed for this context. The difference between the FCM and c-FCM comes in the form of the constraint imposed on the partition matrix where we now have
PPT Slide
Lager Image
Here, B ( x k ) is the linguistic term (fuzzy set) which means the activation level how much the input data x k is involved in the boundary area. Now the optimization problem is formulated in the following form
PPT Slide
Lager Image
The iterative optimization scheme is governed by the two update formulas using which we successively modify the partition matrix and the prototypes
PPT Slide
Lager Image
PPT Slide
Lager Image
- 3.3 pRBFNNs classifier- The use of conditional fuzzy C-Means clustering and a focus on the boundary area
In what follows, we propose the pFRBF NNs classifier developed by the c-FCM clustering supervised by the linguistic term, which specifies the boundary area.
As mentioned earlier, we assume that in order to improve the classification performance one has to locate the pFRBFs within the boundary area. pFRBF NNs is composed of the linear combination of the local models, which are defined on the local areas (receptive fields). In this way, the pFRBF NNs classifier can be regarded as a linear combination of the local boundary surfaces.
The local models of the pFRBF NNs are activated within the receptive fields (pFRBFs). Therefore, the pFRBFs located within the boundary area have the potential to form the “sound” boundary surface. Fig. 5 shows an overall development process of the proposed classifier.
PPT Slide
Lager Image
The overall development of the proposed classifier based on the extended pFRBF NNs, boundary area decision, and c-FCM
As shown in Fig. 2 , the output of the proposed pFRBF NNs classifier comes as the linear combination of the connection weights such as ( f 1 , f 2 , ⋯, fc ) with the activation levels of each node of the hidden layer ( Γ 1 , Γ 2 , ⋯, Γ c ). The way to calculate the output of the network of the proposed classifier is similar to the output of the extended pFRBF NNs. However, the activation levels of each pFRBF of the proposed model are described by using (21) which is quite different from the description provided by (2) and (6).
To estimate the connections we use the orthogonal least square method and the weighted least square estimation method. Proceeding with the optimization details, the objective function of Least Square Estimation (LSE) reads as follows
PPT Slide
Lager Image
where
PPT Slide
Lager Image
PPT Slide
Lager Image
The optimal values of the coefficients are expressed in a well-known manner
PPT Slide
Lager Image
When we use the weighted LSE to estimate the coefficients of local models, we assume that each data patterns comes with its priority and data patterns with high priority significantly affect the estimation process whereas data with low priority participate to a limited degree and can be almost neglected. The activation levels of the linguistic variable defining the boundary area can be considered as the priority index. As said earlier, we emphasize the data positioned within the boundary area.
Unlike the conventional LSE, the objective function of the weighted LSE is defined as follows
PPT Slide
Lager Image
Where,
PPT Slide
Lager Image
In the above expression, q denotes the linguistic modifier of the activation level of the boundary area. If the values of q get higher than 1, we arrive at higher specificity of the underlying linguistic information while an opposite effect becomes present when dealing with the lower values of q [3] . Note that the diagonal partition matrix D is the reduced matrix, which is composed of the activation levels of all data pairs to the linguistic term B as the diagonal elements.
The optimal values of the coefficients by using the weighted LSE are expressed in a well-known manner.
PPT Slide
Lager Image
The final output of the pFRBF NNs comes in the form
PPT Slide
Lager Image
The estimated class label is calculated by using the decision rule
PPT Slide
Lager Image
4. Experimental Study
In order to evaluate and quantify the classification effectiveness of the proposed classifier, the proposed classifier is experimented with by making use of a series of numeric data such as two synthetic datasets and several Machine Learning datasets ( http://www.ics.uci.edu/~mlearn/MLRepository.html ). In the assessment of the performance of the classifiers, we use the error rate of the resulting classifier.
We investigate and report the results of each experiment in terms of the mean and the standard deviation of the performance index. We consider some predefined values of the parameters of the network whose values are summarized in Table 1 . The choice of these particular numeric values has been motivated by the need to come up with a possibility to investigate of the performance of the model in a fairly comprehensive range of scenarios.
Selected Numeric Values of the Parameters of the Proposed Model
PPT Slide
Lager Image
Selected Numeric Values of the Parameters of the Proposed Model
In what follows, we report on several experiments dealing with some machine learning data sets (http://www. ics.uci.edu/~mlearn/MLRepository.html). For simplicity, we deal with two class- problems (the classifier can be extended to deal with more than two classes). The experiments were repeated 10 times using a random split of data into 70%-30% training and testing subsets. Table 2 contrasts the classification error of the proposed classifier with other well-known methods known in the literature [17] . In this experiments, the generic type basic neural networks (NNs), principal component analysis (PCA) and linear discriminant analysis (LDA) are used. Support vector machine (SVM) is available in a MATLAB toolbox, see http://theoval.sys.uea.ac.uk/~gcc/svm/toolbox/ . For the decision tree methods, the code of C4.5 trees was coming from the Classification Toolbox of MATLAB ( http://www.yom-tov.info/cgi-bin/list_uploaded_files.pl ) and the decision trees used some functions coming from the Statistics Toolbox of MATLAB.
Results of comparative analysis (The best results are shown in boldface)
PPT Slide
Lager Image
Results of comparative analysis (The best results are shown in boldface)
Table 3 shows the comparison between the proposed classifier and the classification methods based on the boundary analysis. In this experiments, we use 10 fold cross validation to evaluate the classification abilities and the final correct classification ratio is given in terms of its average and the standard deviation. From the results in Table VI, we can see that the proposed classifier is better than the LBDA based classifiers in terms of the classification abilities achieving higher classification rates.
Results of comparative analysis with the other classification methods based on the boundary analysis (The best results are shown in boldface)
PPT Slide
Lager Image
LBDA - linear boundary discriminant analysis. LBDA+NN(non) uses only non-boundary patterns to train Nearest Neighbor classifier, while LBDA+NN(all) uses all patterns to train the same classifier.
5. Conclusion
In this paper, we proposed the new design methodology of polynomial fuzzy radial basis function neural networks for the classification problem. Unlike the usual design method of RBFs, the proposed design method concentrate on a detailed description of the boundary regions in the feature space. The learning algorithm used to in the development of the conclusion part of the rules takes advantage of the linear discriminant analysis. To evaluate the proposed model for classification problem, we completed several experiments using 2-dimensional synthetic datasets and a number of machine learning datasets.
Acknowledgements
This work was supported by the GRRC program of Gyeonggi province [GRRC Suwon 2014-B2, Center for U-city Security & Surveillance Technology] and by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education, Science and Technology (NRF-2012R1A1B 3003568).
BIO
Seok-Beom Roh received the B.Sc., M.Sc., and Ph.D. degrees in control and instrumentation engineering from Wonkwang University, Korea, in 1994, 1996, and 2006 respectively. He is currently a Senior manager with Wia Corporation Co. His research interests include fuzzy set, neural networks, genetic algorithms, computational intelligence and statistical learning.
Sung-Kwun Oh received the B.Sc., M.Sc., and Ph.D. degrees in electrical engineering from Yonsei University, Seoul, Korea, in 1981, 1983, and 1993, respectively. During 1983-1989, he was a Senior Researcher of R&D Lab. of Lucky-Goldstar Industrial Systems Co., Ltd. From 1996 to 1997, he was a Postdoctoral Fellow with the Department of Electrical and Computer Engineering, University of Manitoba, Winnipeg, MB, Canada. He is currently a Professor with the Department of Electrical Engineering, University of Suwon, Suwon, South Korea. His research interests include fuzzy system, fuzzy-neural networks, automation systems, advanced computational intelligence, and intelligent control. He currently serves as an Associate Editor of the KIEE Transactions on Systems and Control, International Journal of Fuzzy Logic and Intelligent Systems of the KFIS, and Information Sciences.
References
Albrecht S. 2000 “Generalized radial basis function networks for classification and novelty detection: self-organization of optimal Bayesian decision” Neural Networks 13 1075 - 1093    DOI : 10.1016/S0893-6080(00)00060-5
Bishop C. M. 2006 Pattern Recognition and Machine Learning Springer Berlin
Bouchon-Meunier B. 1992 “Linguistic hedges and fuzzy logic” Fuzzy Systems, 1992., IEEE International Conference 247 - 254
Bugmann G. 1998 “Normalized Gaussian radial basis function networks” Neurocomputing 20 97 - 110    DOI : 10.1016/S0925-2312(98)00027-7
Duy N. M. , Cong T.T. 2001 “Numerical solution of differential equations using multiquadric radial basis function networks” Neural Networks 14 185 - 199    DOI : 10.1016/S0893-6080(00)00095-2
Er M.J. , Wu S.Q. , Lu J.W. , Toh H.L. 2002 “Face recognition with radical basis function (RBF) neural networks” IEEE Transactions on Neural Networks 13 (5) 697 - 710    DOI : 10.1109/TNN.2002.1000134
Hong X. , Chen S. , Harris C.J. 2007 “A kernel-based two-class classifier for imbalanced data sets” IEEE Transactions on Neural Networks 18 (1) 28 - 41    DOI : 10.1109/TNN.2006.882812
Jing X.Y. , Yao Y.F. , Zhang D. , Yang J.Y. , Li M. 2007 “Face and palm print pixel level fusion and Kernel DCV-RBF classifier for small sample biometric recognition” Pattern Recognition 40 3209 - 3224    DOI : 10.1016/j.patcog.2007.01.034
Karayiannis N.B. 1999 “Reformulated radial basis neural networks trained by gradient descent” IEEE Transactions on Neural Networks 10 (3) 657 - 671    DOI : 10.1109/72.761725
Krishapuram R. , Keller J.M. 1993 “A possibilistic approach to clustering” IEEE Transactions on Fuzzy Systems 1 (2) 98 - 110    DOI : 10.1109/91.227387
Loone S.M. , Irwin G. 2001 “Improving neural network training solutions using regularization” Neurocomputing 37 71 - 90    DOI : 10.1016/S0925-2312(00)00314-3
Looney C.G. 2002 “Radial basis functional link nets and fuzzy reasoning” Neurocomputing 48 489 - 509    DOI : 10.1016/S0925-2312(01)00613-0
Ma L. , Wahab A. , Ng G.S. , Erdogan S. 2008 “An experimental study of the extended NRBF regression model and its enhancement for classification problem” Neurocomputing 72 458 - 470    DOI : 10.1016/j.neucom.2007.12.011
Lazaro M. , Santamaria I. , Pantaleon C. 2003 “A new EM-based training algorithm for RBF networks” Neural Networks 16 (1) 69 - 77    DOI : 10.1016/S0893-6080(02)00215-0
Na Jin Hee , Park Myoung Soo , Choi Jin Young 2010 “Linear boundary discriminant analysis” Pattern Recognition 43 929 - 936    DOI : 10.1016/j.patcog.2009.09.015
Ng W.W.Y. , Dorado A. , Yeung D.S. , Pedrycz W. , Izquierdo E. 2007 “Image classification with the use of radial basis function neural networks and the minimization of the localized generalization error” Pattern Recognition 40 19 - 32    DOI : 10.1016/j.patcog.2006.07.002
Park B. J. , Pedrycz W. , Oh S.K. 2008 “Polynomialbased radial basis function neural networks (PRBFNNs) and their application to pattern classifycation” Applied Intelligence 32 (1) 27 - 46
Pedrycz W. 1998 “Conditional fuzzy clustering in the design of radial basis function neural networks” IEEE Transaction on Neural Networks 9 (4) 601 - 612    DOI : 10.1109/72.701174
Pedrycz W. , Park H.S. , Oh S.K. 2008 “A granularoriented development of functional radial basis function neural networks” Neurocomputing 72 420 - 435    DOI : 10.1016/j.neucom.2007.12.016
Powell M.J.D. , Mason J.C. , Cox M.G. 1987 Algorithms for Approximation Oxford University Press Oxford Radial basis functions for multivariable interpolation: a review 143 - 167
Senapati M.R. , Vijaya I. , Dash P.K. 2007 “Rule Extraction from Radial Basis Functional Neural Networks by Using Particle Swarm Optimization” Journal of Computer Science 3 (8) 592 - 599    DOI : 10.3844/jcssp.2007.592.599
Cervantes A. , Galvan I.M. , Isasi P. 2009 “AMSPO: A New Particle Swarm Method for Nearest Neighborhood Classification” IEEE Transaction on Systems, Man, and Cybernetics part B 39 (5) 1082 - 1091    DOI : 10.1109/TSMCB.2008.2011816
Bouchon-Meunier B. 1992 “Linguistic hedges and fuzzy logic” Fuzzy Systems, 1992., IEEE International Conference 247 - 254
Forghani Y. , Yazdi H. S. 2014 “Robust support vector machine-trained fuzzy system” Neural Networks 50 154 - 165    DOI : 10.1016/j.neunet.2013.11.013
Frigul Hichem , Bchir Quiem , Baili Naouel 2013 “An Overview of Unsupervised and Semi-Supervised Fuzzy Kernel Clustering” IJFIS 13 (4) 254 - 268