Multi-Choice Wavelet Thresholding Based Binary Classification Method

Seung Hyun Baek*a, Alberto Garcia-Diazb, Yuanshun Daic

Methodology, 2020, Vol. 16(2), 127–146, https://doi.org/10.5964/meth.2787

Received: 2018-08-29. Accepted: 2019-09-14. Published (VoR): 2020-06-18.

*Corresponding author at: Division of Business Administration, 55 Hanyangdaehak-ro, Sangnok-gu, Ansan Gyeonggi-do, 15588, South Korea. E-mail: sbaek4@hanyang.ac.kr

This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

Data mining is one of the most effective statistical methodologies to investigate a variety of problems in areas including pattern recognition, machine learning, bioinformatics, chemometrics, and statistics. In particular, statistically-sophisticated procedures that emphasize on reliability of results and computational efficiency are required for the analysis of high-dimensional data. Optimization principles can play a significant role in the rationalization and validation of specialized data mining procedures. This paper presents a novel methodology which is Multi-Choice Wavelet Thresholding (MCWT) based three-step methodology consists of three processes: perception (dimension reduction), decision (feature ranking), and cognition (model selection). In these steps three concepts known as wavelet thresholding, support vector machines for classification and information complexity are integrated to evaluate learning models. Three published data sets are used to illustrate the proposed methodology. Additionally, performance comparisons with recent and widely applied methods are shown.

Keywords: data mining, search procedures, optimization, classification analysis, multi-choice wavelet thresholding

Data mining procedures are essentially based on statistical principles and machine learning theory both creatively integrated to effect and facilitate the identification of significant informative patterns for a given database. Recurrent strategies used in data mining include preprocessing, data partitioning, machine learning (modeling), and validation. The ultimate goal of these procedures is the disclosure of unknown and valuable information. Adams et al. (2000) have discussed several models and patterns.

As indicated by Meisel and Mattfeld (2010), operations research and data mining are complementary and supportive due to three facts: (i) operations research techniques expedite efficiency of data mining; (ii) data mining methodologies enlarge the scope of operations research applications; and (iii) integration of both data mining and operations research boost systems performance. Furthermore, the key element that allows effective fusion of both areas is the use of optimization algorithms (with particular emphasis on search procedures) to find an accurate model and development of metaheuristics. An example of such procedures is the search algorithm by Olafsson et al. (2008) to find the best variable subset.

Classification analysis methods, based on several types of different algorithms, have been proposed to find successful models for complicated data in an extensive range of application domains. The objective of classification analysis is to identify groups of observations based on the input variables which minimize the within group-variability and maximize the between group-variability. Recently, not only classification area but also other supervised or unsupervised learning areas have faced two challenging issues: (i) the curse of dimensionality; and (ii) nonlinearity. Several researchers developed new classification analysis techniques for preventing problems of the curse of dimensionality; spectral regression discriminant analysis (Cai et al., 2008), automatic non-parameter uncorrelated discriminant analysis (Yang et al., 2008), high-dimensional discriminant analysis (Bouveyron et al., 2007), and for avoiding problems of nonlinearity; adaptive nonlinear discriminant analysis (Kim et al., 2006), kernel Fisher discriminant analysis (Mika, 2002), support vector machines for classification and regression (Vapnik, 1995).

Variable selection is an important area of research in machine learning, pattern recognition, statistics, and related fields. The key idea of variable selection is to find input variables which have predictive information and to eliminate non-informative variables. The use of variable selection techniques is motivated by three reasons: (i) to improve discriminant power; (ii) to find fast and cost-effective variables; and (iii) to reach a better understanding of the application process (Guyon & Elisseeff, 2003). In the case of high-dimensional data, variable selection plays a crucial role because of four challenges (Theodoridis & Koutroumbas, 2006): (i) large set of variables; (ii) existence of irrelevant variables; (iii) presence of redundant variables; and (iv) data noise.

This article proposes a novel methodology based on an integration of both the multi-choice wavelet thresholding (MCWT) and a variable selection method for classification to perform three steps known as perception, decision, and cognition. The proposed procedure will be referred to as a Perception-Decision-Cognition Methodology (PDCM). The main idea of this methodology is to provide better classification power by integrating both optimal search and data mining procedures. The perception step includes five different dimension reduction methods, based on wavelets, to transform original data into a representation form that exhibits orthogonality and de-noising. Multi-choice wavelet thresholding tries to say that the proposed method integrates these five different dimension reduction methods. But in the computational approach, five versions of the proposed procedure are compare and each version uses one of these dimension reduction methods. The decision step uses information complexity to find informative variables which can be used to identify groups based on prior modeling information. The cognition step recognizes the best model based on the support vector machines for classification, a well-known kernel-based statistical data mining approach. As the optimal way, new methodologies are usually tested to check capabilities of the procedures with simulated datasets which have different characteristics. But, the proposed PDCM is directly applied to three real datasets in this article. Three numerical experiments were run to compare the PDCM to other often-used procedures. The results from the experiments show that the proposed method outperforms all the procedures used in the experiments.

The section “The Proposed PDCM” reviews relevant procedures integrated to design the proposed methodology. These procedures can be classified into the following areas:

  1. Wavelet thresholding-based dimension reduction

  2. Variable selection (feature ranking)

  3. Cognition accuracy (model selection)

The performance of the methodology is tested using three published data sets and the corresponding results are documented in section “Experimental Results”. Section “Discussion” concludes this article. The information of three real benchmark datasets is presented in section “Supplementary Materials”.

The Proposed PDCM

Perception-Decision-Cognition Methodology (PDCM)

The proposed Perception-Decision-Cognition Methodology (PDCM) for discriminant analysis is conceptually represented in Figure 1. As indicated in this figure, it consists of three steps:

Click to enlarge
meth.2787-f1
Figure 1

Conceptual View of the Perception-Decision-Cognition Methodology (PDCM)

  1. Perceive environmental information.

  2. Decide on response (actions).

  3. Cognize the accuracy of results to adjust the response.

The algorithm used by the PDCM consists of three steps conceptually described below, after assuming that all data have been classified according to three sets: training set, cognition (validation) set, and test set.

Step 1: Perceive Sample Space and Data Dimensions

Let the sample data be X = ( x 1 , x 2 ,..., x q ) , and the corresponding response be y = ( y 1 , y 2 ,..., y n ) T , where q is the dimension of X and n is the number of samples. Now apply all available MCWT techniques: VisuShrinkUnion, VisuShrinkIntersect, VertiShrink, VET, and MSVET. For each dimension reduction technique, generate a new training set X = ( d 1 , d 2 ,..., d p ) , where the reduced dimension p is the number of coefficients perceived by the reduction techniques (pq).

Step 2: Decide on Variables Given Information Complexity

The procedure can be described as follows. Remove each of the p variables one at a time, and evaluate the corresponding information complexity measure, ICOMPPERF. Once the p - 1 removal procedures are completed, the removed variable resulting in minimum value of ICOMPPERF is identified and assigned the lowest rank (i.e. p). This procedure is repeated for the remaining p - 1 variables for which there is no rank yet. As a result of this, a variable receives rank equal to p - 1. This procedure is repeated until the p variables have been arranged according to their ranks.

Step 3: Cognize Accuracy of Selected Models

Compute the accuracy value of each cognition data set using the SVMC for all possible subsets of the ranked variables selected in Step 2. Specifically, first consider the variable with the highest rank (i.e, rank = 1), and calculate the cognition accuracy value. After this, the two variables with rank = 1 and rank = 2 are considered, and a new cognition accuracy value is calculated. This procedure is repeated until all ranked variables are considered. Finally, the subset of variables resulting in the highest accuracy value is chosen as the best model. The reason of using the two steps for transforming the nonlinear input data with wavelets and for finding informative variables with ICOMPPERF - RFE is to find more accurate and faster models.

Step 1: Wavelet Thresholding-Based Dimension Reduction Techniques

Dimension reduction is a preferred strategy in the area of machine learning. As anticipated, there are several approaches to perform dimensional reduction. The following methods are among the most popular: principal component analysis (Jolliffe, 2002), rotational linear discriminant analysis technique (Sharma & Paliwal, 2008), independent component analysis (Stone, 2004), semi-definite embedding (Weinberger & Saul, 2006), multifactor dimensionality reduction (Ritchie & Motsinger, 2005), factor analysis (Basilevsky, 1994), and wavelet-based dimension reduction (Chang & Vidakovic, 2002; Cho et al., 2009; Donoho & Johnstone, 1994; Jung et al., 2006).

The dimension reduction strategy has important benefits that can be measured not only in terms of computational time savings, but also in accuracy improvement. In the novel PDCM, the wavelet-based dimension reduction is applied in Step 1. The wavelets approach was selected because of several attractive attributes, among which the following two are most relevant: (a) wavelets adapt effectively to spatial features of a function such as discontinuities and varying frequency behavior; (b) wavelets have efficient O(n) algorithms to do transformations (Mallot, 1999). Based on perceived knowledge, wavelet-based techniques are applied to obtain a well-fitted reduced-dimension representation of the original data.

Discrete Wavelet Transformation (DWT) is often used for dimension reduction (also known as shrinkage or threshold). The data constructed with the scaling and wavelet functions based on orthogonal base in time domain is as follows:

f ( t ) = k c L , k ϕ L , k ( t ) + j L k d j , k ψ j , k ( t )
ϕ L , k ( t ) = 2 L / 2 ϕ ( 2 L t k ) ,       L , k Z
ψ j , k ( t ) = 2 j / 2 ψ ( 2 j t k ) ,       j L  and  j , k Z

where Z is the set of all the possible integer values, c L , k is the coarse level coefficient and d L , k is the finer level coefficient. Let y m = [ y m 1 , y m 2 , , y m N ] T is an m t h observed sample. For a single sample, the DWT procedure uses the orthonormal matrix W of dimension N × N to find the wavelet coefficient

d = ( c L , d L , d L + 1 , , d J )

where J > L, L corresponds to the lowest decomposition level.

c L = ( c L ,0 ,..., c L ,2 L 1 ) , d L = ( d L ,0 ,..., d L ,2 L 1 ) ,..., d J = ( d J ,0 ,..., d J ,2 J 1 ) ,

through the transformation

d = W y .

For multiple samples, let vector Y = [ y 1 , y 2 ,..., y M ] be the data set with M observed samples. The wavelet coefficient vector is obtained from the transformation

D = W Y

where D = [ d 1 , d 2 ,..., d M ] , and d m = ( c m L , d m L , d m L + 1 , , d m J ) , m = 1,2,..., M .

Small absolute values of wavelet coefficients are undesirable since they may be influenced more by noise than by information. On the other hand, large absolute values are more influenced by information than noise. This observation motivates the development of threshold methods. There are two threshold rules usually referred to as soft and hard thresholds. The soft rule is a continuous function of the data that shrinks each observation, while the hard rule retains unchanged only large observations (Donoho & Johnstone, 1994). The hard and soft threshold methods are defined as following:

D U , λ = s i g n U max 0, U λ     Soft D U , λ = U , all  U > λ 0, otherwise Hard

here λ is the threshold value. The threshold method can be used not only for data reduction but also for de-noising. One of the 5 different thresholding methods will be selected based on the performance of PDCM (Multi-Choice). Also, there are 5 different wavelet-based thresholding methods which will handle high-dimensional data. Therefore, the dimension reduction method is referred as Multi-Choice Wavelet Thresholding (MCWT).

VisuShrink (VS)

VisuShrink is a soft thresholding technique that applies a universal threshold proposed by Donoho and Johnstone (1994). The VisuShrink threshold is given by σ 2 log N , where N is the number of wavelet coefficients, and σ is the standard deviation of the wavelet coefficients (or noise standard deviation). When ε i is a white noise sequence, independent and identically distributed as N ( 0,1 ) , then as N ,   P { max | ε i | > 2 log N } 0 . That is, the maximum of the N values will most likely be smaller than the universal threshold. The VisuShrink guarantees a noise free reconstruction. However, when setting the threshold large, the degree of data fitting may be unsatisfactory. For multiple curves or samples, the VS procedure uses the union (VisuShrinkUnion, VSU) or intersection (VisuShrinkIntersection, VSI) of data sets in the selection of wavelet coefficients (Jung et al., 2006).

VertiShrink (VERTI)

Chang and Vidakovic (2002) developed a Stein-type shrinkage method, known as VertiShrink, to maximize the predictive density under appropriate model assumptions regarding wavelet coefficients. The main goal of VertShrink is the estimation of the baseline curve by using the average of block vertical coefficients. The estimated wavelet coefficients are given by:

θ ^ = 1 M σ 2 d T d + d

where, d is the wavelet coefficient, d = ( c L , d L , d L + 1 , , d J ) , M is the number of curves and σ is the standard deviation of the wavelet coefficients.

Vertical-Energy-Thresholding (VET)

VET was proposed by Jung et al. (2006). The procedure is based on the concept of energy of a function with some smoothness, since it is often concentrated on few coefficients, while the energy of noise is still spread over all coefficients in the wavelet domain. The vertical energy of wavelet coefficients is defined by

| | d v j | | 2 = d 1 j 2 + d 2 j 2 + ... + d M j 2

where d m j is the wavelet coefficient at the jth wavelet position for the mth data curve, m = 1, 2 ,..., M .

The VET method minimizes the overall relative reconstruction error (ORRE), formulated below, to determine a threshold value, namely λ:

O R R E ( λ ) = j = 1 N E | | d v j ( 1 I ( | | d v j | | 2 > λ ) ) | | 2 j = 1 N E | | d v j | | 2 + j = 1 N E | | I ( | | d v j | | 2 > λ ) | | 2 N

MultiScale-Vertical-Energy-Thresholding (MSVET)

Since the VET procedure does not consider the scale information of wavelets, an improved procedure proposed by Cho et al. (2009) and known as multi-scale vertical energy thresholding (MSVET) obtains a different optimal thresholding value for each scale by extending the idea of the VET procedure. In the MSVET procedure, the multi-scale overall relative reconstruction error (MSORRE) is defined as follows to determine the threshold values, λ i :

M S O R R E ( λ L , λ L + 1 ,..., λ J ) = i = L J j = 1 i + j 2 L E | | d v j i ( 1 I ( | | d v j i | | 2 > λ i ) ) | | 2 j = 1 N E | | d v j i | | 2 + i = L J j = 1 i + j 2 L E I ( | | d v j i | | 2 > λ i ) ) | | 2 N 2 J L

where, d v j i = ( d 1 j i , d 2 j i ,..., d M j i ) , | | d v j i | | 2 = d 1 j i 2 + d 2 j i 2 + ... + d M j i 2 ; d m j i represents the wavelet coefficient at the j t h wavelet position of the i t h scale for the mth curve, m = 1, 2 ,…, M.

Step 2: Variable Selection Based on Information Complexity and Recursive Feature Elimination

Once the reduced sample space is determined in Step 1, the decision regarding which of the remaining variables should be selected for ranking is made on the basis of minimal information complexity values, following the Information Complexity Performance Testing with Recursive Feature Elimination (ICOMPPERF-RFE) procedure proposed by Bozdogan and Baek (2018). Since this procedure resulted in better performance than other RFE-based methods. Also, this procedure essentially generates a stabilized and smoothed covariance estimator to calculate the information complexity measure, and, finally performs ranking using recursive elimination on the remaining variables.

The development of information complexity for the discriminant analysis is evaluated using the modified maximal entropic complexity C1F

C 1 F ( Σ ^ ) = 1 4 λ ¯ a 2 j = 1 s ( λ j λ ¯ a 2 ) ,

where s is the rank of Σ ^ , λ j is the jth eigenvalue of Σ ^ > 0,   j = 1, 2 , , s and λ ¯ a is arithmetic means of the eigenvalues

ICOMPPERF can be evaluated as indicated below:

I C O M P P E R F = n log 2 π + n log ( σ ^ 2 ) + n + 2 C 1 F ( Σ ^ S T A _ C S E )

where lack of fit is assessed by means of the first three terms and complexity by the fourth one. In the above expression, σ ^ 2 is the estimated mean squared error given by

σ ^ 2 = 1 n i = 1 n ( y i y ^ i ) 2 ,

and Σ ^ S T A _ C S E is the stabilized and smoothed convex sum covariance matrix estimator given by

Σ ^ S T A _ C S E = n n + k Σ ^ S T A + 1 n n + k t r a c e ( Σ ^ S T A ) h I h ,

where Σ ^ S T A is the stabilized covariance matrix proposed by Thomaz (2004), h is the number of variables, I h is h×h identity matrix, and k is chosen such that

0 < k < 2 [ h ( 1 + β ) 2 ] h β ,  

and

β = ( t r Σ ^ W ) 2 t r ( Σ ^ W 2 ) .

Specific details on this procedure are provided by Bozdogan and Baek (2018).

Step 3: Cognition Accuracy of Selected Models

When the ranking decision is finished in Step 2, the corresponding accuracies are determined using the corresponding cognition sets and the support vector machines for classification (SVMC) described below. Once the accuracies are calculated for the selected models the most-accurate one is chosen.

The SVMC find an optimal separating hyperplane that maximizes the margin between the classes (Vapnik, 1995). Consider the case of classifying a set of linearly separating data into two groups. Assume a set of training data is given by [ ( x 1 , y 1 ) , ( x 2 , y 2 ) ,..., ( x n , y n ) ] where x i p is an input vector, y i { -1,1 } is a binary class index, and n is the size of the training data set. Then, a decision boundary that partitions the underlying vector space into two classes can be represented by the hyperplane

w T x + b = 0

where w is the weight vector and b is the bias. The objective of SVMC is to find a maximum margin decision boundary between two parallel hyperplanes, w T x + b = 1 and w T x + b = 1 . The dual model of the corresponding primal model can be formulated as

max l ( α ) = i = 1 n α i 1 2 i = 1 n j = 1 n α i α j y i y j K ( x i T , x j )

subject to

i = 1 n α i x i = 0 , 0 α i C , i = 1,2,..., n

where K ( x i T , x j ) is the kernel function and C is a predefined coefficient. Kernel functions used in the numerical experiments are described in Table 1.

Table 1

Kernel Functions

Function K ( x i T , x j ) Parameters
Gaussian exp 1 a b | | x i x j | | 2 c a = 2, b = c = 1
Cauchy 1 + 1 a | | x i x j | | 2 1 a = 1
Inverse Multi-Quadratic | | x i x j | | 2 + a 2 1 / 2 a = 1

The point x o with coordinates corresponding to new data can be classified as indicated below:

Class 1:  i = 1 n α i o v y i K ( x i T , x o ) + b o v < 0

and

Class 2:  i = 1 n α i o v y i K ( x i T , x o ) + b o v > 0

where α o v and b o v are optimal values found based on the training data. A classification example based on the PDCM is illustrated in Figure 2.

Click to enlarge
meth.2787-f2
Figure 2

Classification Example Based on PDCM with SVMC

Experimental Results

In order to emphasize the effectiveness of the PDCM, it will be applied to three different data sets (Heart, Fat, & Handwritten) used for experiments with low-dimension high sample size, and high-dimension low sample size. The data sets are divided as 3 sets; training (50% of total set), cognition (10% of total set), and test (40% of total set) sets. For the wavelet transformation of the three data sets, the linear padding suggested by (Strang & Nguyen, 1997) is applied. Also, well-known MATLAB software was used to get the experimental results. MATLAB software gives several advantages such as easy calculation for data matrix, vector operation, easy plotting, and function operations with MATLAB toolbox. This article documents the comparison of the PDCM to the following procedures:

  1. SVMC recursive feature elimination (SVMC-RFE) (Guyon et al., 2002): The SVMC-RFE is used weight as a criterion to rank each feature using support vector machines for classification and recursive feature elimination algorithm.

  2. Two-stage method (Cho et al., 2009): The Two-stage method is used multi-scale vertical energy thresholding (MSVET) to reduced dimension and applied support vector machines for classification and recursive feature elimination to select important wavelet coefficients based on gradient.

Heart Data (44 Variables)

The data set includes 267 samples and 44 variables on cardiac single proton emission computed tomography (SPECT) images with two categories, i.e., normal and abnormal. There are 55 normal and 212 abnormal classes (Cios & Kurgan, 2001). The data set is divided into 134 samples as a training set, 13 samples as a cognition set, and 120 samples as a test set. The training set has 25 normal and 109 abnormal classes. The cognition set has 1 normal and 12 abnormal classes. The test set has 29 normal and 91 abnormal classes1. Table 2 and Table 3 show comparison results in terms of the variables selected from ICOMPPERF-RFE, the cognition accuracy, and the test accuracy. Cauchy and inverse multi-quadratic kernel functions are used in Table 2 and Table 3, respectively.

Table 2

PDCM Versus Various Ranking Based Method Using Cauchy Kernel

Method Selected Variables from ICOMPPERF-RFE Number of Variables Cognition Accuracy Test Accuracy
PDCM (MSVET) 20, 28, 30 3 100% 92%
PDCM (VET) 20, 28, 50 3 100% 92%
PDCM (VERTI) 19, 34 2 100% 92%
PDCM (VISU UNION) 9, 20, 44, 53 4 100% 92%
PDCM (VISU INTERSECT) 8, 15 2 100% 92%
SVMC-RFE 6, 16, 17, 18, 26, 32, 35 7 92% 76%
Two-Stage 43 1 92% 78%
Table 3

PDCM Versus Various Ranking Based Method Using Inverse Multi-Quadratic Kernel

Method Selected Variables from ICOMPPERF-RFE Number of Variables Cognition Accuracy Test Accuracy
PDCM (MSVET) 6, 10, 18, 29 4 92% 76%
PDCM (VET) 6, 10, 18, 29 4 92% 76%
PDCM (VERTI) 6, 10, 18, 29 4 92% 76%
PDCM (VISU UNION) 9, 10, 16, 29 4 92% 76%
PDCM (VISU INTERSECT) 10, 19, 28 3 92% 78%
SVMC-RFE 22 1 92% 76%
Two-Stage 43 1 92% 78%

As observed in Table 2, PDCM achieves better cognition accuracies comparing to SVMC-RFE and two-stage and yields more accurate results in test. Also, as shown in Table 3, PDCM (visu intersect) and two-stage both reach the highest test accuracy, although two-stage requires fewer variables.

Near Infrared Spectroscopy Data (100 Variables)

These data were collected by a Tecator infratec food and feed analyzer to predict the fat content of a meat sample based on near infrared (NIR) spectroscopy. The data set was divided into two classes defined on the basis of fat content; one class (low-fat) corresponded to 20% or less, and another class (high-fat) to more than this level (Rossi & Villa, 2006). There are 77 low-fat and 138 high-fat classes. The entire data set consists of 215 samples with measured values for each of 100 predictive variables (wavelengths). These samples are divided randomly to configure a training set consisting of 108 samples; a cognition set consisting of 11 samples for cognition; and a test set consisting of the remaining 96 samples. The training set has 39 low-fat and 69 high-fat classes. The cognition set has 4 low-fat and 7 high-fat classes. The test set has 34 low-fat and 62 high-fat classes2. Table 4 and Table 5 show comparison results in terms of the variables selected from ICOMPPERF-RFE, the cognition accuracy, and the test accuracy. Cauchy and Gaussian kernel functions are used in Table 4 and Table 5, respectively. Although both PDCM and two-stage reach the 91% accuracy level for the cognition set in Table 4, the test accuracy of PDCM is higher than that of SVMC-RFE and two-stage except PDCM (verti). Additionally, Table 5 shows that the cognition accuracy of PDCM is 91% and the test accuracy is higher than that of other methods. Furthermore, PDCM uses only less than 9 variables to reach the accuracy levels previously mentioned.

Table 4

PDCM Versus Various Ranking Based Method Using Cauchy Kernel

Method Selected Variables from ICOMPPERF-RFE Number of Variables Cognition Accuracy Test Accuracy
PDCM (MSVET) 2, 5 2 91% 90%
PDCM (VET) 5, 32 2 91% 90%
PDCM (VERTI) 6, 32 2 91% 84%
PDCM (VISU UNION) 1, 5 2 91% 90%
PDCM (VISU INTERSECT) 4, 28 2 91% 91%
SVMC-RFE 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100 96 82% 84%
Two-Stage 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45 13 82% 86%
Table 5

PDCM Versus Various Ranking Based Method Using Gaussian Kernel

Method Selected Variables from ICOMPPERF-RFE Number of Variables Cognition Accuracy Test Accuracy
PDCM (MSVET) 4, 5, 14, 17, 18, 29 6 91% 92%
PDCM (VET) 4, 5, 10, 16, 22, 25, 32 7 91% 91%
PDCM (VERTI) 1, 5, 8, 12, 14, 20, 22, 25, 27 9 91% 86%
PDCM (VISU UNION) 1, 5, 6, 13, 17, 19, 24, 25, 30 9 91% 91%
PDCM (VISU INTERSECT) 5, 9, 10, 16, 19, 28 6 91% 88%
SVMC-RFE 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100 92 73% 85%
Two-Stage 40, 43, 45 3 73% 80%

Handwritten Data (240 Variables)

This data set has features of handwritten numerals from 0 to 9 extracted from a collection of Dutch utility maps. The entire set consists of 200 samples digitized in binary images per numerals and six different variable sets: 76 fourier coefficients, 216 profile correlations, 64 Karhunen-Love coefficients, 240 pixel averages, 47 Zernike moments and 6 morphological features (Van Breukelen et al., 1998). Two of six variable sets (216 profile correlations and 240 pixel averages) are used for the experiment. Two (0 and 9) out of 10 numerals are selected to verify the proposed method. Each numeral has 100 samples which are included in the experimental data set. For the experiment, 100 samples are used as a training set, 10 samples are used as a cognition set and 90 samples are used as a test set. The training set has 45 zero-numeral and 55 nine-numeral classes. The cognition set has 4 zero-numeral and 6 nine-numeral classes. The test set has 51 zero-numeral and 39 nine-numeral classes3. Table 6 and Table 7 show comparison results in terms of the selected variables from ICOMPPERF-RFE, the cognition accuracy, and the test accuracy.

Table 6

PDCM Versus Various Ranking Based Method Using Cauchy Kernel

Method (Wavelet-Based Method) Selected Variables from ICOMPPERF-RFE Number of Variables Cognition Accuracy Test Accuracy
PDCM (MSVET) 112 1 100% 88%
PDCM (VET) 25 1 100% 93%
PDCM (VERTI) 25 1 100% 93%
PDCM (VISU UNION) 25 1 100% 93%
PDCM (VISU INTERSECT) 25 1 100% 93%
SVMC-RFE 324, 334, 339, 340, 344, 349, 456 7 100% 89%
Two-Stage 8, 10, 11, 15 4 80% 80%
Table 7

PDCM Versus Various Ranking Based Method Using Inverse Multi-Quadratic Kernel

Method Selected Variables from ICOMPPERF-RFE Number of Variables Cognition Accuracy Test Accuracy
PDCM (MSVET) 112, 155 2 100% 97%
PDCM (VET) 25 1 100% 100%
PDCM (VERTI) 25 1 100% 100%
PDCM (VISU UNION) 87, 345 2 100% 98%
PDCM (VISU INTERSECT) 24 1 100% 100%
SVMC-RFE 48, 82, 106, 154, 166, 217, 218, 223, 224, 225, 226, 230, 231, 232, 235, 236, 237, 238, 239, 240, 241, 246, 249, 250, 251, 257, 258, 261, 262, 264, 265, 273, 276, 277, 279, 280, 283, 284, 288, 289, 291, 294, 295, 298, 299, 300, 303, 304, 309, 310, 312, 318, 319, 325, 333, 334, 335, 337, 342, 348, 349, 350, 352, 358, 363, 364, 365, 367, 373, 374, 376, 378, 379, 380, 382, 387, 389, 391, 393, 394, 395, 397, 402, 404, 408, 410, 412, 422, 423, 424, 427, 428, 435, 436, 437, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 454, 455, 456 109 100% 92%
Two-Stage 8, 10, 11, 15 4 70% 80%

Cauchy and inverse multi-quadratic kernel functions are used in Table 6 and Table 7, respectively. As shown in the tables, PDCM reaches 100% cognition accuracy levels as did the SVMC-RFE, except two-stage. Furthermore, PDCM achieves a higher accuracy level than the other methods for the test set, except PDCM (msvet) in Cauchy kernel.

Discussion

This article documents the development and application of a novel Perception-Decision-Cognition Methodology (PDCM) for classification analysis based on the MCWT and SVMC with ICOMPPERF-RFE. Five different wavelet-based dimension reduction techniques called MCWT are applied in the perception step. It is shown that the procedure yields a good representation of the original data, using only reduced variables. The decision step is performed using a rank-based variable selection approach, using the information complexity criterion. The information complexity based variable selection approach shows a good ability to achieve reasonable variable ranks, which in turn can affect decision making. In the cognition step, the number of variables and accuracy are cognized for further discrimination.

The PDCM is directly applied to three real datasets instead of using simulated datasets having different characteristics in this article. As supported by the numerical experiments documented in this article, the PDCM outperforms the currently available data mining approaches, and, furthermore, shows to be applicable to various areas, such as bioinformatics, chemometrics, pattern recognition, and other data mining fields. The PDCM has three advantages:

  1. Dimension simplification.

  2. Multiple model choices based on simplified dimension.

  3. Novel rank based variable selection: ICOMPPERF-RFE.

Funding

The authors have no funding to report.

Competing Interests

The authors have declared that no competing interests exist.

Acknowledgments

The authors have no support to report.

Author note

This paper was modified from the Ph.D. dissertation (Baek, 2010) of the first author.

Data Availability

Datasets used for this article are freely available (see Supplementary Materials section).

Supplementary Materials

  1. Dataset on cardiac Single Proton Emission Computed Tomography (SPECT) images.

  2. Dataset on near infrared absorbance spectrum of fat content in meats.

  3. Dataset on features of handwritten numerals (`0'--`9') extracted from a collection of Dutch utility maps.

Index of Supplementary Materials

References

  • Adams, N., Blunt, G., Hand, D., & Kelly, M. (2000). Data mining for fun and profit. Statistical Science, 15(2), 111-131. https://doi.org/10.1214/ss/1009212753

  • Baek, S. H. (2010). Kernel-based data mining approach with variable selection for nonlinear high-dimensional data [Unpublished doctoral dissertation]. University of Tennessee, Knoxville, USA.

  • Basilevsky, A. (1994). Statistical factor analysis and related methods: Theory and application. New York, NY, USA: Wiley.

  • Bouveyron, C., Girard, S., & Schmid, C. (2007). High-dimensional discriminant analysis. Communications in Statistics. Theory and Methods, 36, 2607-2623. https://doi.org/10.1080/03610920701271095

  • Bozdogan, H., & Baek, S. H. (2018). Hybridized support vector machine and recursive feature elimination with information complexity. Statistics, Optimization & Information Computing, 6(2), 159-177. https://doi.org/10.19139/soic.v6i2.327

  • Cai, D., He, X., & Han, J. (2008). An efficient algorithm for large-scale discriminant analysis. IEEE Transactions on Knowledge and Data Engineering, 20(1), 1-12. https://doi.org/10.1109/TKDE.2007.190669

  • Chang, W., & Vidakovic, B. (2002). Wavelet estimation a baseline signal from repeated noisy measurements by vertical block shrinkage. Computational Statistics & Data Analysis, 40, 317-328. https://doi.org/10.1016/S0167-9473(02)00053-1

  • Cho, H., Baek, S. H., Youn, E., Jeong, M. K., & Taylor, A. (2009). A two-stage classification procedure for near-infrared spectra based on multi-scale vertical energy wavelet thresholding and SVM-based gradient-recursive feature elimination. The Journal of the Operational Research Society, 60, 1107-1115. https://doi.org/10.1057/jors.2008.179

  • Cios, K. J., & Kurgan, L. (2001). Hybrid inductive machine learning: An overview of CLIP algorithms. In L. C. Jain & J. Kacprzyk (Eds.), New learning paradigms in soft computing (pp. 276-321). Heidelberg, Germany: Physica-Verlag.

  • Donoho, D. L., & Johnstone, I. M. (1994). Ideal spatial adaptation by wavelet shrinkage. Biometrika, 81, 425-455. https://doi.org/10.1093/biomet/81.3.425

  • Guyon, I., & Elisseeff, A. (2003). An introduction to variable and feature selection. Journal of Machine Learning Research, 3(1), 1157-1182.

  • Guyon, I., Weston, J., Barnhill, S., & Vapnik, V. (2002). Gene selection for cancer classification using support vector machines. Machine Learning, 46(1/3), 389-422. https://doi.org/10.1023/A:1012487302797

  • Jolliffe, I. T. (2002). Principal component analysis. New York, NY, USA: Springer.

  • Jung, U., Jeong, M. K., & Lu, J. C. (2006). A vertical-energy-thresholding procedure for data reduction with multiple complex curves. IEEE Transactions on Cybernetics, 36(5), 1128-1138. https://doi.org/10.1109/TSMCB.2006.874681

  • Kim, H., Drake, B. L., & Park, H. (2006). Adaptive nonlinear discriminant analysis by regularized minimum squared errors. IEEE Transactions on Knowledge and Data Engineering, 18(5), 603-612. https://doi.org/10.1109/TKDE.2006.72

  • Mallot, S. (1999). A wavelet tour of signal processing. San Diego, CA, USA: Academic Press.

  • Meisel, S., & Mattfeld, D. C. (2010). Synergies of operations research and data mining. European Journal of Operational Research, 206(1), 1-10. https://doi.org/10.1016/j.ejor.2009.10.017

  • Mika, S. (2002). Kernel fisher discriminants [Unpublished doctoral dissertation]. Technical University of Berlin, Berlin, Germany.

  • Olafsson, S., Li, X., & Wu, S. (2008). Operations research and data mining. European Journal of Operational Research, 187(3), 1429-1448. https://doi.org/10.1016/j.ejor.2006.09.023

  • Ritchie, M. D., & Motsinger, A. A. (2005). Multifactor dimensionality reduction for detecting gene-gene and gene-environment interactions in pharmacogenomics studies. Pharmacogenomics, 6(8), 823-834. https://doi.org/10.2217/14622416.6.8.823

  • Rossi, F., & Villa, N. (2006). Support vector machine for functional data classification. Neurocomputing, 69(7-9), 730-742. https://doi.org/10.1016/j.neucom.2005.12.010

  • Sharma, A., & Paliwal, K. K. (2008). Rotational linear discriminant analysis. IEEE Transactions on Knowledge and Data Engineering, 20(10), 1336-1347. https://doi.org/10.1109/TKDE.2008.101

  • Stone, J. (2004). Independent component analysis: A tutorial introduction. London, United Kingdom: MIT press.

  • Strang, G., & Nguyen, T. (1997). Wavelets and filter banks. Wellesley, MA, USA: Wellesley-Cambridge Press.

  • Theodoridis, S., & Koutroumbas, K. (2006). Pattern recognition. San Diego, CA, USA: Academic Press.

  • Thomaz, C. (2004). Maximum entropy covariance estimate for statistical pattern recognition [Unpublished doctoral dissertation]. University of London, London, United Kingdom.

  • Van Breukelen, M., Duin, R. P. W., Tax, D. M. J., & Den Hartog, J. E. (1998). Handwritten digit recognition by combined classifiers. Kybernetika, 34(4), 381-386.

  • Vapnik, V. (1995). The nature of statistical learning theory. New York, NY, USA: Springer-Verlag.

  • Weinberger, K. Q., & Saul, L. K. (2006). Unsupervised learning of image manifolds by semidefinite programming. International Journal of Computer Vision, 70(1), 77-90. https://doi.org/10.1007/s11263-005-4939-z

  • Yang, W. H., Dai, D. Q., & Yan, H. (2008). Feature extraction and uncorrelated discriminant analysis for high-dimensional data. IEEE Transactions on Knowledge and Data Engineering, 20(5), 601-614. https://doi.org/10.1109/TKDE.2007.190720