Kernel Based Nonlinear Weighted Least Squares Regression

Embed Size (px)

Citation preview

  • 8/3/2019 Kernel Based Nonlinear Weighted Least Squares Regression

    1/8

    Kernel Based Nonlinear Weighted LeastSquares RegressionAntoni Wibowo and Mohamad Ishak Desa

    AbstractIn this paper, we consider a regression model with heteroscedastic errors in which the prediction of ordinary least

    squares (OLS) based regressions can be inappropriate. Weighted least squares (WLS) is widely used to handle in theheteroscedastic model by transforming an original model into a new model that satisfies homoscedasticity assumption. However,WLS yields a linear prediction and has no guarantee to avoid the negative effect of multicollinearity. Under this circumstance, wepropose a method to overcome these difficulties using the hybridization of WLS regression with kernel method. We use WLS tohandle the heteroscedastic errors and use kernel method to perform a nonlinear model and to eliminate the negative effect ofmulticollinearity in this regression model. Then, we compare the performance of the proposed method with the WLS regressionand it gives better results than WLS regression.

    IndexTerms Kernel principal component regression, kernel trick, multicollinearity, heteroscedastic, nonlinear regression,weighted least squares.

    .

    1 INTRODUCTION

    etus consider aheteroscedastic regressionmodel as

    follows:

    , (1) , where

    , ,

    , , , ,and , , , with and arepositivereal numbers and 1 , 2 , , . The weight wi isestimatedusingthedataand,seeforexample[2],[6]and[7].Thesizesof,,,, andarep1,N1,Np,N(p+1),(p+1)1andN1,respectively,whereisaN1vectorwithallelementsequaltooneandandisthesetofrealnumbers.Thevectordenotesthetransposeofthevector .Animplicationoftheassumption

    isthat

    the ordinary least squares (OLS) estimator

    canbe inappropriate and it isnot thebestlinearunbiasedestimator (BLUE)of, seeforexample[1]

    and [8].That is, among all theunbiasedestimators,OLS

    doesnotprovide theestimatewith thesmallestvariance.

    Depending on the nature of the heteroscedasticity,

    significancetestscanbetoohighortoolowwhichimplies

    thesignificance testhas lowpower.Thesedifficultiesare

    usually handled by transforming model 1 into a new

    model that satisfies homoscedasticity assumption and

    calleditweightedleastsquares(WLS)method.Thismethod

    also yields a linearpredictionmodel,however, theWLS

    solutionispreferredtotheOLSsolution[1].

    Furthermore,we

    say

    that

    multicollinearity exists

    on

    regressorsmatrixX if ) isanearlysingularmatrix,i.e., if some eigenvalues of are close to zero. Ifmulticollinearity existsonX then canbe a largenumber and under the assumption that i is normally

    distributed,thetestsforinferencesj(j=0,1,...,p)have

    lowpowerand the confidence intervalcanbe large [15].

    Therefore, it will be difficult to decide if a variable xj

    makes a significant contribution to the regression.These

    implicationsareknownastheeffectofmulticollinearity.

    We can use principal component regression (PCR) to

    eliminate the effects ofmulticollinearity.However, PCR

    yields linearpredictionmodelswhichhave limitation in

    applications sincemost real problems are nonlinear. In

    recentyears,[3],[4],[9],[10],[11]and[16]haveproposed

    kernelprincipal component regression (KPCR) to overcome

    this linearity and to avoid the negative effect of

    multicollinearity.KPCRwas constructedbased on kernel

    principal component analysis (KPCA) andhomoscedasticity

    assumption, see [12] and [13] for the detailed ofKPCA,

    whichwasperformedbymappinganoriginalinputspace

    intoahigherdimensionalfeaturespace.Therefore,KPCR

    AntoniWibowoiswiththeFacultyofComputerScienceandInformationSystems,UniversitiTeknologiMalaysia,81310UTMJohorBahru,Johor,Malaysia.

    Mohammad Ishak Desa is with the Faculty of Computer Science andInformation Systems, Universiti Teknologi Malaysia, 81310 UTM JohorBahru, Johor, Malaysia.

    L

    JOURNAL OF COMPUTING, VOLUME 3, ISSUE 11, NOVEMBER 2011, ISSN 2151-9617

    HTTPS://SITES.GOOGLE.COM/SITE/JOURNALOFCOMPUTING

    WW.JOURNALOFCOMPUTING.ORG 56

  • 8/3/2019 Kernel Based Nonlinear Weighted Least Squares Regression

    2/8

    can be inappropriate to be used in regression analysis

    whenhomoscedasticityassumptionisnotsatisfied.

    Under the circumstance, we propose a nonlinear

    methodwhichcaneliminate theeffectofmulticolinearity

    and compromises the heteroscedasticity assumption. In

    thismethod, an original data set is transformed into a

    higherdimensionalfeaturespaceandtocreateamultiple

    linearregression

    with

    heteroscedasticity

    assumption

    in

    this space.Then,weperformWLSmethodfor this linear

    regression to obtain a multiple linear regression with

    homoscedasticityassumption.Furthermore,weuseatrick

    to have an explicitly homoscedastic multiple linear

    regression andperform the similarprocedure ofPCR in

    this featurespace toobtainanonlinearpredictionand to

    eliminate the effect of multicollineary. We refer the

    proposed technique as the weighted leastsquares KPCR

    (WLSKPCR).

    The rest of manuscript are organized as follows:

    Section 2, we present theories and methods of WLS

    regression,

    kernel

    trick,

    WLS

    KPCR

    and

    WLS

    KPCRs

    algorithm.InSection3,wecomparetheperformanceWLS

    regression andWLSKPCR.Then,wegiveconclusionsin

    Section4.

    2 THEORIES AND METHODS

    2.1 Weighted Least Squares Regression

    WLS method in the linear regression is performed as

    follows. First, we define , , , whichimpliesthat

    , and , , ).Itisevidentthat

    (2)Let , and . It is easy toverify that and and model 1becomes

    , (3) ,

    .We can see that the error in model 3 satisfieshomoscedasticityassumption.Toobtaintheestimatorof

    inmodel3wesolve

    min (4)

    with respect to.Let be the solution to theproblem4whichsatisfiestheleastsquaresnormalequations

    . (5)It is evident that if the row vectors of are linearlyindependent, then the rowvectorsof

    arealso linearly

    independent.Hence,

    isinvertibleandweobtain . (6)

    andcalledittheWLSestimatorof.Let be the observed data

    correspondingtoand bethevalueof wheninEq.(6)isreplacedby.Thefittedvalueof, say,isgivenby

    (7)

    andthepredictionofWLSbasedlinearregressionisgivenby

    , (8)where isa function from into .We shouldnoticethat the elements ofmatrix inmodel 1 canbe chosensuch that multicollinearity is not present in .Unfortunately, eigenvalues of are not equal toeigenvaluesof which implies there isnoguaranteethatmulticollinearitydoesnotexist in.Thepresentofmulticollinearity in can seriously deteriorate thepredictionofWLS regression.

    2.2 Heteroscedastic Regression Model in FeatureSpace

    The procedure for constructing WLS KPCR is almost

    similar toWLS regression that previously explained in

    Subsection 2.1. First,we transform our data set into an

    Euclidean space of higher dimensionby a function.The

    importantpointhere is that the function isnotexplicitly

    defined. Then, we construct a heteroscedastic linear

    regressionmodelinthisspaceanduseWLSsprocedureto

    obtain a homoscedastic linear regression model, and

    followed by performing a trick to obtain the explicit

    homoscedasticregressionandusethesimilarprocedureof

    PCR to eliminate the effect of multicollineary in this

    regressionmodel.

    Assumingwe have a function : where iscalled the feature space which is an Euclidean space of

    higherdimensionthanp,say.Then,wedefine ,

    JOURNAL OF COMPUTING, VOLUME 3, ISSUE 11, NOVEMBER 2011, ISSN 2151-9617

    HTTPS://SITES.GOOGLE.COM/SITE/JOURNALOFCOMPUTING

    WW.JOURNALOFCOMPUTING.ORG 57

  • 8/3/2019 Kernel Based Nonlinear Weighted Least Squares Regression

    3/8

    1 and

    wheresizesof

    , and

    areNpF,pFpFandNN,

    respectively.We

    assume

    that

    . If isinfinitedimensional, we consider the linear operator insteadofthematrix[13].Theheteroscedasticityregressionmodelinthefeaturespace

    isgivenby

    , (9) , ,where is a vector of regressioncoefficients in the feature space, isavectorofrandomerrors,

    )

    and

    ,

    , ,

    whereandarepositiverealnumbersfor i=1,2,...,N.Theweight isestimatedusing thedata ) and . We notice that we cannot use thegeneralized inversematrix to obtain the estimator of since isnotknownexplicitly.Then, we define , , , which

    implies

    , and , , )andobtain

    , (10) , ,where , and .Furthermore,wedefine twomatrices and . The relation of eigenvalues andeigenvectors of thematrices and are related in thefollowingtheorem.

    Theorem 1. [16] Suppose 0 and }.Thefollowingstatementsareequivalent:

    1. andsatisfy .2. andsatisfy and ),forsome }.

    3. and satisfy and ),forsome }.

    Let betherankofwhere min(N,).Sincetherankof isequal to therankof and the rankof ,then therankof and therankof areequal to .

    We should notice that is symmetric and positivesemidefinite which implies the eigenvalues of arenonnegativerealnumbers.Let 1 2 1... r r

    1F Fp p 0

    N betheeigenvaluesof, = ( . . . be the matrix of the corresponding

    normalizedeigenvectors

    1, 2 , . . . , of

    ,

    and forl=1,2,... .Then, according to Theorem 1we obtain eigenvaluesandeigenvectorsproblemasfollows

    1,2, . . . , 1 , , 1,2, , 0 . .

    Since therankofT isequal , then theremaining( )eigenvaluesofT arezero.Let k (k= 1,

    2 ,...,

    ),bethezeroeigenvaluesof T and

    bethe

    normalizedeigenvectors

    of

    T

    corresponding

    to

    k .Then, we define which is anorthogonalmatrix. Itisnotdifficulttoverifythat

    ,where ,

    0

    0

    .andOisazeromatrix.Since ,wecanrewritethemodel(10)as

    , (11) , ,where and .Let

    and ,with sizes of

    ,

    ,

    , and

    are

    , , 1 and 1,respectively.Themodel(11)canbewrittenas , (12) , .

    Since ,weobtain

    JOURNAL OF COMPUTING, VOLUME 3, ISSUE 11, NOVEMBER 2011, ISSN 2151-9617

    HTTPS://SITES.GOOGLE.COM/SITE/JOURNALOFCOMPUTING

    WW.JOURNALOFCOMPUTING.ORG 58

  • 8/3/2019 Kernel Based Nonlinear Weighted Least Squares Regression

    4/8

    , ,and .Since is equal tozero, we see that

    is equal to 0.

    Consequently,the

    model

    (12)

    is

    simplified

    to

    , (13) , .Letus assume that

    1 2, , ..., Fr r p are close to zero

    anddefine

    , and

    ,with

    0 0 ,

    0 0 ,andsizes of, , , , , 1and( r)1,respectively.Themodel(13)cannowbewrittenas

    (14) , .It is evident that the estimator of , say . . . , isgivenby

    (15)andthevarianceof 1 , . . ., is

    (16)Since , , . . . , are close to zero, the diagonalelementsof andalso thevarianceof 1 , . . . , will be very large numbers. Thus, we

    encounter the ill effect ofmulticollinearity in themodel

    (14).To avoid the effects ofmulticollinearitywedispose

    theterm asin[15]andobtain (17)

    where is a random vector influenced by dropping inthemodel(17).Weusuallydisposetheterm to handle the effects ofmulticollinearityonthePCR.Wecanusetheratio

    1

    l (l

    =1,2,, )todetectthepresenceofmulticollinearityon. If 1l is smaller than, say

  • 8/3/2019 Kernel Based Nonlinear Weighted Least Squares Regression

    5/8

    then there exists such that , .The function iscalled thekernel function.Insteadofchoosingexplicitly,wechooseakernelandemploythecorrespondingfunctionas.Let , .Hence,weobtain

    It isevident that isknownexplicitlynow.This impliesthat= ,andarealsoknownexplicitly.2.4 WLS KPCR

    Now, let us consider model 17 again. Since we have

    known that = , and explicitly,we canobtain theregressioncoefficientsof thismodel.Then, the

    fittedvalueof inthetransformedscale,say,isgivenby

    (21) andresidualbetweenandisgivenby

    . (22)Thefittedvalueofintheoriginalscale,say,isgivenby

    (23)and

    the

    theWLSKPCRspredictionis

    given

    by

    , , (24)where 1 , is a function from into and( . . . . The number is called theretainednumberofnonlinearPCsfortheWLSKPCR.

    2.5 WLS KPCR Algorithms

    We summarize the procedure in subsection 2.22.4 to

    obtaintheWLSKPCRspredictionasfollows:1. Given 1 2( , , ,..., ),i i i ip y x x x ,i=1,2,...,N.

    2. Calculate (1/ )1TN y N y

    and ( (1/ ) )To N N N I N y 1 1 y .

    3. Estimate V andfind L .

    4. Calculate1

    o o

    z = L y .

    5. Chooseakernel .

    6. Contruct ( , ), ( )ij i j ijK K x x K

    and

    1 1 K = L KL .

    7. Diagonalize K . Let 1 2 ... r

    1... ... 0 p F p F N be the

    eigenvalues of K and 1 2, ..., Nb b b be the

    correspondingnormalizedeigenvectorsof K .

    8. Detect multicollinearity on K . Let r be theretained number of nonlinear PCs such that

    1

    1max .

    1000

    sr s

    9. Construct ll

    l

    b

    forl=1,2,...,rand ( )r =

    1 2( ... ) r .

    10. Calculate*

    ( ) ( ) ( ),

    r r r KU 1( ) ( )Tr r o D U z and

    1 2( ... )T

    Nc c cc =1 *

    ( ) ( )

    r rL

    11. Given a vector x , the WLS KPCRspredictionisgivenby

    1

    ( ) ( , )N

    i i

    i

    x y c k x xg

    Note that the above algorithm works under the

    assumption1

    ( ) 0N

    i

    i

    x

    . When1

    ( ) 0N

    i

    i

    x

    we

    replaceKby NK =K EK KE+EKEinStep6,whereEis

    the

    NNmatrix

    with

    all

    elements

    equal

    to

    1 ,

    Nand

    workbasedon NK inthesubsequentsteps

    3 CASE STUDY

    In our case study, we use the Gaussian kernel

    2( , ) exp( / )k x y x - y where is the parameter

    of thekernel.Asmentionedbefore that thereare several

    methodtoestimatetheweight iw seeforexample[2],[6],

    [7] and [14].Weuse themethodbasedon replication to

    estimatethe

    weight

    iw as follows.First,wearrange the

    data x inorderofincreasing iy andmakesomegroups,

    say M (

  • 8/3/2019 Kernel Based Nonlinear Weighted Least Squares Regression

    6/8

    ikx and variance of iky respectively.Thenwemake

    the prediction from the set 2,k kx s , say

    2 1 ( ) o f x c c x ,where 2f is a function from into

    and 1 ,oc c

    . Furthermore,we calculate the estimated

    varianceofi

    y byusingthepredictor 2 ( )if x .Theweight

    iw

    ischoseninverselyfromtheestimatedvarianceof .iy

    Thesimilarprocedure isperformed toobtaintheweights

    ofWLSKPCR . Weonly replacei

    y byoi

    y whereoi

    y is

    theithelementofo

    y .

    We also use the averagemonthly income from food

    sales (y) and the corresponding annual advertising

    expenses (x) for30 restaurants [6].Thedataaregiven in

    Table1andarecalledthetrainingdata.Weaddthenoises

    inthe

    response

    values

    of

    the

    training

    data

    which

    are

    generatedby a normally distributed random noisewith

    zeromeanandstandarddeviation 1 0,1 .Afterward,

    we use some of the data to test the predictionbyWLS

    basedlinearregressionandWLSKPCRandcallthemthe

    testingdata.Wealsoaddthenoisesintheresponsevalues

    of the testing data which are generated by a normally

    distributed random noisewith zeromean and standard

    deviation 2 where 2 0,1 . For the sake of

    comparison, we set 1 and 2 equal to 0.25 and 0.5,

    respectively.Then,

    we

    generated

    1000

    sets

    for

    both

    training and testing data to test the performance of the

    WLSbasedregressionandWLSKPCR.

    Table1:Therestaurantfoodssalesdata(yi 100)

    i xi yi(100)

    i xi yi(100)

    i xi yi(100)

    1 3.00 81.464 11 9.00 131.434 21 15.050 178.187

    2 3.150 72.661 12 11.345 140.564 22 178.187 185.304

    3 3.085 72.344 13 12.275 151.352 23 15.150 155.931

    4 5.225 90.743 14 12.400 146.426 24 16.800 172.579

    5 5.350 98.588 15 12.525 130.963 25 16.500 188.851

    6 6.090 96.507 16 12.310 144.630 26 17.830 192.424

    7 8.925 126.574 17 13.700 147.041 27 19.500 203.112

    8

    9.015114.133

    18

    15.000

    179.021

    28

    19.200 192.482

    9 8.885 115.814 19 15.175 166.200 29 19.000 218.715

    10 8.950 123.181 20 14.995 180.732 30 19.350 214.317

    Let ie and iy betheresidualandthepredictionofOLS

    method,respectively.Itiswellknownthattheplotofthe

    residual ie anditscorresponding iy isusefultocheckthe

    assumption of constant variance. Our choice of WLS

    solutionisalsobasedonthepatternofresiduals.Theplot

    of ie and iy

    isshowninFigure1(a)inwhichthevariation

    of the residuals increases significantly as the prediction

    valuesincrease.Hence,thisplotindicatesviolationof the

    assumption of constant variance. Consequently, the

    ordinary

    least

    squares

    fit

    inappropriate

    and

    the

    estimator

    ofOLS isnot thebest linearunbiased estimator (BLUE).

    Fortheshakeofcomparison,thevaluesofMarechosento

    betwo,fourandfive.ForinstancewhenM=2,itmeans

    (a)

    (b)

    Figure1:

    Plot

    of

    residual

    and

    its

    corresponding

    predicted

    valuefortrainingdata:(a)OLSbasedlinearregression,(b)

    WLSKPCR.

    that the ordered data is divided into two groupswhere

    eachgroup contains 50percentof theordereddata.The

    plotofresidual 2ie anditscorrespondingpredictionvalue

    oiz withM=2and =0.5isshowninFigure1(b).Figure

    JOURNAL OF COMPUTING, VOLUME 3, ISSUE 11, NOVEMBER 2011, ISSN 2151-9617

    HTTPS://SITES.GOOGLE.COM/SITE/JOURNALOFCOMPUTING

    WW.JOURNALOFCOMPUTING.ORG 61

  • 8/3/2019 Kernel Based Nonlinear Weighted Least Squares Regression

    7/8

    1(b) shows a residual plot with no systematic pattern

    around zero. It seems that the assumption of constant

    variance is satisfied for the data associated with this

    residualplot.Wenotice that the eigenvalues of are5.2012e+03 and 4.3800 and the eigenvalues of are1.05717e03and6.4251e07.Theratiooftheeigenvaluesof

    and are 4.3800/5.2012e+03 = 8.4228e04 and6.4251e07/1.05717e03=6.0779e04, respectively. Hence,multicollinearity exists in the regressionmatrices and

    . It implies that the prediction of OLS based linearregression and the prediction ofWLS regression canbe

    inappropriatetobeused.

    We also notice that the estimator of WLS KPCR is

    BLUE for regression model with heteroscedastic errors

    andprovidestheestimatewiththesmallestvariance.The

    comparisonofthetwomethodsisshowninTable2where

    thevaluesinthetableareaveragesofRMSEfor1000sets

    ofthe

    training

    data

    and

    the

    testing

    data

    in

    the

    original

    scale.FromTable2,weseethattheRMSEsofWLSKPCR

    are smaller compared to RMSEs ofWLS regression. For

    these data, the choice = 0.05 yieldsbetter results than

    othervaluesof sandWLSKPCRreduces RMSEofWLS

    regressiondowntomorethan50percent.

    Table2:RMSEofWLSregressionandWLSKPCRforthe

    restaurantfoodssalesdata.

    Data Method RSME

    M=2 M=4 M=5

    Training

    WLS

    regression

    870.3370 870.0094 869.9483

    WLSKPCR

    (=0.05,r=23)

    387.4919 387.4247 387.4118

    WLSKPCR

    (=0.1,r=21)

    430.6168 430.5629 430.5506

    WLSKPCR

    (=0.5,r=18)

    601.4776 601.4338 601.4254

    WLSKPCR

    (=1,r=17)

    624.2837 624.2512 624.2440

    Testing

    WLS

    regression

    835.0128 834.8192 835.2938

    WLSKPCR

    (=0.05,r=23)

    398.7448 398.7926 398.7979

    WLSKPCR

    (=0.1,r=21)

    463.4775 463.4887 463.4777

    WLSKPCR

    (=0.5,r=18)

    689.8822 689.9098 689.8971

    WLSKPCR

    (=1,r=17)

    721.3939 721.3949 721.3855

    4 CONCLUSIONS

    WLS is a technique to be used in case of a regression

    model with non constant variances. However, this

    techniqueyieldsalinearpredictionandhasnoguarantee

    thatitcanhandlethenegativeeffectsofmulticollinearity.

    Inthis

    paper,

    we

    proposed

    WLS

    KPCR

    to

    be

    used

    in

    a

    heteroscedastic regressionmodel and to overcome those

    limitations.We should notice thatWLS KPCR yields a

    nonlinear prediction and can avoid the effects of

    multicollinearity inheteroscedasticregressionmodel.The

    estimateofWLSKPCRsregressioncoefficientsisthebest

    linear unbiased estimator in which the proof of the

    unbiasedestimatorofWLSKPCRfollowstheproofofthe

    unbiased estimator ofPCR (See [15]). In our case study,

    WLS KPCR outperforms WLS regression in a

    heteroscedasticmodel.

    ACKNOWLEGDEMENT

    The authors sincerely thank to Universiti Teknologi

    Malaysia and Ministry of Higher Education (MOHE)

    Malaysia for Research University Grant (RUG) with

    numberQ.J130000.7128.02J88. In addition,we also thankto TheResearchManagement Center (RMC) UTM for

    supportingthisresearchproject.

    REFERENCES

    [1]S.ChattefueeandA.S.Hadi.RegressionbyExample:FouthEdition.

    JohnWileyandSons,2006.

    [2]J.J.Faraway.LinearModelswithR.ChapmanandHall/CRC,2005.

    [3] L. Hoegaerts,J.A.K. Suykens, J. Vandewalle, and B. D.Moor.

    Subsetbased least squares subspace in reproducing kernel hilbert

    space.Neurocomputing,pages293323,2005.

    [4]A.M.Jade,B.Srikanth,B.DKulkari,J.PJog,andL.Priya.Feature

    extraction and denoising using kernel pca. Chemical Engineering

    Sciences,58:44414448,2003.

    [5]

    H.

    Q.

    Minh,

    P.

    Niyogi,

    and

    Y.

    Yao.

    Mercers

    theorem,

    feature

    maps, and smoothing. Lecture Notes in Computer Science, Springer

    Berling,4005/2006,2009.

    [6]D.C.Montgomery,ElizabethA. Peck, andG.GeoffreyVining.

    IntroductiontoLinearRegression.WileyInterscience,2006.

    [7] D. R. Norman and H. Smith.Applied RegressionAnalysis.John

    WileyandSons,1998.

    JOURNAL OF COMPUTING, VOLUME 3, ISSUE 11, NOVEMBER 2011, ISSN 2151-9617

    HTTPS://SITES.GOOGLE.COM/SITE/JOURNALOFCOMPUTING

    WW.JOURNALOFCOMPUTING.ORG 62

  • 8/3/2019 Kernel Based Nonlinear Weighted Least Squares Regression

    8/8

    [8]J.O.Rawlings,SastryG.Pantula,andDavidA.Dickey.Applied

    regressionanalysis:aresearchtool.Springer,1998.

    [9]R.Rosipal,M.Girolami,L.J.Trejo,andAndrzejCichoki.Kernel

    pca for feature extraction and denoising in nonlinear regression.

    NeuralComputingandApplications,pages231243,2001.

    [10]

    R.

    Rosipal

    and

    L.

    J.

    Trejo.

    Kernel

    partial

    least

    squares

    regression

    in reproducing kernel hilbert space. Journal of Machine Learning

    Research,2:97123,2002.

    [11] R. Rosipal, L. J. Trejo, and A. Cichoki. Kernel principal

    component regression with em approach to nonlinear principal

    componentextraction.TechnicalReport,UniversityofPaisley,UK,2001.

    [12]B.Scholkopf,A.Smola,andK.R.Muller.Nonlinearcomponent

    analysisasakerneleigenvalueproblem.NeuralComputation,10:1299

    1319,1998.

    [13] B. Scholkopf and A.J. Smola. Learning with kernels. TheMIT

    Press.,2002.

    [14]G.A.F.SeberandA.J.Lee.LinearRegressionAnalysis.JohnWiley

    andSons,Inc.,2003.

    [15]M.S.Srivastava.MethodsofMultivariateStatistics.JohnWileyand

    Sons,Inc.,2002.

    [16] A. Wibowo and Y. Yamamoto. A note of kernel principal

    component regression. To appear in Computational Mathematics and

    Modeling.

    Antoni Wibowo is currently working as a senior lecturer in the facultyof computer science and information systems in UTM. He receivedB.Sc in Math Engineering from University of Sebelas Maret (UNS)Indonesia and M.Sc in Computer Science from University of

    Indonesia. He also received M. Eng and Dr. Eng in System andInformation Engineering from University of Tsukuba Japan. Hisinterests are in the field of computational intelligence, machinelearning, operations research and data analysis.

    Mohamad Ishak Desa is a professor in the faculty of computerscience and information systems in UTM. He received B.Sc. inMathematics from UKM in Malaysia, along an advance diploma insystem analysis from Aston University. He received a M.A. inMathematics from University of Illinois, and then, a PhD in operationsresearch from Salford University in UK. He is currently the Head ofOperation Business Intelligence Research Group in UTM. Hisinterests are operations research, optimization, logistic and supplychain.

    JOURNAL OF COMPUTING, VOLUME 3, ISSUE 11, NOVEMBER 2011, ISSN 2151-9617

    HTTPS://SITES.GOOGLE.COM/SITE/JOURNALOFCOMPUTING

    WW.JOURNALOFCOMPUTING.ORG 63