On low-rank regularized least squares for scalable nonlinear classification

Zhouyu Fu, Guojun Lu, Kai-Ming Ting, Dengsheng Zhang

    Research output: Chapter in Book / Conference PaperConference Paperpeer-review

    2 Citations (Scopus)

    Abstract

    In this paper, we revisited the classical technique of Regularized Least Squares (RLS) for the classification of large-scale nonlinear data. Specifically, we focus on a low-rank formulation of RLS and show that it has linear time complexity in the data size only and does not rely on the number of labels and features for problems with moderate feature dimension. This makes low-rank RLS particularly suitable for classification with large data sets. Moreover, we have proposed a general theorem for the closed-form solutions to the Leave-One-Out Cross Validation (LOOCV) estimation problem in empirical risk minimization which encompasses all types of RLS classifiers as special cases. This eliminates the reliance on cross validation, a computationally expensive process for parameter selection, and greatly accelerate the training process of RLS classifiers. Experimental results on real and synthetic large-scale benchmark data sets have shown that low-rank RLS achieves comparable classification performance while being much more efficient than standard kernel SVM for nonlinear classification. The improvement in efficiency is more evident for data sets with higher dimensions.
    Original languageEnglish
    Title of host publicationNeural Information Processing: 18th International Conference, ICONIP 2011, Shanghai, China, November 13-17, 2011: Proceedings, Part II
    PublisherSpringer
    Pages490-499
    Number of pages10
    ISBN (Print)9783642249570
    DOIs
    Publication statusPublished - 2011
    EventICONIP (Conference) -
    Duration: 13 Nov 2011 → …

    Publication series

    Name
    ISSN (Print)0302-9743

    Conference

    ConferenceICONIP (Conference)
    Period13/11/11 → …

    Fingerprint

    Dive into the research topics of 'On low-rank regularized least squares for scalable nonlinear classification'. Together they form a unique fingerprint.

    Cite this