Skip to Main Content
 

Global Search Box

 
 
 
 

Files

ETD Abstract Container

Abstract Header

Model Selection in Kernel Methods

Abstract Details

2011, Doctor of Philosophy, Ohio State University, Electrical and Computer Engineering.

Kernel methods have been extensively studied in pattern recognition and machine learning over the last decade, and they have been successfully used in a variety of applications. A main advantage of kernel methods is that nonlinear problems such as classification and regression can be efficiently solved using classical linear approaches. The performance of kernel methods greatly depends on the selected kernel model. The model is defined by the kernel mapping and its parameters. Different models result in different generalization performance. Hence, model selection in kernel methods is an important problem and remains a challenge in the literature. In this dissertation, we propose several approaches to address this problem. Our approaches can determine good learning models by optimizing both the kernels and all other parameters in the kernel-based algorithms.

In classification, we develop an algorithm yielding class distributions that are linearly separable in the kernel space. The idea is to enforce the homoscedasticity and separability of the pairwise class distributions simultaneously in the kernel space. We show how this approach can be employed to optimize kernels in discriminant analysis. We then derive a criterion to search for a good kernel representation by directly minimizing the Bayes classification error over different kernel mappings.

In regression, we derive a model selection approach to directly balance the model fit and model complexity using the framework of multiobjective optimization. We develop an algorithm to obtain the Pareto-optimal solutions which balance the trade-off between the model fit and model complexity. We show how the proposed method is related to minimizing the predicted generalization error of the learning function.

In our final algorithm, the kernel matrix is recursively learned with genetic algorithms until the classification/prediction error falls below a threshold. We derive a family of adaptive kernels to better fit the data with various densities and show their superiority over the commonly used fixed-shape kernels.

Extensive experimental results demonstrate that the proposed approaches are superior to the state of the art.

Aleix Martinez (Advisor)
Yuan Zheng (Committee Member)
Yoonkyung Lee (Committee Member)
186 p.

Recommended Citations

Citations

  • You, D. (2011). Model Selection in Kernel Methods [Doctoral dissertation, Ohio State University]. OhioLINK Electronic Theses and Dissertations Center. http://rave.ohiolink.edu/etdc/view?acc_num=osu1322581224

    APA Style (7th edition)

  • You, Di. Model Selection in Kernel Methods. 2011. Ohio State University, Doctoral dissertation. OhioLINK Electronic Theses and Dissertations Center, http://rave.ohiolink.edu/etdc/view?acc_num=osu1322581224.

    MLA Style (8th edition)

  • You, Di. "Model Selection in Kernel Methods." Doctoral dissertation, Ohio State University, 2011. http://rave.ohiolink.edu/etdc/view?acc_num=osu1322581224

    Chicago Manual of Style (17th edition)