이 제출물을 팔로우합니다
- 팔로우하는 게시물 피드에서 업데이트를 확인할 수 있습니다
- 정보 수신 기본 설정에 따라 이메일을 받을 수 있습니다
Refer: An Introduction to Support Vector Machines and Other Kernel-based Learning Methods by Nello Cristianini and John Shawe-Taylor]
In this demo: training or cross-validation of a support vector machine (SVM) model for two-class (binary) classification on a low dimensional data set.
The training algorithm only depend on the data through dot products in H, i.e. on functions of the form Φ(x_i)·Φ(x_j). Now if there were a “kernel function” K such that
K(x_i,x_j) = Φ(x_i)·Φ(x_j),
we would only need to use K in the training algorithm, and would never need to explicitly even know what Φ is. One example is radial basis functions (RBF) or gaussian kernels where, H is infinite dimensional, so it would not be very easy to work with Φ explicitly.
Training the model requires the choice of:
• the kernel function, that determines the shape of the decision surface
• parameters in the kernel function (eg: for gaussian kernel:variance of the Gaussian, for polynomial kernel: degree of the polynomial)
• the regularization parameter λ.
Related Examples:
1. AdaBoost
https://in.mathworks.com/matlabcentral/fileexchange/63156-adaboost
2. SVM using various kernels
https://in.mathworks.com/matlabcentral/fileexchange/63033-svm-using-various-kernels
3. SVM for nonlinear classification
https://in.mathworks.com/matlabcentral/fileexchange/63024-svm-for-nonlinear-classification
4. SMO
https://in.mathworks.com/matlabcentral/fileexchange/63100-smo--sequential-minimal-optimization-
인용 양식
Bhartendu (2026). Support Vector Machine (https://kr.mathworks.com/matlabcentral/fileexchange/63158-support-vector-machine), MATLAB Central File Exchange. 검색 날짜: .
| 버전 | 퍼블리시됨 | 릴리스 정보 | Action |
|---|---|---|---|
| 1.0.0.0 |
