|
||||||||||
PREV PACKAGE NEXT PACKAGE | FRAMES NO FRAMES |
See:
Description
Class Summary | |
---|---|
Cache | SVM Kernel Cache, implementing a least recently used (LRU) policy. |
Kernel | Abstract interface to model an SVM Kernel. |
ONECLASSQ | Inner class representing a Kernel matrix for SVM Distribution Estimation. |
Solver | Generalized SMO+SVMlight algorithm Solves: min 0.5(\alpha^T Q \alpha) + b^T \alpha y^T \alpha = \delta y_i = +1 or -1 0 <= alpha_i <= Cp for y_i = 1 0 <= alpha_i <= Cn for y_i = -1 Given: Q, b, y, Cp, Cn, and an initial feasible point \alpha l is the size of vectors and matrices eps is the stopping criterion solution will be put in \alpha, objective value will be put in obj |
Solver.SolutionInfo | |
SolverNU | Solver for nu-svm classification and regression additional constraint: e^T \alpha = constant |
SVCQ | Inner class representing a Kernel matrix for Support Vector classification. |
SVM | Construct and solve various formulations of the support vector machine (SVM) problem. |
SVM.decisionFunction | Inner class modeling the data for the SVM decision function, used for classifying points with respect to the hyperplane. |
SVMCategorizer | Simple, easy-to-use, and efficient software for SVM classification and regression, based on the LIBSVM implementation of Chin-Chung Chang and Chin-Jen Lin. |
SVMModel | SVMModel encondes a classification model, describing both the model parameters and the Support Vectors. |
SVMNode | SVMNode is used to model dimentions in vectors. |
SVMParameter | Constants and Parameters used used in the SVM package. |
SVMProblem | Class to model an SVM Problem, containing both the training vectors and the class (value) associated with each vector. |
SVRQ | Inner class representing a Kernel matrix for Support Vector regression. |
Implementation of Support Vector Machines classification and regression that
can be used to categorize text files using N-Grams as features.
A support vector machine is a supervised learning algorithm developed over the past
decade by Vapnik and others (Vapnik, Statistical Learning Theory, 1998).
It operates by mapping the given training set into a possibly high-dimensional feature
space and attempting to locate in that space a plane that separates the positive
from the negative examples.
|
||||||||||
PREV PACKAGE NEXT PACKAGE | FRAMES NO FRAMES |