### 2011

Santiago-Mozos, Ricardo; Perez-Cruz, Fernando; Artés-Rodríguez, Antonio

Extended Input Space Support Vector Machine Artículo de revista

En: IEEE transactions on neural networks / a publication of the IEEE Neural Networks Council, vol. 22, no. 1, pp. 158–163, 2011, ISSN: 1941-0093.

Resumen | Enlaces | BibTeX | Etiquetas: Algorithms, Artificial Intelligence, Automated, Automated: standards, Computer Simulation, Computer Simulation: standards, Neural Networks (Computer), Pattern recognition, Problem Solving, Software Design, Software Validation

@article{Santiago-Mozos2011,

title = {Extended Input Space Support Vector Machine},

author = {Ricardo Santiago-Mozos and Fernando Perez-Cruz and Antonio Art\'{e}s-Rodr\'{i}guez},

url = {http://www.tsc.uc3m.es/~antonio/papers/P38_2011_Extended Input Space Support Vector Machine.pdf

http://www.ncbi.nlm.nih.gov/pubmed/21095866},

issn = {1941-0093},

year = {2011},

date = {2011-01-01},

journal = {IEEE transactions on neural networks / a publication of the IEEE Neural Networks Council},

volume = {22},

number = {1},

pages = {158--163},

abstract = {In some applications, the probability of error of a given classifier is too high for its practical application, but we are allowed to gather more independent test samples from the same class to reduce the probability of error of the final decision. From the point of view of hypothesis testing, the solution is given by the Neyman-Pearson lemma. However, there is no equivalent result to the Neyman-Pearson lemma when the likelihoods are unknown, and we are given a training dataset. In this brief, we explore two alternatives. First, we combine the soft (probabilistic) outputs of a given classifier to produce a consensus labeling for K test samples. In the second approach, we build a new classifier that directly computes the label for K test samples. For this second approach, we need to define an extended input space training set and incorporate the known symmetries in the classifier. This latter approach gives more accurate results, as it only requires an accurate classification boundary, while the former needs an accurate posterior probability estimate for the whole input space. We illustrate our results with well-known databases.},

keywords = {Algorithms, Artificial Intelligence, Automated, Automated: standards, Computer Simulation, Computer Simulation: standards, Neural Networks (Computer), Pattern recognition, Problem Solving, Software Design, Software Validation},

pubstate = {published},

tppubtype = {article}

}

In some applications, the probability of error of a given classifier is too high for its practical application, but we are allowed to gather more independent test samples from the same class to reduce the probability of error of the final decision. From the point of view of hypothesis testing, the solution is given by the Neyman-Pearson lemma. However, there is no equivalent result to the Neyman-Pearson lemma when the likelihoods are unknown, and we are given a training dataset. In this brief, we explore two alternatives. First, we combine the soft (probabilistic) outputs of a given classifier to produce a consensus labeling for K test samples. In the second approach, we build a new classifier that directly computes the label for K test samples. For this second approach, we need to define an extended input space training set and incorporate the known symmetries in the classifier. This latter approach gives more accurate results, as it only requires an accurate classification boundary, while the former needs an accurate posterior probability estimate for the whole input space. We illustrate our results with well-known databases.