English
 
User Manual Privacy Policy Disclaimer Contact us
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT
  Support Vector Machines and Kernels for Computational Biology

Ben-Hur, A., Ong, C., Sonnenburg, S., Schölkopf, B., & Rätsch, G. (2008). Support Vector Machines and Kernels for Computational Biology. PLoS Computational Biology, 4(10), 1-10. doi:10.1371/journal.pcbi.1000173.

Item is

Basic

show hide
Item Permalink: http://hdl.handle.net/11858/00-001M-0000-0013-C6B3-5 Version Permalink: http://hdl.handle.net/21.11116/0000-0003-2CD8-9
Genre: Journal Article

Files

show Files

Creators

show
hide
 Creators:
Ben-Hur, A, Author
Ong, CS1, 2, 3, Author              
Sonnenburg, S1, 2, Author              
Schölkopf, B1, 2, Author              
Rätsch, G3, Author              
Affiliations:
1Department Empirical Inference, Max Planck Institute for Biological Cybernetics, Max Planck Society, ou_1497795              
2Max Planck Institute for Biological Cybernetics, Max Planck Society, ou_1497794              
3Friedrich Miescher Laboratory, Max Planck Society, Max-Planck-Ring 9, 72076 Tübingen, DE, ou_2575692              

Content

show
hide
Free keywords: -
 Abstract: The increasing wealth of biological data coming from a large variety of platforms and the continued development of new high-throughput methods for probing biological systems require increasingly more sophisticated computational approaches. Putting all these data in simple-to-use databases is a first step; but realizing the full potential of the data requires algorithms that automatically extract regularities from the data, which can then lead to biological insight. Many of the problems in computational biology are in the form of prediction: starting from prediction of a gene's structure, prediction of its function, interactions, and role in disease. Support vector machines (SVMs) and related kernel methods are extremely good at solving such problems [1]–[3]. SVMs are widely used in computational biology due to their high accuracy, their ability to deal with high-dimensional and large datasets, and their flexibility in modeling diverse sources of data [2], [4]–[6]. The simplest form of a prediction problem is binary classification: trying to discriminate between objects that belong to one of two categories—positive (+1) or negative (−1). SVMs use two key concepts to solve this problem: large margin separation and kernel functions. The idea of large margin separation can be motivated by classification of points in two dimensions (see Figure 1). A simple way to classify the points is to draw a straight line and call points lying on one side positive and on the other side negative. If the two sets are well separated, one would intuitively draw the separating line such that it is as far as possible away from the points in both sets (see Figures 2 and 3). This intuitive choice captures the idea of large margin separation, which is mathematically formulated in the section Classification with Large Margin.

Details

show
hide
Language(s):
 Dates: 2008-10
 Publication Status: Published in print
 Pages: -
 Publishing info: -
 Table of Contents: -
 Rev. Method: -
 Identifiers: DOI: 10.1371/journal.pcbi.1000173
eDoc: e1000173
BibTex Citekey: 5586
 Degree: -

Event

show

Legal Case

show

Project information

show

Source 1

show
hide
Title: PLoS Computational Biology
Source Genre: Journal
 Creator(s):
Affiliations:
Publ. Info: San Francisco, CA : Public Library of Science
Pages: - Volume / Issue: 4 (10) Sequence Number: - Start / End Page: 1 - 10 Identifier: ISSN: 1553-734X
CoNE: https://pure.mpg.de/cone/journals/resource/1000000000017180_1