日本語
 
Help Privacy Policy ポリシー/免責事項
  詳細検索ブラウズ

アイテム詳細


公開

会議論文

An Efficient Method for Gradient-Based Adaptation of Hyperparameters in SVM Models

MPS-Authors
/persons/resource/persons83855

Chapelle,  O
Department Empirical Inference, Max Planck Institute for Biological Cybernetics, Max Planck Society;
Max Planck Institute for Biological Cybernetics, Max Planck Society;

Fulltext (restricted access)
There are currently no full texts shared for your IP range.
フルテキスト (公開)
公開されているフルテキストはありません
付随資料 (公開)
There is no public supplementary material available
引用

Keerthi, S., Sindhwani, V., & Chapelle, O. (2007). An Efficient Method for Gradient-Based Adaptation of Hyperparameters in SVM Models. Advances in Neural Information Processing Systems 19, 673-680.


引用: https://hdl.handle.net/11858/00-001M-0000-0013-CBD3-F
要旨
We consider the task of tuning hyperparameters in SVM models based on minimizing a smooth performance validation function, e.g., smoothed k-fold cross-validation error,
using non-linear optimization techniques. The key computation in this approach is that of the gradient of the validation function with respect to hyperparameters. We show that for large-scale problems involving a wide choice of kernel-based models and validation functions, this computation can be very efficiently done; often within just a fraction of the training time. Empirical results show that a near-optimal set of hyperparameters can be identified by our approach with very few training rounds and gradient computations.