Convex Loss Functions for Support Vector Machines (SVMs) and Neural Networks
About
We propose a new convex loss for Support Vector Machines, both for the binary classification and for the regression models. Therefore, we show the mathematical derivation of the dual problems and we experiment with them on several small datasets. The minimal dimension of those datasets is due to the difficult scalability of the SVM method to bigger instances. This preliminary study should prove that using pattern correlations inside the loss function could enhance the generalisation performances. Our method consistently achieved comparable or superior performance, with improvements of up to 2.0% in F1 scores for classification tasks and 1.0% reduction in Mean Squared Error (MSE) for regression tasks across various datasets, compared to standard losses. Coherently, results show that generalisation measures are never worse than the standard losses and several times they are better. In our opinion, it should be considered a careful study of this loss, coupled with shallow and deep neural networks. In fact, we present some novel results obtained with those architectures.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Graph Classification | PROTEINS (test) | Accuracy77.6213 | 180 | |
| Graph Classification | NCI1 (test) | Accuracy77.7778 | 174 | |
| Image Classification | Masks | Mean Accuracy94.1165 | 40 | |
| Binary Classification | Heart | Mean F10.8751 | 8 | |
| Binary Classification | Iono | Mean F10.9754 | 8 | |
| Binary Classification | Haberman | Mean F10.6095 | 8 | |
| Binary Classification | Breast | Mean F1 Score97.6717 | 8 | |
| Binary Classification | Sonar (test) | Mean F1 Score90.2096 | 8 | |
| Binary Classification | German | Mean F1 Score65.5664 | 8 | |
| Binary Classification | WDBC | Mean F1 Score95.7907 | 8 |