Hi,
Well, I have a hint which can help. There is a new bound for the number of the
Mistakes the Perceptron algorithm makes. In terms of SVM the bound is
proportional to (for all w)
(1/2)||w||^2 R^2 + \sum_i \max\{0, 1-y_i(w*x_i)\}
where it is assumed that all the examples are in a ball of radius R around the
origin.
If you take the factor R^2 out of the bound you get the SVM optimization
problem, where C=1/R^2 ; a max operator is used rather than the mean operator
you asked about. I am not sure if the bound still holds if we replace the
max with a mean. In practice, the mean is more stable than the max. (Although I would try to use
also the median).
Regards, Koby
============================================
Koby Crammer [log in to unmask]
http://www.cs.huji.ac.il/~kobics
============================================
On Thu, 27 May 2004, Dave Lewis wrote:
| Hi - SVM Light has as the default value of its regularization parameter
| (tradeoff between training error and margin) the reciprocal of the mean
| 2-norm of the training examples. Does anyone know what the statistical
| justification for this particular choice is? I can't seem to find this in
| Thorsten Joachim's papers.
|
| Regards, Dave
| http://www.DavidDLewis.com
|
|