i just prepare some papers regarding support vector machines. As it is well know the kernel-trick enables us to transform data implicitly from the input space to some (potentially infinite dimensional) feature space.
As a short reference you can use Cristianini, Nello ; Shawe-Taylor, John: An Introduction to Support Vector Machines and Other Kernel-based Learning Methods. Cambridge: Cambridge University Press, 2000.
Since we then do not know the corresponding feature map, i wonder if there are any estimations about the dimensionality of the feature space, when we use kernels. Especially i would be interested if there are any results, stating when the data is linear separable in the resulting feature space. Maybe somebody knows some (recent) papers about this topic. I would be really interested!