Let X be a nonempty set, K a positive-definite real-valued kernel on XΓX with corresponding reproducing kernel Hilbert space Hkβ, and let R:HkββR be a differentiable regularization function. Then given a training sample (x1β,y1β),β¦,(xnβ,ynβ)βXΓR and an arbitrary error function E:(XΓR2)mβRβͺ{β}, a minimizer fβ=fβHkβargminβ{E((x1β,y1β,f(x1β)),β¦,(xnβ,ynβ,f(xnβ)))+R(f)} of the regularized empirical risk admits a representation of the form fβ(β )=βi=1nβΞ±iβk(β ,xiβ), Why itβs cool
Representer theorems are useful from a practical standpoint because they dramatically simplify the regularized ERM problem.
In most interesting applications, the search domain Hkβ for the minimization will be an infinite-dimensional subspace of L2β(X) and therefore the search (as written) does not admit implementation on finite-memory and finite-precision computers.
In contrast, the representation of fβ(β ) afforded by a representer theorem reduces the original (infinite-dimensional) minimization problem to a search for the optimal n-dimensional vector of coefficients Ξ±; it can then be obtained by applying any standard function minimization algorithm.