A string kernel is a mathematical tool used in large scale data analysis and mining, where sequence data are to be clustered or classified (concerning especially the popular research fields of text and gene analysis[1]). Kernels are often used in with support vector machines to transform data from its original space to one where it can be more easily separated and grouped. This may enable non-linear classification.
Contents |
Suppose one wants to compare some text passages automatically and indicate their relative similarity. For many applications, it might be sufficient to find some keywords which match exactly. One example where exact matching is not always enough is found in spam detection[2]; another would be in computational gene analysis, where homologous genes have mutated, resulting in common subsequences along with deleted, inserted or replaced symbols.
Since several well-proven data clustering, classification and information retrieval methods (for example support vector machines) are designed to work on vectors (i.e. data are elements of a vector space), using a string kernel allows the extension of these methods to handle sequence data.
The string kernel method is to be contrasted with earlier approaches for text classification where feature vectors only indicated the presence or absence of a word. Not only does it improve on these approaches, but it is an example for a whole class of kernels adapted to data structures, which began to appear at the turn of the 21st century. A survey of such methods has been compiled by Gärtner.[3]
A kernel on a domain is a function satisfying some conditions (being symmetric in the arguments, continuous and positive semidefinite in a certain sense).
Mercer's theorem asserts that can then be expressed as with mapping the arguments into an inner product space.
We can now reproduce the definition of a string subsequence kernel[4] on strings over an alphabet . Coordinate-wise, the mapping is defined as follows:
The are multiindices and is a string of length : subsequences can occur in a non-contiguous manner, but gaps are penalized. The parameter may be set to any value between (gaps are not allowed) and (even widely-spread "occurrences" are weighted the same as appearances as a contiguous substring).
For several relevant algorithms, data enters into the algorithm only in expressions involving an inner product of feature vectors, hence the name kernel methods. A desirable consequence of this is that one does not need to explicitly calculate the transformation , only the inner product via the kernel, which may be a lot quicker, especially when approximated.[4]