Data Mining: The Textbook



Yüklə 17,13 Mb.
səhifə192/423
tarix07.01.2024
ölçüsü17,13 Mb.
#211690
1   ...   188   189   190   191   192   193   194   195   ...   423
1-Data Mining tarjima

APPROX. DECISION































BOUNDARY
















1.6





































1.5

0.6

0.7

0.8

0.9

1

1.1

1.2

1.3

1.4

1.5







0.5






















FEATURE X



















Figure 10.8: Nonlinear decision surface


The decision boundary of Eq. 10.60 can be expressed linearly in terms of the variables z1 . . . z4, by expanding Eq. 10.60 in terms of x1, x21, x2, and x22:
8x2 16x + 50x2 200x + 207 = 0
1 1 2 2
Thus, each training data point is now expressed in terms of these four newly transformed dimensions, and the classes will be linearly separable in this space. The SVM optimization formulation can then be solved in the transformed space as a linear model, and used to classify test instances that are also transformed to 4-dimensional space. It is important to note that the complexity of the problem effectively increased because of the increase in the size of the hyperplane coefficient vector W .

In general, it is possible to approximate any polynomial decision boundary by adding an additional set of dimensions for each exponent of the polynomial. High-degree polynomials have significant expressive power in approximating many nonlinear functions well. This kind of transformation can be very effective in cases where one does not know whether the decision boundary is linear or nonlinear. This is because the additional degrees of freedom in the model, in terms of the greater number of coefficients to be learned, can determine the linearity or nonlinearity of the decision boundary in a data-driven way. In our previous example, if the decision boundary had been linear, the coefficients for z1 and z3 would automatically have been learned to be almost 0, given enough training data. The price for this additional flexibility is the increased computational complexity of the training problem, and the larger number of coefficients that need to be learned. Furthermore, if enough training data is not available, then this may result in overfitting where even a simple linear decision boundary is incorrectly approximated as a nonlinear one. A different approach, which is sometimes used to learn nonlinear decision boundaries, is known as the “kernel trick.” This approach is able to learn arbitrary decision boundaries without performing the transformation explicitly.




Yüklə 17,13 Mb.

Dostları ilə paylaş:
1   ...   188   189   190   191   192   193   194   195   ...   423




Verilənlər bazası müəlliflik hüququ ilə müdafiə olunur ©azkurs.org 2024
rəhbərliyinə müraciət

gir | qeydiyyatdan keç
    Ana səhifə


yükləyin