Difference between revisions of "Orange: SVM"

From OnnoWiki
Jump to navigation Jump to search
(Created page with "Sumber: https://docs.biolab.si//3/visual-programming/widgets/model/svm.html Support Vector Machines map inputs to higher-dimensional feature spaces. Inputs Data: input...")
 
 
(9 intermediate revisions by the same user not shown)
Line 2: Line 2:
  
  
Support Vector Machines map inputs to higher-dimensional feature spaces.
+
Widget SVM (Support Vector Machines) melakukan mapping dari input ke higher-dimensional feature space.
  
Inputs
+
==Input==
  
    Data: input dataset
+
Data: input dataset
 +
Preprocessor: preprocessing method(s)
  
    Preprocessor: preprocessing method(s)
+
==Output==
  
Outputs
+
Learner: linear regression learning algorithm
 +
Model: trained model
 +
Support Vectors: instances used as support vectors
  
    Learner: linear regression learning algorithm
+
Support Vector machine (SVM) adalah teknik machine learning yang memisahkan ruang atribut dengan hyperplane, sehingga memaksimalkan margin antara instance class yang berbeda atau nilai class. Teknik ini sering menghasilkan hasil kinerja prediksi tertinggi. Orange menanamkan implementasi populer SVM dari paket LIBSVM. Widget SVM adalah graphical user interface-nya.
  
    Model: trained model
+
Untuk task regression, SVM akan melakukan linear regression di high dimension feature space menggunakan  ε-insensitive loss. Dia akan mengestimasi keakuratan tergantung pada settingan parameter C, ε dan kernel. Widget SVM akan mengeluarkan prediksi class berdasarkan SVM Regression.
  
    Support Vectors: instances used as support vectors
+
Widget SVM dapat berfungsi untuk task classification dan regression.
  
Support vector machine (SVM) is a machine learning technique that separates the attribute space with a hyperplane, thus maximizing the margin between the instances of different classes or class values. The technique often yields supreme predictive performance results. Orange embeds a popular implementation of SVM from the LIBSVM package. This widget is its graphical user interface.
+
[[File:SVM-stamped.png|center|200px|thumb]]
  
For regression tasks, SVM performs linear regression in a high dimension feature space using an ε-insensitive loss. Its estimation accuracy depends on a good setting of C, ε and kernel parameters. The widget outputs class predictions based on a SVM Regression.
+
* The learner can be given a name under which it will appear in other widgets. The default name is “SVM”.
 +
* SVM type with test error settings. SVM and ν-SVM are based on different minimization of the error function. On the right side, you can set test error bounds:
  
The widget works for both classification and regression tasks.
+
** SVM:
 +
*** Cost: penalty term for loss and applies for classification and regression tasks.
 +
*** ε: a parameter to the epsilon-SVR model, applies to regression tasks. Defines the distance from true values within which no penalty is associated with predicted values.
  
../../_images/SVM-stamped.png
+
** ν-SVM:
 +
*** Cost: penalty term for loss and applies only to regression tasks
 +
*** ν: a parameter to the ν-SVR model, applies to classification and regression tasks. An upper bound on the fraction of training errors and a lower bound of the fraction of support vectors.
  
    The learner can be given a name under which it will appear in other widgets. The default name is “SVM”.
+
* Kernel is a function that transforms attribute space to a new feature space to fit the maximum-margin hyperplane, thus allowing the algorithm to create the model with Linear, Polynomial, RBF and Sigmoid kernels. Functions that specify the kernel are presented upon selecting them, and the constants involved are:
 +
** g for the gamma constant in kernel function (the recommended value is 1/k, where k is the number of the attributes, but since there may be no training set given to the widget the default is 0 and the user has to set this option manually),
 +
** c for the constant c0 in the kernel function (default 0), and
 +
** d for the degree of the kernel (default 3).
  
    SVM type with test error settings. SVM and ν-SVM are based on different minimization of the error function. On the right side, you can set test error bounds:
+
* Set permitted deviation from the expected value in Numerical Tolerance. Tick the box next to Iteration Limit to set the maximum number of iterations permitted.
 +
* Produce a report.
 +
* Click Apply to commit changes. If you tick the box on the left side of the Apply button, changes will be communicated automatically.
  
        SVM:
+
==Contoh==
  
            Cost: penalty term for loss and applies for classification and regression tasks.
+
Untuk contoh regresi, kita menggunakan dataset housing dan membagi data menjadi dua subset data (Data Sampel dan Data Sisa / Remaining) menggunakan widget Data Sampler. Sampel dikirim ke Widget SVM yang menghasilkan Model, yang kemudian digunakan dalam Widget Predictions untuk memprediksi nilai pada Data Sisa. Skema serupa dapat digunakan jika data sudah dalam dua file terpisah; dalam hal ini, dua widget File akan digunakan daripada memecah data menggunakan Widget File - Widget Data Sampler.
  
            ε: a parameter to the epsilon-SVR model, applies to regression tasks. Defines the distance from true values within which no penalty is associated with predicted values.
+
[[File:SVM-Predictions.png|center|600px|thumb]]
  
        ν-SVM:
+
Contoh berikut menunjukkan cara menggunakan widget SVM dalam kombinasi dengan widget Scatter Plot. Workflow berikut men-train model widget SVM pada data iris dan mengeluarkan support vector, yang merupakan contoh data yang digunakan sebagai support vector pada fase learning. Kita dapat mengamati contoh data ini dalam visualisasi widget Scatter Plot. Perhatikan bahwa agar workflow berfungsi dengan benar, kita harus mengatur link antara widget seperti yang ditunjukkan pada screenshot di bawah ini.
  
            Cost: penalty term for loss and applies only to regression tasks
+
[[File:SVM-support-vectors.png|center|600px|thumb]]
  
            ν: a parameter to the ν-SVR model, applies to classification and regression tasks. An upper bound on the fraction of training errors and a lower bound of the fraction of support vectors.
+
==Youtube==
  
    Kernel is a function that transforms attribute space to a new feature space to fit the maximum-margin hyperplane, thus allowing the algorithm to create the model with Linear, Polynomial, RBF and Sigmoid kernels. Functions that specify the kernel are presented upon selecting them, and the constants involved are:
+
* [https://www.youtube.com/watch?v=mbuEa8L24fw YOUTUBE: ORANGE Prediksi Menggunakan SVM]
  
        g for the gamma constant in kernel function (the recommended value is 1/k, where k is the number of the attributes, but since there may be no training set given to the widget the default is 0 and the user has to set this option manually),
 
  
        c for the constant c0 in the kernel function (default 0), and
+
==Referensi==
 
 
        d for the degree of the kernel (default 3).
 
 
 
    Set permitted deviation from the expected value in Numerical Tolerance. Tick the box next to Iteration Limit to set the maximum number of iterations permitted.
 
 
 
    Produce a report.
 
 
 
    Click Apply to commit changes. If you tick the box on the left side of the Apply button, changes will be communicated automatically.
 
 
 
Examples
 
 
 
In the first (regression) example, we have used housing dataset and split the data into two data subsets (Data Sample and Remaining Data) with Data Sampler. The sample was sent to SVM which produced a Model, which was then used in Predictions to predict the values in Remaining Data. A similar schema can be used if the data is already in two separate files; in this case, two File widgets would be used instead of the File - Data Sampler combination.
 
 
 
../../_images/SVM-Predictions.png
 
 
 
The second example shows how to use SVM in combination with Scatter Plot. The following workflow trains a SVM model on iris data and outputs support vectors, which are those data instances that were used as support vectors in the learning phase. We can observe which are these data instances in a scatter plot visualization. Note that for the workflow to work correctly, you must set the links between widgets as demonstrated in the screenshot below.
 
 
 
../../_images/SVM-support-vectors.png
 
References
 
  
 
Introduction to SVM on StatSoft.
 
Introduction to SVM on StatSoft.
 
 
 
  
 
==Referensi==
 
==Referensi==

Latest revision as of 04:46, 13 April 2020

Sumber: https://docs.biolab.si//3/visual-programming/widgets/model/svm.html


Widget SVM (Support Vector Machines) melakukan mapping dari input ke higher-dimensional feature space.

Input

Data: input dataset
Preprocessor: preprocessing method(s)

Output

Learner: linear regression learning algorithm
Model: trained model
Support Vectors: instances used as support vectors

Support Vector machine (SVM) adalah teknik machine learning yang memisahkan ruang atribut dengan hyperplane, sehingga memaksimalkan margin antara instance class yang berbeda atau nilai class. Teknik ini sering menghasilkan hasil kinerja prediksi tertinggi. Orange menanamkan implementasi populer SVM dari paket LIBSVM. Widget SVM adalah graphical user interface-nya.

Untuk task regression, SVM akan melakukan linear regression di high dimension feature space menggunakan ε-insensitive loss. Dia akan mengestimasi keakuratan tergantung pada settingan parameter C, ε dan kernel. Widget SVM akan mengeluarkan prediksi class berdasarkan SVM Regression.

Widget SVM dapat berfungsi untuk task classification dan regression.

SVM-stamped.png
  • The learner can be given a name under which it will appear in other widgets. The default name is “SVM”.
  • SVM type with test error settings. SVM and ν-SVM are based on different minimization of the error function. On the right side, you can set test error bounds:
    • SVM:
      • Cost: penalty term for loss and applies for classification and regression tasks.
      • ε: a parameter to the epsilon-SVR model, applies to regression tasks. Defines the distance from true values within which no penalty is associated with predicted values.
    • ν-SVM:
      • Cost: penalty term for loss and applies only to regression tasks
      • ν: a parameter to the ν-SVR model, applies to classification and regression tasks. An upper bound on the fraction of training errors and a lower bound of the fraction of support vectors.
  • Kernel is a function that transforms attribute space to a new feature space to fit the maximum-margin hyperplane, thus allowing the algorithm to create the model with Linear, Polynomial, RBF and Sigmoid kernels. Functions that specify the kernel are presented upon selecting them, and the constants involved are:
    • g for the gamma constant in kernel function (the recommended value is 1/k, where k is the number of the attributes, but since there may be no training set given to the widget the default is 0 and the user has to set this option manually),
    • c for the constant c0 in the kernel function (default 0), and
    • d for the degree of the kernel (default 3).
  • Set permitted deviation from the expected value in Numerical Tolerance. Tick the box next to Iteration Limit to set the maximum number of iterations permitted.
  • Produce a report.
  • Click Apply to commit changes. If you tick the box on the left side of the Apply button, changes will be communicated automatically.

Contoh

Untuk contoh regresi, kita menggunakan dataset housing dan membagi data menjadi dua subset data (Data Sampel dan Data Sisa / Remaining) menggunakan widget Data Sampler. Sampel dikirim ke Widget SVM yang menghasilkan Model, yang kemudian digunakan dalam Widget Predictions untuk memprediksi nilai pada Data Sisa. Skema serupa dapat digunakan jika data sudah dalam dua file terpisah; dalam hal ini, dua widget File akan digunakan daripada memecah data menggunakan Widget File - Widget Data Sampler.

SVM-Predictions.png

Contoh berikut menunjukkan cara menggunakan widget SVM dalam kombinasi dengan widget Scatter Plot. Workflow berikut men-train model widget SVM pada data iris dan mengeluarkan support vector, yang merupakan contoh data yang digunakan sebagai support vector pada fase learning. Kita dapat mengamati contoh data ini dalam visualisasi widget Scatter Plot. Perhatikan bahwa agar workflow berfungsi dengan benar, kita harus mengatur link antara widget seperti yang ditunjukkan pada screenshot di bawah ini.

SVM-support-vectors.png

Youtube


Referensi

Introduction to SVM on StatSoft.

Referensi

Pranala Menarik