Difference between revisions of "Orange: CN2 Rule Induction"

From OnnoWiki
Jump to navigation Jump to search
Line 3: Line 3:
 
Induce rules from data using CN2 algorithm.
 
Induce rules from data using CN2 algorithm.
  
Inputs
+
==Input==
  
    Data: input dataset
+
Data: input dataset
    Preprocessor: preprocessing method(s)
+
Preprocessor: preprocessing method(s)
  
Outputs
+
==Output==
  
    Learner: CN2 learning algorithm
+
Learner: CN2 learning algorithm
    CN2 Rule Classifier: trained model
+
CN2 Rule Classifier: trained model
  
 
The CN2 algorithm is a classification technique designed for the efficient induction of simple, comprehensible rules of form “if cond then predict class”, even in domains where noise may be present.
 
The CN2 algorithm is a classification technique designed for the efficient induction of simple, comprehensible rules of form “if cond then predict class”, even in domains where noise may be present.
Line 19: Line 19:
 
[[File:CN2-stamped.png|center|200px|thumb]]
 
[[File:CN2-stamped.png|center|200px|thumb]]
  
    Name under which the learner appears in other widgets. The default name is CN2 Rule Induction.
+
* Name under which the learner appears in other widgets. The default name is CN2 Rule Induction.
 +
* Rule ordering:
 +
** Ordered: induce ordered rules (decision list). Rule conditions are found and the majority class is assigned in the rule head.
 +
** Unordered: induce unordered rules (rule set). Learn rules for each class individually, in regard to the original learning data.
  
    Rule ordering:
+
* Covering algorithm:
        Ordered: induce ordered rules (decision list). Rule conditions are found and the majority class is assigned in the rule head.
+
** Exclusive: after covering a learning instance, remove it from further consideration.
        Unordered: induce unordered rules (rule set). Learn rules for each class individually, in regard to the original learning data.
+
** Weighted: after covering a learning instance, decrease its weight (multiplication by gamma) and in-turn decrease its impact on further iterations of the algorithm.
  
    Covering algorithm:
+
* Rule search:
        Exclusive: after covering a learning instance, remove it from further consideration.
+
** Evaluation measure: select a heuristic to evaluate found hypotheses:
        Weighted: after covering a learning instance, decrease its weight (multiplication by gamma) and in-turn decrease its impact on further iterations of the algorithm.
+
*** Entropy (measure of unpredictability of content)
 +
*** Laplace Accuracy
 +
*** Weighted Relative Accuracy
 +
** Beam width; remember the best rule found thus far and monitor a fixed number of alternatives (the beam).
  
    Rule search:
+
* Rule filtering:
        Evaluation measure: select a heuristic to evaluate found hypotheses:
+
** Minimum rule coverage: found rules must cover at least the minimum required number of covered examples. Unordered rules must cover this many target class examples.
            Entropy (measure of unpredictability of content)
+
** Maximum rule length: found rules may combine at most the maximum allowed number of selectors (conditions).
            Laplace Accuracy
+
** Default alpha: significance testing to prune out most specialised (less frequently applicable) rules in regard to the initial distribution of classes.
            Weighted Relative Accuracy
+
** Parent alpha: significance testing to prune out most specialised (less frequently applicable) rules in regard to the parent class distribution.
        Beam width; remember the best rule found thus far and monitor a fixed number of alternatives (the beam).
+
* Tick ‘Apply Automatically’ to auto-communicate changes to other widgets and to immediately train the classifier if learning data is connected. Alternatively, press ‘Apply‘ after configuration.
 
 
    Rule filtering:
 
        Minimum rule coverage: found rules must cover at least the minimum required number of covered examples. Unordered rules must cover this many target class examples.
 
        Maximum rule length: found rules may combine at most the maximum allowed number of selectors (conditions).
 
        Default alpha: significance testing to prune out most specialised (less frequently applicable) rules in regard to the initial distribution of classes.
 
        Parent alpha: significance testing to prune out most specialised (less frequently applicable) rules in regard to the parent class distribution.
 
    Tick ‘Apply Automatically’ to auto-communicate changes to other widgets and to immediately train the classifier if learning data is connected. Alternatively, press ‘Apply‘ after configuration.
 
  
 
==Contoh==
 
==Contoh==
Line 56: Line 55:
 
==Referensi==
 
==Referensi==
  
    Fürnkranz, Johannes. “Separate-and-Conquer Rule Learning”, Artificial Intelligence Review 13, 3-54, 1999.
+
* Fürnkranz, Johannes. “Separate-and-Conquer Rule Learning”, Artificial Intelligence Review 13, 3-54, 1999.
 
+
* Clark, Peter and Tim Niblett. “The CN2 Induction Algorithm”, Machine Learning Journal, 3 (4), 261-283, 1989.
    Clark, Peter and Tim Niblett. “The CN2 Induction Algorithm”, Machine Learning Journal, 3 (4), 261-283, 1989.
+
* Clark, Peter and Robin Boswell. “Rule Induction with CN2: Some Recent Improvements”, Machine Learning - Proceedings of the 5th European Conference (EWSL-91),151-163, 1991.
 
+
* Lavrač, Nada et al. “Subgroup Discovery with CN2-SD”,Journal of Machine Learning Research 5, 153-188, 2004
    Clark, Peter and Robin Boswell. “Rule Induction with CN2: Some Recent Improvements”, Machine Learning - Proceedings of the 5th European Conference (EWSL-91),151-163, 1991.
 
 
 
    Lavrač, Nada et al. “Subgroup Discovery with CN2-SD”,Journal of Machine Learning Research 5, 153-188, 2004
 
  
  

Revision as of 10:30, 28 January 2020

Sumber: https://docs.biolab.si//3/visual-programming/widgets/model/cn2ruleinduction.html

Induce rules from data using CN2 algorithm.

Input

Data: input dataset
Preprocessor: preprocessing method(s)

Output

Learner: CN2 learning algorithm
CN2 Rule Classifier: trained model

The CN2 algorithm is a classification technique designed for the efficient induction of simple, comprehensible rules of form “if cond then predict class”, even in domains where noise may be present.

CN2 Rule Induction works only for classification.

CN2-stamped.png
  • Name under which the learner appears in other widgets. The default name is CN2 Rule Induction.
  • Rule ordering:
    • Ordered: induce ordered rules (decision list). Rule conditions are found and the majority class is assigned in the rule head.
    • Unordered: induce unordered rules (rule set). Learn rules for each class individually, in regard to the original learning data.
  • Covering algorithm:
    • Exclusive: after covering a learning instance, remove it from further consideration.
    • Weighted: after covering a learning instance, decrease its weight (multiplication by gamma) and in-turn decrease its impact on further iterations of the algorithm.
  • Rule search:
    • Evaluation measure: select a heuristic to evaluate found hypotheses:
      • Entropy (measure of unpredictability of content)
      • Laplace Accuracy
      • Weighted Relative Accuracy
    • Beam width; remember the best rule found thus far and monitor a fixed number of alternatives (the beam).
  • Rule filtering:
    • Minimum rule coverage: found rules must cover at least the minimum required number of covered examples. Unordered rules must cover this many target class examples.
    • Maximum rule length: found rules may combine at most the maximum allowed number of selectors (conditions).
    • Default alpha: significance testing to prune out most specialised (less frequently applicable) rules in regard to the initial distribution of classes.
    • Parent alpha: significance testing to prune out most specialised (less frequently applicable) rules in regard to the parent class distribution.
  • Tick ‘Apply Automatically’ to auto-communicate changes to other widgets and to immediately train the classifier if learning data is connected. Alternatively, press ‘Apply‘ after configuration.

Contoh

For the example below, we have used zoo dataset and passed it to CN2 Rule Induction. We can review and interpret the built model with CN2 Rule Viewer widget.

CN2-visualize.png

The second workflow tests evaluates CN2 Rule Induction and Tree in Test & Score.

CN2-classification.png


Referensi

  • Fürnkranz, Johannes. “Separate-and-Conquer Rule Learning”, Artificial Intelligence Review 13, 3-54, 1999.
  • Clark, Peter and Tim Niblett. “The CN2 Induction Algorithm”, Machine Learning Journal, 3 (4), 261-283, 1989.
  • Clark, Peter and Robin Boswell. “Rule Induction with CN2: Some Recent Improvements”, Machine Learning - Proceedings of the 5th European Conference (EWSL-91),151-163, 1991.
  • Lavrač, Nada et al. “Subgroup Discovery with CN2-SD”,Journal of Machine Learning Research 5, 153-188, 2004


Referensi

Pranala Menarik