Difference between revisions of "Orange: CN2 Rule Induction"
Jump to navigation
Jump to search
Onnowpurbo (talk | contribs) (Created page with "Sumber: https://docs.biolab.si//3/visual-programming/widgets/model/cn2ruleinduction.html Induce rules from data using CN2 algorithm. Inputs Data: input dataset Pre...") |
Onnowpurbo (talk | contribs) |
||
Line 6: | Line 6: | ||
Data: input dataset | Data: input dataset | ||
− | |||
Preprocessor: preprocessing method(s) | Preprocessor: preprocessing method(s) | ||
Line 12: | Line 11: | ||
Learner: CN2 learning algorithm | Learner: CN2 learning algorithm | ||
− | |||
CN2 Rule Classifier: trained model | CN2 Rule Classifier: trained model | ||
Line 19: | Line 17: | ||
CN2 Rule Induction works only for classification. | CN2 Rule Induction works only for classification. | ||
− | + | [[File:CN2-stamped.png|center|200px|thumb]] | |
Name under which the learner appears in other widgets. The default name is CN2 Rule Induction. | Name under which the learner appears in other widgets. The default name is CN2 Rule Induction. | ||
Rule ordering: | Rule ordering: | ||
− | |||
Ordered: induce ordered rules (decision list). Rule conditions are found and the majority class is assigned in the rule head. | Ordered: induce ordered rules (decision list). Rule conditions are found and the majority class is assigned in the rule head. | ||
− | |||
Unordered: induce unordered rules (rule set). Learn rules for each class individually, in regard to the original learning data. | Unordered: induce unordered rules (rule set). Learn rules for each class individually, in regard to the original learning data. | ||
Covering algorithm: | Covering algorithm: | ||
− | |||
Exclusive: after covering a learning instance, remove it from further consideration. | Exclusive: after covering a learning instance, remove it from further consideration. | ||
− | |||
Weighted: after covering a learning instance, decrease its weight (multiplication by gamma) and in-turn decrease its impact on further iterations of the algorithm. | Weighted: after covering a learning instance, decrease its weight (multiplication by gamma) and in-turn decrease its impact on further iterations of the algorithm. | ||
Rule search: | Rule search: | ||
− | |||
Evaluation measure: select a heuristic to evaluate found hypotheses: | Evaluation measure: select a heuristic to evaluate found hypotheses: | ||
− | |||
Entropy (measure of unpredictability of content) | Entropy (measure of unpredictability of content) | ||
− | |||
Laplace Accuracy | Laplace Accuracy | ||
− | |||
Weighted Relative Accuracy | Weighted Relative Accuracy | ||
− | |||
Beam width; remember the best rule found thus far and monitor a fixed number of alternatives (the beam). | Beam width; remember the best rule found thus far and monitor a fixed number of alternatives (the beam). | ||
Rule filtering: | Rule filtering: | ||
− | |||
Minimum rule coverage: found rules must cover at least the minimum required number of covered examples. Unordered rules must cover this many target class examples. | Minimum rule coverage: found rules must cover at least the minimum required number of covered examples. Unordered rules must cover this many target class examples. | ||
− | |||
Maximum rule length: found rules may combine at most the maximum allowed number of selectors (conditions). | Maximum rule length: found rules may combine at most the maximum allowed number of selectors (conditions). | ||
− | |||
Default alpha: significance testing to prune out most specialised (less frequently applicable) rules in regard to the initial distribution of classes. | Default alpha: significance testing to prune out most specialised (less frequently applicable) rules in regard to the initial distribution of classes. | ||
− | |||
Parent alpha: significance testing to prune out most specialised (less frequently applicable) rules in regard to the parent class distribution. | Parent alpha: significance testing to prune out most specialised (less frequently applicable) rules in regard to the parent class distribution. | ||
− | |||
Tick ‘Apply Automatically’ to auto-communicate changes to other widgets and to immediately train the classifier if learning data is connected. Alternatively, press ‘Apply‘ after configuration. | Tick ‘Apply Automatically’ to auto-communicate changes to other widgets and to immediately train the classifier if learning data is connected. Alternatively, press ‘Apply‘ after configuration. | ||
− | + | ==Contoh== | |
For the example below, we have used zoo dataset and passed it to CN2 Rule Induction. We can review and interpret the built model with CN2 Rule Viewer widget. | For the example below, we have used zoo dataset and passed it to CN2 Rule Induction. We can review and interpret the built model with CN2 Rule Viewer widget. | ||
− | + | [[File:CN2-visualize.png|center|200px|thumb]] | |
The second workflow tests evaluates CN2 Rule Induction and Tree in Test & Score. | The second workflow tests evaluates CN2 Rule Induction and Tree in Test & Score. | ||
− | + | [[File:CN2-classification.png|center|200px|thumb]] | |
− | + | ||
+ | |||
+ | ==Referensi== | ||
Fürnkranz, Johannes. “Separate-and-Conquer Rule Learning”, Artificial Intelligence Review 13, 3-54, 1999. | Fürnkranz, Johannes. “Separate-and-Conquer Rule Learning”, Artificial Intelligence Review 13, 3-54, 1999. |
Revision as of 08:45, 23 January 2020
Sumber: https://docs.biolab.si//3/visual-programming/widgets/model/cn2ruleinduction.html
Induce rules from data using CN2 algorithm.
Inputs
Data: input dataset Preprocessor: preprocessing method(s)
Outputs
Learner: CN2 learning algorithm CN2 Rule Classifier: trained model
The CN2 algorithm is a classification technique designed for the efficient induction of simple, comprehensible rules of form “if cond then predict class”, even in domains where noise may be present.
CN2 Rule Induction works only for classification.
Name under which the learner appears in other widgets. The default name is CN2 Rule Induction.
Rule ordering: Ordered: induce ordered rules (decision list). Rule conditions are found and the majority class is assigned in the rule head. Unordered: induce unordered rules (rule set). Learn rules for each class individually, in regard to the original learning data.
Covering algorithm: Exclusive: after covering a learning instance, remove it from further consideration. Weighted: after covering a learning instance, decrease its weight (multiplication by gamma) and in-turn decrease its impact on further iterations of the algorithm.
Rule search: Evaluation measure: select a heuristic to evaluate found hypotheses: Entropy (measure of unpredictability of content) Laplace Accuracy Weighted Relative Accuracy Beam width; remember the best rule found thus far and monitor a fixed number of alternatives (the beam).
Rule filtering: Minimum rule coverage: found rules must cover at least the minimum required number of covered examples. Unordered rules must cover this many target class examples. Maximum rule length: found rules may combine at most the maximum allowed number of selectors (conditions). Default alpha: significance testing to prune out most specialised (less frequently applicable) rules in regard to the initial distribution of classes. Parent alpha: significance testing to prune out most specialised (less frequently applicable) rules in regard to the parent class distribution. Tick ‘Apply Automatically’ to auto-communicate changes to other widgets and to immediately train the classifier if learning data is connected. Alternatively, press ‘Apply‘ after configuration.
Contoh
For the example below, we have used zoo dataset and passed it to CN2 Rule Induction. We can review and interpret the built model with CN2 Rule Viewer widget.
The second workflow tests evaluates CN2 Rule Induction and Tree in Test & Score.
Referensi
Fürnkranz, Johannes. “Separate-and-Conquer Rule Learning”, Artificial Intelligence Review 13, 3-54, 1999.
Clark, Peter and Tim Niblett. “The CN2 Induction Algorithm”, Machine Learning Journal, 3 (4), 261-283, 1989.
Clark, Peter and Robin Boswell. “Rule Induction with CN2: Some Recent Improvements”, Machine Learning - Proceedings of the 5th European Conference (EWSL-91),151-163, 1991.
Lavrač, Nada et al. “Subgroup Discovery with CN2-SD”,Journal of Machine Learning Research 5, 153-188, 2004