Predicate selection for structural decision trees
We study predicate selection functions (also known as splitting rules) for structural decision trees and propose two improvements to existing schemes. The first is in classification learning, where we reconsider the use of accuracy as a predicate selection function and show that, on practical grounds, it is a better alternative to other commonly used functions. The second is in regression learning, where we consider the standard mean squared error measure and give a predicate pruning result for...[Show more]
|Collections||ANU Research Publications|
|Source:||Inductive Logic Programming: Proceedings of the 15th International Conference on Inductive Logic Programming|
Items in Open Research are protected by copyright, with all rights reserved, unless otherwise indicated.