regression_tree
Regression tree regressor supporting continuous and mixed-feature
datasets. The library implements the regressor_protocol defined in
the regression_protocols library and learns a binary regression tree
using recursive variance-reduction splits that select the encoded
feature threshold maximizing variance reduction.
API documentation
Open the ../../apis/library_index.html#regression_tree link in a web browser.
Loading
To load this library, load the loader.lgt file:
| ?- logtalk_load(regression_tree(loader)).
Testing
To test this library predicates, load the tester.lgt file:
| ?- logtalk_load(regression_tree(tester)).
To run the performance benchmark suite, load the
tester_performance.lgt file:
| ?- logtalk_load(regression_tree(tester_performance)).
Export header format
The shared exporter in the regressor_common category writes a header
before the exported clauses in the following format:
% exported regressor predicate: Functor/Arity
% training dataset: Dataset
% target: Target
% attributes: Attributes
% diagnostics: Diagnostics
% Functor(Encoders, FeatureLabels, Tree, Diagnostics)
Functor(Encoders, FeatureLabels, Tree, Diagnostics)
The exported clauses serialize the learned regressor state so that
loading the file gives a regressor term that can be passed directly to
the predict/3 predicate.
When exporting a serialized regressor term, using a noun such as
regressor/4 or model/4 is recommended.
Features
Variance-Reduction Splits: Selects binary thresholds over encoded features to reduce target variance.
Continuous and Mixed Features: Supports continuous attributes and categorical attributes.
Categorical Features Encoding: Uses reference-level dummy coding derived from the declared dataset attribute values, with a missing-value indicator, and the resulting encoded features are treated as ordinary numeric split features.
Missing Values: Missing feature values represented using anonymous variables or omitted attribute-value pairs are encoded using explicit missing-value indicator features during both training and prediction.
Per-Split Feature Sampling: Optionally samples a subset of dataset attributes at each split before searching for the best partition.
Optional Feature Scaling: Continuous attributes can be standardized using z-score scaling before tree induction.
Diagnostics Metadata: Learned regressors record model name, target, training example count, encoded feature count, and effective options, accessible using the shared regression diagnostics predicates.
Model Export: Learned regressors can be exported as predicate clauses or written to a file.
Readable Trees: Includes a pretty-printer for inspecting learned tree structure.
Reference Benchmarks: Includes a dedicated performance suite reporting training time, RMSE, and MAE for representative regression datasets.
Regressor representation
The learned regressor is represented by default as:
regression_tree_regressor(Encoders, FeatureLabels, Tree, Diagnostics)
In this representation, Tree is built from leaf(Prediction) and
node(Index, Threshold, FallbackPrediction, Left, Right) terms and
Diagnostics stores training metadata including the effective
options.
Diagnostics syntax
The diagnostics/2 predicate returns a list of metadata terms with
the form:
[
model(regression_tree),
target(Target),
training_example_count(TrainingExampleCount),
options(Options),
encoded_feature_count(FeatureCount)
]
Where:
model(regression_tree)identifies the learning algorithm that produced the regressor.target(Target)stores the target attribute name declared by the training dataset.training_example_count(TrainingExampleCount)stores the number of examples used during training.options(Options)stores the effective learning options after merging the user options with the library defaults.encoded_feature_count(FeatureCount)stores the number of numeric features induced by the encoder list, including missing-value indicator features.
Use the regression_protocols diagnostic/2 and
regressor_options/2 helper predicates when you only need a single
metadata term or the effective options.
Options
The learn/3 predicate accepts the following options:
maximum_depth/1: Maximum depth allowed for the induced regression tree. Lower values yield smaller trees; higher values allow more detailed partitioning of the training data. The default is10.minimum_samples_leaf/1: Minimum number of training examples required in a leaf. This option also prevents candidate splits that would create child nodes smaller than the requested size. The default is1.minimum_variance_reduction/1: Minimum variance-reduction gain required for accepting a split. Higher values make the learner more conservative by pruning weak splits during induction. The default is0.0.maximum_features_per_split/1: Number of dataset attributes sampled at each split when searching for the best partition. Accepted values are a positive integer orall. The default isall.feature_scaling/1: Controls z-score standardization of continuous attributes before tree induction. Accepted values aretrueandfalse. The default isfalse.