sgd_classifier
Linear stochastic-gradient classifier for tabular datasets using a one-vs-rest scheme with configurable losses. The implementation reuses the shared linear encoder pipeline so continuous and categorical features, including missing values, are represented consistently with the existing linear classifiers.
The library implements the classifier_protocol defined in the
classification_protocols library. It provides predicates for
learning a classifier from a dataset, using it to make predictions,
estimating class probabilities, and exporting it as a list of predicate
clauses or to a file.
Datasets are represented as objects implementing the
dataset_protocol protocol from the classification_protocols
library. Continuous, categorical, and mixed-feature datasets are
supported.
API documentation
Open the ../../docs/library_index.html#sgd_classifier link in a web browser.
Loading
To load this library, load the loader.lgt file:
| ?- logtalk_load(sgd_classifier(loader)).
Testing
To test this library predicates, load the tester.lgt file:
| ?- logtalk_load(sgd_classifier(tester)).
Features
Binary and Multiclass Classification: Learns one-vs-rest linear models and predicts the class with the highest decision score.
Multiple Losses: Supports
log_loss,hinge,squared_hinge,modified_huber, andperceptronlosses.Mixed Features: Reuses the shared tabular encoders for continuous and categorical attributes, including missing-value indicators.
Probability Estimation: Provides class probabilities using a softmax over linear decision scores.
Configurable Optimization: Exposes learning-rate scheduling, convergence tolerance, and L2 regularization options.
Classifier Export: Learned classifiers can be exported as predicate clauses or written to a file.
Options
The learn/3 predicate supports these options:
loss/1- optimization loss, one oflog_loss,hinge,squared_hinge,modified_huber, orperceptron(default:log_loss)learning_rate/1- base learning rate for optimization (default:0.05)learning_schedule/1- learning-rate schedule, eitherconstantorinverse_scaling(Power)(default:constant)maximum_iterations/1- maximum number of optimization epochs (default:100)tolerance/1- convergence threshold for the maximum parameter update (default:1.0e-5)l2_regularization/1- L2 penalty factor applied to the weight vectors (default:0.0001)feature_scaling/1- whether to standardize continuous attributes before encoding (default:true)
Usage
Learning a classifier
| ?- sgd_classifier::learn(weather, Classifier).
| ?- sgd_classifier::learn(mixed, Classifier, [loss(hinge), maximum_iterations(250)]).
Making predictions
| ?- sgd_classifier::learn(weather, Classifier),
sgd_classifier::predict(Classifier, [outlook-rainy, temperature-mild, humidity-normal, windy-false], Class).
| ?- sgd_classifier::learn(missing_mixed, Classifier),
sgd_classifier::predict_probabilities(Classifier, [age-38, income-72000, student-yes, credit_rating-fair], Probabilities).
Exporting the classifier
| ?- sgd_classifier::learn(weather, Classifier),
sgd_classifier::export_to_clauses(weather, Classifier, classify, Clauses).
| ?- sgd_classifier::learn(weather, Classifier),
sgd_classifier::export_to_file(weather, Classifier, classify, 'classifier.pl').
Classifier representation
The learned classifier is represented as a compound term with the form:
sgd_classifier(Classes, Encoders, Loss, Models, Options)
Where:
Classes: list of class labelsEncoders: list of continuous scaling descriptors and categorical value encodersLoss: selected optimization lossModels: list ofclass_model(Class, Bias, Weights)termsOptions: merged training options used to learn the classifier
When exported using export_to_clauses/4 or export_to_file/4,
this classifier term is serialized directly as the single argument of
the generated predicate clause so that the exported model can be loaded
and reused as-is.
References
Bottou, L. (2010). “Large-Scale Machine Learning with Stochastic Gradient Descent”.
Shalev-Shwartz, S. and Ben-David, S. (2014). “Understanding Machine Learning”. Chapter 15.
Hastie, T., Tibshirani, R. and Friedman, J. (2009). “The Elements of Statistical Learning”. Chapter 12.