Tutorial K: Creating a new baseline for (supervised) Seldonian Experiments
Contents
Introduction
In a Seldonian Experiment, it is often useful to be able to compare the Seldonian model(s) you have trained to other machine learning models, which we refer to as "baselines" from here on out. These baseline models fall into two categories: i) "standard" ML models that are not fairness-aware and ii) fairness-aware (or safety aware) ML models. This tutorial tackles how to compare to baseline models of the first kind, i.e., standard ML models, in Seldonian Experiments.
If you have gone through any of the previous tutorials which run Seldonian Experiments, such as the Fair Loans Tutorial, then you have probably noticed that the Seldonian Experiments library has some pre-defined baseline models already available. These existing baselines include models for logistic regression, linear regression, and a convolutional neural network implemented with PyTorch. The source code for these baselines can be found here. We could not possibly include all models with which users might want to compare, so we designed the Experiments library to make it straightforward to add your own baseline models.
In this tutorial, we will demonstrate how a user might add a new baseline model for their specific supervised learning problem. We will implement a random forest classifier baseline model for binary classification and show how to run an experiment comparing a Seldonian model to it.
Requirements for a new baseline
Baselines are defined as Python classes. There are three requirements for defining a new baseline class.
-
The class must define a
model_name
string attribute in its __init__()
method. Make this string unique and descriptive for each baseline.
-
The class must have a
train()
method that takes features and labels as input and trains the model. This method should return the trained model parameters, if there are any. If not (e.g., for non-parametric models), return None. If returning None, save the trained model as a class attribute so that it can be referenced in the predict()
method (see next requirement).
-
The class must have a
predict()
method which takes the fitted model parameters (the return value of train()
) and features as input and returns the predicted labels. For classification problems, the predictions should be probabilities for each class, as opposed to the integer class label.
Note: These requirements are similar to those for the underlying supervised models used in the Seldonian Engine, and it turns out that those models (without constraints) can be used as baselines with the simple addition of a train()
method. One method of achieving this is to create a new class for the baseline which inherits from the Seldonian model and then add a train()
method. It is often useful to compare the model trained subject to constraints with the engine to the same model trained without constraints, and this trick allows you do that easily. For an example of how this method works, see the source code for the built-in logistic regression baseline model: logistic_regression.py. Note that the predict()
method is a parent method, so it is inheritied from the parent and not reiterated in the child class.
Creating a random forest classifier baseline
To create a class for a random forest classifier baseline, we will wrap scikit-learn's random forest classifier functionality to do all of the hard work of training and making predictions. Below is all of the code we need to implement this baseline.
In __init__()
, we define the model name, which fulfills requirement #1. __init__()
takes in one optional argument, n_estimators
, which is the number of trees we want to use in the random forest classifier. This argument was added as an example of how you would pass hyperparameters to the classifier. scikit-learn's classifiers have lots of hyperparameters, and in practice if you wanted the ability to specify more hyperparameters, you would simply add more arguments (or a dictionary of arguments) to __init__()
.
The train()
method takes features X
and labels y
, trains the model and returns None
, fulfilling requirement #2. Looking into the body of this method, we make a class attribute self.trained_model
, an instantiation of a scikit-learn RandomForestClassifier
class, so that we can refer to it later in the predict()
method. Note that when we instantiate this class, we pass in the hyperparameter n_estimators
via the class attribute we set in __init__()
. Next, we train the model using the fit()
method of the scikit-learn model. It is important that a new classifier is instantiated in each call to train()
because we want to make sure we train a new model in each experiment trial, instead of retraining the same model instance in each trial.
The predict()
method takes the fitted model parameters theta
and the test features X
as input and returns the predicted class probabilities, fulfilling requirement #3. For random forest models, we do not have fitted model parameters, which is why train()
returned None
. As a result, theta
will be passed into this method as None
, and notice that we don't use theta
anywhere in the body of the method. The reason that theta
is a required input to predict()
is because for other model types (such as logistic regression), the fitted parameters may be needed to make the predictions, and the experiments library does not know ahead of time whether your model is parametric or non-parametric.
Instead of using theta
in predict()
, we obtain the fitted model via the self.trained_model
attribute. train()
is run before predict()
in each experiment trial, so self.trained_model
will be set by the time predict()
is called. To make the class probability predictions, we use scikit-learn's predict_proba()
method. For binary classification problems, predict_proba()
returns a two dimensional array, where the first column is the probabilities of the 0th class and the second column is the probabilities of the 1st class. In the Seldonian Toolkit (both the Engine and the Experiments libraries), a convention we adopt for binary classification problems is that the predict()
methods only return the probabilities of the 1st class (the "positive" class), since the 0th are redundant with these. This is why we include the last line: probs_posclass = probs_bothclasses[:,1]
Using the random forest classifier baseline in an experiment
Using this new baseline in an experiment is simple. Let's say we saved the baseline we created above to a file called random_forest.py
. In our experiment script, all we need to do is import the baseline class and tell the plot generator to use an instance of this class. Below is a minimal example showing how to do this with the loans dataset described in the Fair Loans Tutorial. Our experiment will consist of 10 trials with 10 different data fractions ranging from 0.001 to 1.0. We will specify the n_estimator
hyperparameter to be different than the default. We will reference a specification file (contains the "spec" object) that we created in that tutorial, which can be found in this directory.
Below is the code to run the experiment.
If we save the above code in a file called experiment_with_random_forest_baseline.py
, and run it from the command line, it will generate a plot file called results/loans_random_forest/loans_random_forest.pdf
, which should look like this:
Summary
In this tutorial, we demonstrated how to create a new baseline for Seldonian Experiments. We described the general requirements for implementing a new baseline and provided an example implementation in a Seldonian Experiment using the dataset from the Fair Loans Tutorial. Our implementation wrapped scikit-learn's random forest classifier, and the method we used can be used to implement any scikit-learn supervised ML model as a baseline in a Seldonian Experiment. Baselines using other Python librarires can similarly be implemented using the patterns described in this tutorial. If you develop a baseline that you think could be generally useful to others, feel free to submit a pull request on the
Experiments library Github page.