Skip to content

MLPRegressor

Multi-layer Perceptron for regression.

This model is still work in progress. Here are some features that still need implementing:

  • learn_one and predict_one just cast the input dict to a single row dataframe and then

    call learn_many and predict_many respectively. This is very inefficient. - Not all of the optimizers in the optim module can be used as they are not all vectorised.

  • Emerging and disappearing features are not supported. Each instance/batch has to have the

    same features. - The gradient haven't been numerically checked.

Parameters

  • hidden_dims

    The dimensions of the hidden layers. For example, specifying (10, 20) means that there are two hidden layers with 10 and 20 neurons, respectively. Note that the number of layers the network contains is equal to the number of hidden layers plus two (to account for the input and output layers).

  • activations

    The activation functions to use at each layer, including the input and output layers. Therefore you need to specify three activation if you specify one hidden layer.

  • loss

    Typeoptim.losses.Loss | None

    DefaultNone

    Loss function. Defaults to optim.losses.Squared.

  • optimizer

    Typeoptim.base.Optimizer | None

    DefaultNone

    Optimizer. Defaults to optim.SGD with the learning rate set to 0.01.

  • seed

    Typeint | None

    DefaultNone

    Random number generation seed. Set this for reproducibility.

Attributes

  • n_layers

    Return the number of layers in the network. The number of layers is equal to the number of hidden layers plus 2. The 2 accounts for the input layer and the output layer.

Examples

from river import datasets
from river import evaluate
from river import neural_net as nn
from river import optim
from river import preprocessing as pp
from river import metrics

model = (
    pp.StandardScaler() |
    nn.MLPRegressor(
        hidden_dims=(5,),
        activations=(
            nn.activations.ReLU,
            nn.activations.ReLU,
            nn.activations.Identity
        ),
        optimizer=optim.SGD(1e-3),
        seed=42
    )
)

dataset = datasets.TrumpApproval()

metric = metrics.MAE()

evaluate.progressive_val_score(dataset, model, metric)
MAE: 1.589827

You can also use this to process mini-batches of data.

model = (
    pp.StandardScaler() |
    nn.MLPRegressor(
        hidden_dims=(10,),
        activations=(
            nn.activations.ReLU,
            nn.activations.ReLU,
            nn.activations.ReLU
        ),
        optimizer=optim.SGD(1e-4),
        seed=42
    )
)

dataset = datasets.TrumpApproval()
batch_size = 32

for epoch in range(10):
    for xb in pd.read_csv(dataset.path, chunksize=batch_size):
        yb = xb.pop('five_thirty_eight')
        y_pred = model.predict_many(xb)
        model = model.learn_many(xb, yb)

model.predict_many(xb)
      five_thirty_eight
992           39.361609
993           46.398536
994           42.094086
995           40.195802
996           40.782954
997           40.839678
998           40.896403
999           48.362659
1000          42.021849

Methods

call

Make predictions.

Parameters

  • X'pd.DataFrame'

learn_many

Train the network.

Parameters

  • X'pd.DataFrame'
  • y'pd.DataFrame'

learn_one

Fits to a set of features x and a real-valued target y.

Parameters

  • x'dict'
  • y'base.typing.RegTarget'

Returns

Regressor: self

predict_many
predict_one

Predict the output of features x.

Parameters

  • x'dict'

Returns

base.typing.RegTarget: The prediction.