iter_progressive_val_score¶
Evaluates the performance of a model on a streaming dataset and yields results.
This does exactly the same as evaluate.progressive_val_score
. The only difference is that this function returns an iterator, yielding results at every step. This can be useful if you want to have control over what you do with the results. For instance, you might want to plot the results.
Parameters¶
-
dataset
Type → base.typing.Dataset
The stream of observations against which the model will be evaluated.
-
model
The model to evaluate.
-
metric
Type → metrics.base.Metric
The metric used to evaluate the model's predictions.
-
moment
Type → str | typing.Callable | None
Default →
None
The attribute used for measuring time. If a callable is passed, then it is expected to take as input a
dict
of features. IfNone
, then the observations are implicitly timestamped in the order in which they arrive. -
delay
Type → str | int | dt.timedelta | typing.Callable | None
Default →
None
The amount to wait before revealing the target associated with each observation to the model. This value is expected to be able to sum with the
moment
value. For instance, ifmoment
is adatetime.date
, thendelay
is expected to be adatetime.timedelta
. If a callable is passed, then it is expected to take as input adict
of features and the target. If astr
is passed, then it will be used to access the relevant field from the features. IfNone
is passed, then no delay will be used, which leads to doing standard online validation. -
step
Default →
1
Iteration number at which to yield results. This only takes into account the predictions, and not the training steps.
-
measure_time
Default →
False
Whether or not to measure the elapsed time.
-
measure_memory
Default →
False
Whether or not to measure the memory usage of the model.
-
yield_predictions
Default →
False
Whether or not to include predictions. If step is 1, then this is equivalent to yielding the predictions at every iterations. Otherwise, not all predictions will be yielded.
Examples¶
Take the following model:
from river import linear_model
from river import preprocessing
model = (
preprocessing.StandardScaler() |
linear_model.LogisticRegression()
)
We can evaluate it on the Phishing
dataset as so:
from river import datasets
from river import evaluate
from river import metrics
steps = evaluate.iter_progressive_val_score(
model=model,
dataset=datasets.Phishing(),
metric=metrics.ROCAUC(),
step=200
)
for step in steps:
print(step)
{'ROCAUC': ROCAUC: 90.20%, 'Step': 200}
{'ROCAUC': ROCAUC: 92.25%, 'Step': 400}
{'ROCAUC': ROCAUC: 93.23%, 'Step': 600}
{'ROCAUC': ROCAUC: 94.05%, 'Step': 800}
{'ROCAUC': ROCAUC: 94.79%, 'Step': 1000}
{'ROCAUC': ROCAUC: 95.07%, 'Step': 1200}
{'ROCAUC': ROCAUC: 95.07%, 'Step': 1250}
The yield_predictions
parameter can be used to include the predictions in the results:
import itertools
steps = evaluate.iter_progressive_val_score(
model=model,
dataset=datasets.Phishing(),
metric=metrics.ROCAUC(),
step=1,
yield_predictions=True
)
for step in itertools.islice(steps, 100, 105):
print(step)
{'ROCAUC': ROCAUC: 94.68%, 'Step': 101, 'Prediction': {False: 0.966..., True: 0.033...}}
{'ROCAUC': ROCAUC: 94.75%, 'Step': 102, 'Prediction': {False: 0.035..., True: 0.964...}}
{'ROCAUC': ROCAUC: 94.82%, 'Step': 103, 'Prediction': {False: 0.043..., True: 0.956...}}
{'ROCAUC': ROCAUC: 94.89%, 'Step': 104, 'Prediction': {False: 0.816..., True: 0.183...}}
{'ROCAUC': ROCAUC: 94.96%, 'Step': 105, 'Prediction': {False: 0.041..., True: 0.958...}}