KumoRFM provides an evaluation mode that automatically measures prediction quality by performing a train/test split on context examples and computing relevant metrics.
Running an Evaluation
UseKumoRFM.evaluate() with the same PQL syntax as `KumoRFM.predict()`:
EVALUATE keyword in the query string directly:
Available Metrics
The metrics returned depend on the detected task type:| Task Type | Supported Metrics |
|---|---|
| Binary Classification | accuracy, precision, recall, f1, mrr, auc |
| Multi-Class Classification | acc, precision, recall, f1, mrr |
| Regression / Forecasting | mae, mape, mse, rmse, smape, r2 |
Evaluation Parameters
TheKumoRFM.evaluate() method accepts the same parameters as KumoRFM.predict(), plus:
metrics: A list of metric names to compute. If not specified, all applicable metrics for the task type are computed.
run_mode, anchor_time, num_hops, and other parameters work identically to KumoRFM.predict(). See configuration for details on run modes.
Evaluation with TaskTable
For advanced use cases, you can construct aTaskTable explicitly and use `KumoRFM.evaluate_task()`:
Interpreting Results
The evaluation returns a dictionary mapping metric names to values:r2, accuracy, precision, recall, f1, and auc. Lower values are better for mae, mape, mse, rmse, smape.