KumoRFM training did not incorporate datasets from RelBench. The examples on this page evaluate KumoRFM on RelBench tasks without prior model training on RelBench benchmark data.
- it gives you benchmark datasets that are already structured for relational prediction
- it provides task definitions that make it easy to compare KumoRFM against other methods
What the RelBench Notebook Covers
The RelBench notebook shows how to evaluate KumoRFM on RelBench tasks in two different ways:- with a custom context table that turns a RelBench task into a missing value imputation problem
- with Predictive Query Language (PQL) that reproduces the task directly in KumoRFM
rel-f1 dataset as the main worked example and the driver-top3 task as the primary classification task.
What RelBench Includes
RelBench includes a growing set of realistic relational databases across domains such as sports, e-commerce, medicine, social platforms, and scientific publishing. The website highlights that the benchmark includes diverse databases, standardized evaluators, and automated data loading. The notebook focuses on RelBench as an evaluation framework rather than a single dataset. It demonstrates how the same KumoRFM workflow can transfer across multiple RelBench tasks.Notebook Workflow
At a high level, the notebook does the following:- Installs the required packages, including
kumoaiand RelBench. - Authenticates with KumoRFM and initializes the SDK.
- Imports RelBench dataset helpers.
- Defines a helper function that converts any RelBench dataset into a
LocalGraph. - Loads the
rel-f1dataset and builds its graph. - Prints metadata, links, and a visualization to verify the graph.
- Demonstrates the custom context table approach for the
driver-top3task. - Builds a context table by concatenating split-specific task tables and masking test labels.
- Adds the context table to the graph and links it to the task entity table.
- Runs a KumoRFM prediction query against that context table.
- Evaluates the predictions with AUROC.
- Demonstrates the PQL approach for the same task.
- Uses
model.get_train_table(...)to debug and validate label generation. - Groups test entities by anchor timestamp and predicts in batches.
- Evaluates the final predictions and compares the setup with the context table approach.
Two Evaluation Approaches
The notebook explicitly compares two ways to evaluate KumoRFM on RelBench tasks.Custom Context Table
This approach creates a dedicated context table that contains:- the entity identifier
- the anchor timestamp
- the target label
- you want a general recipe that works across many task types
- you want direct control over the context rows and labels
- you are evaluating a custom task that does not already have a clean PQL form
Predictive Query Language
The notebook also shows how to express RelBench tasks directly in PQL. For thedriver-top3 task on rel-f1, the notebook uses:
Practical Notes from the Notebook
The RelBench notebook highlights a few operational details that matter in practice:run_mode='best'can improve benchmark performance when runtime is less important than accuracyanchor_time='entity'is useful for context-table workflows where each row carries its own timestamp- batching by shared anchor timestamp helps KumoRFM reuse context efficiently
max_pq_iterationsmay need to be increased for more restrictive queries so that enough valid context examples are found
More Reading
- RelBench notebook with KumoRFM
- RelBench homepage
- RelBench quick start
- rel-f1 dataset page
- SALT If you want a more enterprise-oriented example after RelBench, the SALT notebook is a good next step because it shows how the same KumoRFM evaluation pattern applies to a large ERP-style sales dataset.