Zero-shot prediction makes accurate predictions on tasks never seen during training. No task-specific labels. No fine-tuning. No feature engineering. You point a foundation model at a new database and it generates predictions immediately. This is possible because graph foundation models learn universal relational patterns during pre-training: what declining customer engagement looks like, how transaction velocity anomalies predict fraud, how multi-hop relational patterns correlate with outcomes.
When the model encounters a new e-commerce database it has never seen, it recognizes the relational structure (customers linked to orders linked to products) and applies its learned patterns. Customers whose order frequency is declining, whose product preferences are shifting, and whose support interactions are increasing are likely churners. The model knows this from pre-training on other databases where similar patterns preceded churn.
How zero-shot works for relational data
# KumoRFM zero-shot prediction (conceptual)
# 1. Point at your database
database = connect_to_database('postgresql://...')
# 2. Specify what to predict (PQL - Predictive Query Language)
query = """
PREDICT customer.will_churn
FROM customer, order, product
WHERE prediction_date = '2026-04-01'
"""
# 3. Get predictions immediately (no training)
predictions = kumo_rfm.predict(database, query)
# Under the hood:
# - Reads schema: tables, columns, foreign keys, timestamps
# - Builds heterogeneous temporal graph automatically
# - Applies pre-trained relational graph transformer
# - Generates per-customer churn probabilities
# predictions['customer_id_123'].churn_probability = 0.82
# No labels needed. No model training. No feature engineering.Zero-shot prediction: connect to database, write PQL query, get predictions. The pre-trained model handles everything.
What makes zero-shot possible
Zero-shot prediction works because relational patterns are universal:
- Declining engagement: whether it is order frequency, login frequency, or feature usage, a declining trajectory predicts churn across industries.
- Velocity anomalies: whether it is transaction speed, login attempts, or support ticket frequency, abnormal velocity predicts fraud or issues across domains.
- Relational proximity: entities connected to high-risk entities are higher risk themselves. This holds for fraud networks, churn clusters, and product return patterns.
- Temporal patterns: recency, frequency, and monetary patterns predict behavior across any transactional system.
The zero-shot to fine-tuned progression
Zero-shot is not the end state; it is the starting point:
- Day 1: Zero-shot - Immediate predictions. No labels needed. Good enough for ranking and prioritization. (76.71 AUROC)
- Week 2: Few-shot - As investigators confirm cases or analysts label examples, incorporate 10-100 labels. Performance improves incrementally.
- Month 1: Fine-tuned - With hundreds of accumulated labels, fine-tune the model for maximum accuracy. (81.14 AUROC)
- Ongoing: Continuous - As new labels arrive, periodically re-fine-tune. The model improves continuously with human feedback.
Enterprise example: new market expansion
A fintech company expands to a new country. They have no local transaction history, no local fraud labels, and no local data science team. But they have KumoRFM pre-trained on their other markets.
Zero-shot deployment:
- Connect KumoRFM to the new market's database
- The model recognizes the relational schema (accounts, transactions, merchants)
- It applies universal fraud patterns learned from other markets
- Day-one fraud scoring for every transaction, without a single local label
As the local team investigates flagged transactions and confirms fraud cases, those labels feed back into fine-tuning, progressively improving accuracy for the local market's specific fraud patterns.